url
stringlengths
11
2.25k
text
stringlengths
88
50k
ts
timestamp[s]date
2026-01-13 08:47:33
2026-01-13 09:30:40
https://www.linkedin.com/products/ovhgroup-ovh-antiddos-protection/?trk=products_seo_search
OVHcloud Anti-DDoS Protection | LinkedIn Skip to main content LinkedIn OVHcloud in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in OVHcloud Anti-DDoS Protection DDoS Protection Software by OVHcloud See who's skilled in this Add as skill Learn more Report this product About Mitigate your risk of DDoS attacks and keep your services available at all times. Free with every OVH service. Learn more. Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less OVHcloud products OVHcloud Block Storage OVHcloud Block Storage Block Storage Software OVHcloud Marketplace OVHcloud Marketplace Marketplace Platforms OVHcloud Object Storage OVHcloud Object Storage Object Storage Software OVHcloud Zerto DRP OVHcloud Zerto DRP Disaster Recovery as a Service (DRaaS) Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-arbor-edge-defense/?trk=products_seo_search
Arbor Edge Defense | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Arbor Edge Defense DDoS Protection Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Arbor Edge Defense is an inline security appliance deployed at the network perimeter that can automatically detect and block inbound threats and outbound malicious communication using highly scalable, stateless technology and unique, global threat intelligence. This product is intended for Cyber Security Engineer Chief Executive Officer Chief Information Officer Network Operations Center Head of Security Security Engineer Director of Security Information Technology Specialist Cyber Security Specialist Media Products media viewer No more previous content Smart Perimeter Protection With NETSCOUT Arbor Edge Defense (AED) NETSCOUT Arbor Edge Defense (AED) blocks inbound threats such as DDoS attacks, and outbound communication from compromised internal hosts - acting as a first and last line of smart, automated, perimeter defense. Perimeter Defense Best Practices Using Arbor Edge Defense NETSCOUT Arbor Edge Defense acts as a first and last line of smart, automated perimeter defense for an organization. It’s deployed on-premise, where it acts as a first line of defense, by blocking inbound DDoS attacks, protecting stateful security devices. Demo: Blocking Ransomware Attack with Arbor Edge Defense Acting as a last line of defense, AED detects and blocks outbound indicators of compromise (or IoCs) that has been missed by other tools in your security stack to stop the proliferation of malware before a data breach. NETSCOUT’s Arbor Edge Defense: The First and Last Line of Defense Adam Bixler, Director, Product Management at NETSCOUT, discusses NETSCOUT AED’s unique functionality and how it augments and strengthens traditional endpoint security. Learn about the value of integrating threat intelligence and DDoS defense as a first and last line of defense. Visit the product page for more information: http://www.netscout.link/6001EPXxb Get VPN Protection with NETSCOUT Arbor Edge Defense A DDoS attack poses a major threat to the availability of the VPN gateway. As employees continue to work from home, protecting the availability of your VPN gateway from DDoS attacks is critical. Unlike a cloud-based DDoS protection solution, NETSCOUT Arbor Edge Defense is a stateless, on-premise solution that can instantaneously detect and mitigate DDoS attacks against the VPN gateway, enabling you to maintain home-based employee productivity and business continuity. No more next content Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less NETSCOUT products Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/uk/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d0%b8-Git-%d0%92%d0%b7%d0%b0%d1%94%d0%bc%d0%be%d0%b4%d1%96%d1%8f-%d0%b7-%d0%b2%d1%96%d0%b4%d0%b4%d0%b0%d0%bb%d0%b5%d0%bd%d0%b8%d0%bc%d0%b8-%d1%81%d1%85%d0%be%d0%b2%d0%b8%d1%89%d0%b0%d0%bc%d0%b8
Git - Взаємодія з віддаленими сховищами About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Вступ 1.1 Про систему контролю версій 1.2 Коротка історія Git 1.3 Основи Git 1.4 Git, зазвичай, тільки додає дані 1.5 Три стани 1.6 Командний рядок 1.7 Інсталяція Git 1.8 Початкове налаштування Git 1.9 Отримання допомоги 1.10 Підсумок 2. Основи Git 2.1 Створення Git-сховища 2.2 Запис змін до репозиторія 2.3 Перегляд історії комітів 2.4 Скасування речей 2.5 Взаємодія з віддаленими сховищами 2.6 Теґування 2.7 Псевдоніми Git 2.8 Підсумок 3. Галуження в git 3.1 Гілки у кількох словах 3.2 Основи галуження та зливання 3.3 Управління гілками 3.4 Процеси роботи з гілками 3.5 Віддалені гілки 3.6 Перебазовування 3.7 Підсумок 4. Git на сервері 4.1 Протоколи 4.2 Отримання Git на сервері 4.3 Генерація вашого публічного ключа SSH 4.4 Налаштування Серверу 4.5 Демон Git 4.6 Розумний HTTP 4.7 GitWeb 4.8 GitLab 4.9 Варіанти стороннього хостингу 4.10 Підсумок 5. Розподілений Git 5.1 Розподілені процеси роботи 5.2 Внесення змін до проекту 5.3 Супроводжування проекту 5.4 Підсумок 6. GitHub 6.1 Створення та налаштування облікового запису 6.2 Як зробити внесок до проекту 6.3 Супроводжування проєкту 6.4 Керування організацією 6.5 Скриптування GitHub 6.6 Підсумок 7. Інструменти Git 7.1 Вибір ревізій 7.2 Інтерактивне індексування 7.3 Ховання та чищення 7.4 Підписання праці 7.5 Пошук 7.6 Переписування історії 7.7 Усвідомлення скидання (reset) 7.8 Складне злиття 7.9 Rerere 7.10 Зневадження з Git 7.11 Підмодулі 7.12 Пакування 7.13 Заміна 7.14 Збереження посвідчення (credential) 7.15 Підсумок 8. Налаштування Git 8.1 Конфігурація Git 8.2 Атрибути Git 8.3 Гаки (hooks) Git 8.4 Приклад політики користування виконуваної Git-ом 8.5 Підсумок 9. Git and Other Systems 9.1 Git як клієнт 9.2 Міграція на Git 9.3 Підсумок 10. Git зсередини 10.1 Кухонні та парадні команди 10.2 Об’єкти Git 10.3 Посилання Git 10.4 Файли пакунки 10.5 Специфікація посилань (refspec) 10.6 Протоколи передачі 10.7 Супроводження та відновлення даних 10.8 Змінні середовища 10.9 Підсумок A1. Додаток A: Git в інших середовищах A1.1 Графічні інтерфейси A1.2 Git у Visual Studio A1.3 Git в Eclipse A1.4 Git у Bash A1.5 Git у Zsh A1.6 Git у Powershell A1.7 Підсумок A2. Додаток B: Вбудовування Git у ваші застосунки A2.1 Git з командного рядка A2.2 Libgit2 A2.3 JGit A2.4 go-git A3. Додаток C: Команди Git A3.1 Налаштування та конфігурація A3.2 Отримання та створення проектів A3.3 Базове збереження відбитків A3.4 Галуження та зливання A3.5 Поширення й оновлення проектів A3.6 Огляд та порівняння A3.7 Зневаджування A3.8 Латання (patching) A3.9 Електронна пошта A3.10 Зовнішні системи A3.11 Адміністрування A3.12 Кухонні команди 2nd Edition 2.5 Основи Git - Взаємодія з віддаленими сховищами Взаємодія з віддаленими сховищами Задля співпраці з будь-яким проєктом Git, вам необхідно знати, як керувати віддаленими сховищами. Віддалені сховища — це версії вашого проєкту, що розташовані в Інтернеті, або десь у мережі. Їх може бути декілька, кожне зазвичай або тільки для читання, або для читання та змін. Співпраця з іншими вимагає керування цими віддаленими сховищами, надсилання ( pushing ) та стягування ( pulling ) даних до та з них, коли ви хочете зробити внесок. Керування віддаленими сховищами потребує знань про додавання віддалених сховищ, вилучення сховищ, що більше не потрібні, керування різноманітними віддаленими гілками та визначення стежити за ними чи ні, і багато іншого. У цій секції, ми пройдемо ці вміння керування віддаленими сховищами. Зауваження Віддалені сховища можуть розташовуватися на вашій локальній машині. Цілком можливо, що ви працюватимете з віддаленим'' сховищем, що, насправді, міститься на тій саме машині, що ви за нею працюєте. Слово віддалений'' не обовʼязково означає, що сховище зберігається десь в мережі чи Інтернеті — лише що воно деінде. Взаємодія з таким віддаленим сховищем все одно включатиме звичні операції push , pull і fetch  — як і з будь-яким іншим віддаленим сховищем. Дивимось на ваші сховища Щоб побачити, які віддалені сервера ви налаштували, ви можете виконати команду git remote . Вона виводить список коротких імен кожного віддаленого сховища, яке ви задали. Якщо ви отримали своє сховище клонуванням, ви маєте побачити хоча б origin  — таке ім’я Git дає серверу, з якого ви зробили клон: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Ви також можете дати опцію -v , яка покаже вам посилання, які Git зберігає та використовує при читанні та записі до цього сховища: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Якщо у вас більш ніж одне віддалене сховище, ця команда описує їх усі. Наприклад, сховище з декількома віддаленими сховищами для роботи з багатьма співробітниками може виглядати так. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Це означає, що ми можемо отримувати ( pull ) внески з будь-якого з цих користувачів доволі легко. Ми також можемо мати дозвіл на надсилання змін до якихось з них, хоч ми й не можемо цього тут визначити. Завважте, що ці сховища використовують різноманітні протоколи. Ми більше про це поговоримо в Отримання Git на сервері . Додавання віддалених сховищ Ми згадували і навіть продемонстрували, як команда git clone неявно додає віддалене сховище origin . Тут розкажемо як додати нове віддалене сховище явно. Щоб додати нове віддалене Git сховище під заданим ім’ям, на яке ви можете легко посилатись, виконайте git remote add <ім’я> <посилання> : $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Тепер ви можете використати рядок pb в командному рядку замість повного посилання. Наприклад, якщо ви хочете здобути ( fetch ) усю інформацію, яке є в Пола, проте її нема у вашому сховищі, ви можете виконати git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Гілка master Пола тепер доступна локально як pb/master  — ви можете злити її з однією з ваших гілок, або зробити з неї локальну гілку, якщо хочете оглянути її. (Ми розповімо що таке гілки та як ними користуватися набагато докладніше в Галуження в git .) Здобуття ( fetching ) та стягування ( pulling ) з ваших віддалених сховищ Як ви щойно побачили, щоб отримати дані з ваших віддалених проєктів, ви можете виконати: $ git fetch <remote> Ця команда заходить на віддалений проєкт та забирає звідти усі дані, котрих у вас досі нема. Після цього, у вас будуть посилання на всі гілки з того сховища, які ви можете зливати або оглядати в будь-який час. Якщо ви зробили клон сховища, команда автоматично додає це віддалене сховище під ім’ям `origin''. Отже, `git fetch origin здобуває будь-яку нову працю, що її виклали на той сервер після того, як ви зробили його клон (або востаннє отримували зміни з нього). Важливо зауважити, що команда git fetch лише завантажує дані до вашого локального сховища — вона автоматично не зливає їх з вашою роботою, та не змінює вашу поточну працю. Вам буде потрібно вручну її злити, коли ви будете готові. Якщо ваша поточна гілка налаштована стежити за віддаленою гілкою (докладніше в наступній секції та Галуження в git ), ви можете виконати команду git pull щоб автоматично отримати зміни та злити віддалену гілку до вашої поточної гілки. Це може бути легшим та зручнішим методом для вас. Та команда git clone автоматично налаштовує вашу локальну гілку master стежити за віддаленою гілкою master (хоча вона може називатись і по іншому) на віддаленому сервері, з якого ви зробили клон. Виконання git pull зазвичай здобуває дані з серверу, з якого ви зробили клон, та намагається злити їх з кодом, над яким ви зараз працюєте. Зауваження Починаючи з версії Git 2.27 і вище, git pull повертатиме попередження, якщо параметр pull.rebase не встановлено. Git попереджатиме вас допоки ви не встановите значення для цього параметру. Якщо ви потребуєте типову поведінку від Git (перемотання, якщо можливе, або ж створення коміту злиття), виконайте команду: git config --global pull.rebase "false" Якщо ви бажаєте перебазовувати, коли стягуєте зміни: git config --global pull.rebase "true" Надсилання змін до ваших віддалених сховищ Коли ви довели свій проєкт до стану, що хочете ним поділитись, вам треба надіслати ( push ) ваші зміни нагору ( upstream ). Це робиться простою командою: git push <назва сховища> <назва гілки> . Якщо ви бажаєте викласти свою гілку master до вашого серверу origin (клонування зазвичай налаштовує обидва імені для вас автоматично), ви можете виконати наступне для надсилання всіх зроблених комітів до сервера: $ git push origin master Ця команда спрацює тільки в разі, якщо ви зробили клон з серверу, до якого у вас є доступ на запис, та ніхто не оновлював його після цього. Якщо хтось інший зробив клон та надіслав щось назад перед вами, вашій спробі буде слушно відмовлено. Вам доведеться спершу отримати їхню працю й вбудувати її до вашої до того, як вам дозволять надіслати свої зміни. Докладніше про надсилання змін до віддалених серверів у Галуження в git . Оглядання віддаленого сховища Якщо ви бажаєте більше дізнатись про окреме віддалене сховище, ви можете використати команду git remote show <назва сховища> . Якщо ви виконаєте цю команду з окремим ім’ям, наприклад origin , ви отримаєте щось на кшталт: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Вона виводить посилання для віддаленого сховища, а також інформацію про стеження за гілками. Команда ґречно розповідає вам, що якщо ви на гілці master та виконаєте команду git pull , вона автоматично зіллє гілку master з віддаленою після того, як отримає всі дані з віддаленого сховища. Також видано список усіх віддалених посилань, які були забрані. Ви напевно зустрінете такий простий приклад. Втім, коли ви почнете працювати з Git інтенсивніше, ви можете побачити набагато більше інформації від git remote show : $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Ця команда показує, до яких гілок автоматично надсилаються ваші зміни, коли ви виконуєте git push , доки перебуваєте на певній гілці. Вона також показує, яких віддалених гілок з серверу у вас нема, які віддалені гілки, що у вас є, були вилучені з серверу, і декілька локальних гілок, що можуть автоматично зливатися з віддаленими гілками, за якими вони стежать, коли ви виконуєте git pull . Перейменування та вилучення віддалених сховищ Ви можете виконати git remote rename , щоб перейменувати віддалене сховище. Наприклад, щоб перейменувати pb на paul , ви можете зробити це за допомогою git remote rename : $ git remote rename pb paul $ git remote origin paul Варто зазначити, що це змінює і всі назви ваших віддалених гілок. Що раніше мало назву pb/master , тепер називається paul/master . Якщо ви з якоїсь причини бажаєте вилучити віддалене сховище — ви перемістили сервер або більше не використовуєте якесь дзеркало, або, можливо, хтось припинив співпрацю — ви можете використати git remote remove або git remote rm : $ git remote remove paul $ git remote origin Як тільки ви вилучаєте посилання на віддалене сховище в такий спосіб, усі віддалені гілки та налаштування, що пов’язані із цим віддаленим сховищем, також будуть вилучені. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-Je-publieke-SSH-sleutel-genereren
Git - Je publieke SSH sleutel genereren About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.3 Git op de server - Je publieke SSH sleutel genereren Je publieke SSH sleutel genereren Er zijn vele Git servers die authenticeren met een publieke SSH sleutel. Om een publieke sleutel te kunnen aanleveren, zal iedere gebruiker in je systeem er een moeten genereren als ze er nog geen hebben. Dit proces is bij alle operating systemen vergelijkbaar. Als eerste moet je controleren of je er niet al een hebt. Standaard staan de SSH sleutels van de gebruikers in hun eigen ~/.ssh directory. Je kunt makkelijk nagaan of je al een sleutel hebt door naar die directory te gaan en de inhoud te bekijken: $ cd ~/.ssh $ ls authorized_keys2 id_dsa known_hosts config id_dsa.pub Je bent op zoek naar een aantal bestanden genaamd als id_dsa of id_rsa met een bestand met gelijke naam en een .pub extensie. Het .pub bestand is je publieke sleutel en het andere bestand is de bijbehorende private sleutel. Als je deze bestanden niet hebt (of als je zelfs geen .ssh directory hebt), dan kun je ze aanmaken door een applicatie genaamd ssh-keygen uit te voeren, deze wordt met het SSH pakket op Linux/macOS systemen meegeleverd en met het Git for Windows pakket: $ ssh-keygen -o Generating public/private rsa key pair. Enter file in which to save the key (/home/schacon/.ssh/id_rsa): Created directory '/home/schacon/.ssh'. Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/schacon/.ssh/id_rsa. Your public key has been saved in /home/schacon/.ssh/id_rsa.pub. The key fingerprint is: d0:82:24:8e:d7:f1:bb:9b:33:53:96:93:49:da:9b:e3 schacon@mylaptop.local Eerst wordt de lokatie waar je de sleutel wordt opgeslagen ( .ssh/id_rsa ) aangegeven, en vervolgens vraagt het tweemaal om een wachtwoord, die je leeg kunt laten als je geen wachtwoord wilt intypen op het moment dat je de sleutel gebruikt. Echter, als je echt een wachtwoord gebruikt, zorg ervoor dat je de -o optie toevoegt; daarmee bewaar je de private key in een formaat dat beter bestand is tegen brute-force wachtwoord kraken dan het standaard formaat. Je kunt ook de het ssh-agent tool gebruiken om te voorkomen dat je elke keer je wachtwoord moet intypen. Iedere gebruiker die dit doet, moet zijn sleutel sturen naar jou of degene die de Git server beheert (aangenomen dat je een SSH server gebruikt die publieke sleutels vereist). Het enige dat ze hoeven doen is de inhoud van het .pub bestand kopiëren en e-mailen. De publieke sleutel ziet er ongeveer zo uit: $ cat ~/.ssh/id_rsa.pub ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAklOUpkDHrfHY17SbrmTIpNLTGK9Tjom/BWDSU GPl+nafzlHDTYW7hdI4yZ5ew18JH4JW9jbhUFrviQzM7xlELEVf4h9lFX5QVkbPppSwg0cda3 Pbv7kOdJ/MTyBlWXFCR+HAo3FXRitBqxiX1nKhXpHAZsMciLq8V6RjsNAQwdsdMFvSlVK/7XA t3FaoJoAsncM1Q9x5+3V0Ww68/eIFmb1zuUFljQJKprrX88XypNDvjYNby6vw/Pb0rwert/En mZ+AW4OZPnTPI89ZPmVMLuayrD2cE86Z/il8b+gw3r3+1nKatmIkjn2so1d01QraTlMqVSsbx NrRFi9wrf+M7Q== schacon@mylaptop.local Voor een uitgebreidere tutorial over het aanmaken van een SSH sleutel op meerdere operating systemen, verwijzen we je naar de GitHub handleiding over SSH sleutels op https://help.github.com/articles/generating-ssh-keys . prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/uk/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d0%b8-Git-%d0%a2%d0%b5%d2%91%d1%83%d0%b2%d0%b0%d0%bd%d0%bd%d1%8f
Git - Теґування About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Вступ 1.1 Про систему контролю версій 1.2 Коротка історія Git 1.3 Основи Git 1.4 Git, зазвичай, тільки додає дані 1.5 Три стани 1.6 Командний рядок 1.7 Інсталяція Git 1.8 Початкове налаштування Git 1.9 Отримання допомоги 1.10 Підсумок 2. Основи Git 2.1 Створення Git-сховища 2.2 Запис змін до репозиторія 2.3 Перегляд історії комітів 2.4 Скасування речей 2.5 Взаємодія з віддаленими сховищами 2.6 Теґування 2.7 Псевдоніми Git 2.8 Підсумок 3. Галуження в git 3.1 Гілки у кількох словах 3.2 Основи галуження та зливання 3.3 Управління гілками 3.4 Процеси роботи з гілками 3.5 Віддалені гілки 3.6 Перебазовування 3.7 Підсумок 4. Git на сервері 4.1 Протоколи 4.2 Отримання Git на сервері 4.3 Генерація вашого публічного ключа SSH 4.4 Налаштування Серверу 4.5 Демон Git 4.6 Розумний HTTP 4.7 GitWeb 4.8 GitLab 4.9 Варіанти стороннього хостингу 4.10 Підсумок 5. Розподілений Git 5.1 Розподілені процеси роботи 5.2 Внесення змін до проекту 5.3 Супроводжування проекту 5.4 Підсумок 6. GitHub 6.1 Створення та налаштування облікового запису 6.2 Як зробити внесок до проекту 6.3 Супроводжування проєкту 6.4 Керування організацією 6.5 Скриптування GitHub 6.6 Підсумок 7. Інструменти Git 7.1 Вибір ревізій 7.2 Інтерактивне індексування 7.3 Ховання та чищення 7.4 Підписання праці 7.5 Пошук 7.6 Переписування історії 7.7 Усвідомлення скидання (reset) 7.8 Складне злиття 7.9 Rerere 7.10 Зневадження з Git 7.11 Підмодулі 7.12 Пакування 7.13 Заміна 7.14 Збереження посвідчення (credential) 7.15 Підсумок 8. Налаштування Git 8.1 Конфігурація Git 8.2 Атрибути Git 8.3 Гаки (hooks) Git 8.4 Приклад політики користування виконуваної Git-ом 8.5 Підсумок 9. Git and Other Systems 9.1 Git як клієнт 9.2 Міграція на Git 9.3 Підсумок 10. Git зсередини 10.1 Кухонні та парадні команди 10.2 Об’єкти Git 10.3 Посилання Git 10.4 Файли пакунки 10.5 Специфікація посилань (refspec) 10.6 Протоколи передачі 10.7 Супроводження та відновлення даних 10.8 Змінні середовища 10.9 Підсумок A1. Додаток A: Git в інших середовищах A1.1 Графічні інтерфейси A1.2 Git у Visual Studio A1.3 Git в Eclipse A1.4 Git у Bash A1.5 Git у Zsh A1.6 Git у Powershell A1.7 Підсумок A2. Додаток B: Вбудовування Git у ваші застосунки A2.1 Git з командного рядка A2.2 Libgit2 A2.3 JGit A2.4 go-git A3. Додаток C: Команди Git A3.1 Налаштування та конфігурація A3.2 Отримання та створення проектів A3.3 Базове збереження відбитків A3.4 Галуження та зливання A3.5 Поширення й оновлення проектів A3.6 Огляд та порівняння A3.7 Зневаджування A3.8 Латання (patching) A3.9 Електронна пошта A3.10 Зовнішні системи A3.11 Адміністрування A3.12 Кухонні команди 2nd Edition 2.6 Основи Git - Теґування Теґування Як і більшість СКВ, Git дозволяє поставити теґ на окремому моменті історії, що чимось видатний. Зазвичай ця функціональність використовується щоб позначити релізи (v1.0 тощо). У цій секції, ви дізнаєтесь, як отримати список доступних теґів, як створювати нові теґи, та які типи теґів існують. Показати ваші теґи Отримати список доступних теґів у Git елементарно. Просто наберіть git tag (з опціональним -l чи --list ): $ git tag v0.1 v1.3 Ця команда виводить список теґів в алфавітному порядку. Цей порядок насправді неважливий. Ви також можете шукати теґи, що відповідають певному шаблону. Наприклад, сховище Git містить більш ніж 500 теґів. Якщо вас цікавлять виключно версії 1.8.5, ви можете виконати: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Зауваження Пошук за шаблонами із спеціальними символами потребує опції -l чи --list Якщо ви хочете отримати повний список теґів, то команда git tag без додаткових аргументів виконує саме це. Використання -l чи --list опціонально. Втім, якщо ви використовуєте шаблон зі спеціальними символами, то використання -l чи --list є обовʼязковим. Створення теґів Git підтримує два головних типи теґів: легкі та анотовані . Легкий теґ дуже схожий на гілку, що не змінюється — це просто вказівник на певний коміт. Анотовані теґи Створити анотований теґ у Git просто. Найлегший спосіб — додати -a до команди tag : $ git tag -a v1.4 -m "моя версія 1.4" $ git tag v0.1 v1.3 v1.4 -m визначає повідомлення теґу, що в ній буде збережено. Якщо ви не вкажете повідомлення анотованого теґу, Git запустить ваш редактор щоб ви могли його набрати. Ви можете побачити дані теґу та коміт, на який він вказує, за допомогою команди git show : $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Це показує інформацію про автора теґу, дату створення теґу, та повідомлення перед інформацією про коміт. Легкі теґи Другий спосіб позначати коміти — за допомогою легких позначок. Це просто хеш коміту збережений у файлі — ніякої іншої інформації не зберігається. Щоб створити легкий теґ, не додавайте жодної з опцій -a , -s та -m , вкажіть лише назву теґу: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Цього разу, якщо ви виконаєте git show з теґом, ви не побачите додаткової інформації про теґ. Команда покаже тільки коміт: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Створення теґів пізніше Ви також можете зробити теґ до комітів, від котрих ви вже пішли. Припустімо, що ваша історія комітів виглядає так: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 beginning write support 0d52aaab4479697da7686c15f77a3d64d9165190 one more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc added a commit function 4682c3261057305bdd616e23b64b0857d832627b added a todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a started write support 9fceb02d0ae598e95dc970b74767f19372d61af8 updated rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a updated readme Тепер, припустимо ви забули створити теґ до версії проекту v1.2, що має бути на коміті ``updated rakefile''. Ви можете додати теґ і зараз. Щоб створити теґ до коміту, вам треба дописати суму коміту (чи її частину) наприкінці команди: $ git tag -a v1.2 9fceb02 Ви можете побачити, що ви створили теґ до коміту: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... Розповсюдження теґів Без додаткових опцій команда git push не передає теґи на віддалені сервери. Вам доведеться явно надсилати теґи на спільний сервер після створення. Цей процес не відрізняється від розповсюдження віддалених гілок — вам треба виконати git push origin <назва теґу> . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Якщо у вас багато теґів, та ви хочете надіслати їх разом, ви також можете використати опцію --tags команди git push . Це передасть усі ваші теґи до віддаленого серверу, яких там досі нема. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Тепер, коли хтось інший зробить клон або отримає зміни з вашого сховища, він отримає також усі ваші теґи. Переключення до теґів Якщо ви бажаєте переглянути версії файлів, на які вказує теґ, виконайте git checkout , хоча тоді ви отримаєте стан ``відокремлений HEAD'' (detached HEAD), що має декілька обмежень: $ git checkout 2.0.0 Note: checking out '2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch> HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout 2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... add atlas.json and cover image Якщо ви щось зміните й створите коміт у стані ``відокремлений HEAD'', теґ залишиться незмінним, а а новий коміт не належатиме жодній гілці й буде досяжним лише за допомогою хеша коміту. Відповідно, якщо вам треба здійснити зміни — скажімо, ви виправляєте ваду у старшій версії — зазвичай варто створити гілку: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Якщшо ви це зробите й створите коміт, ваша гілка version2 буде трохи відрізнятися від вашого теґу v2.0.0 , адже вона переміститься вперед до ваших нових змін, отже будьте обережні. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-arbor-edge-defense/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_full-click
Arbor Edge Defense | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Arbor Edge Defense DDoS Protection Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Arbor Edge Defense is an inline security appliance deployed at the network perimeter that can automatically detect and block inbound threats and outbound malicious communication using highly scalable, stateless technology and unique, global threat intelligence. This product is intended for Cyber Security Engineer Chief Executive Officer Chief Information Officer Network Operations Center Head of Security Security Engineer Director of Security Information Technology Specialist Cyber Security Specialist Media Products media viewer No more previous content Smart Perimeter Protection With NETSCOUT Arbor Edge Defense (AED) NETSCOUT Arbor Edge Defense (AED) blocks inbound threats such as DDoS attacks, and outbound communication from compromised internal hosts - acting as a first and last line of smart, automated, perimeter defense. Perimeter Defense Best Practices Using Arbor Edge Defense NETSCOUT Arbor Edge Defense acts as a first and last line of smart, automated perimeter defense for an organization. It’s deployed on-premise, where it acts as a first line of defense, by blocking inbound DDoS attacks, protecting stateful security devices. Demo: Blocking Ransomware Attack with Arbor Edge Defense Acting as a last line of defense, AED detects and blocks outbound indicators of compromise (or IoCs) that has been missed by other tools in your security stack to stop the proliferation of malware before a data breach. NETSCOUT’s Arbor Edge Defense: The First and Last Line of Defense Adam Bixler, Director, Product Management at NETSCOUT, discusses NETSCOUT AED’s unique functionality and how it augments and strengthens traditional endpoint security. Learn about the value of integrating threat intelligence and DDoS defense as a first and last line of defense. Visit the product page for more information: http://www.netscout.link/6001EPXxb Get VPN Protection with NETSCOUT Arbor Edge Defense A DDoS attack poses a major threat to the availability of the VPN gateway. As employees continue to work from home, protecting the availability of your VPN gateway from DDoS attacks is critical. Unlike a cloud-based DDoS protection solution, NETSCOUT Arbor Edge Defense is a stateless, on-premise solution that can instantaneously detect and mitigate DDoS attacks against the VPN gateway, enabling you to maintain home-based employee productivity and business continuity. No more next content Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less NETSCOUT products Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/az/v2/GitHub-Q%c4%b1sa-M%c9%99zmun
Git - Qısa Məzmun About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlanğıc 1.1 Versiyaya Nəzarət Haqqında 1.2 Git’in Qısa Hekayəsi 1.3 Git Nədir? 1.4 Əmr Sətiri 1.5 Git’i Quraşdırmaq 1.6 İlk Dəfə Git Quraşdırması 1.7 Kömək Almaq 1.8 Qısa Məzmun 2. Git’in Əsasları 2.1 Git Deposunun Əldə Edilməsi 2.2 Depoda Dəyişikliklərin Qeyd Edilməsi 2.3 Commit Tarixçəsinə Baxış 2.4 Ləğv Edilən İşlər (Geri qaytarılan) 2.5 Uzaqdan İşləmək 2.6 Etiketləmə 2.7 Git Alias’lar 2.8 Qısa Məzmun 3. Git’də Branch 3.1 Nutshell’də Branch’lar 3.2 Sadə Branching və Birləşdirmə 3.3 Branch İdarəedilməsi 3.4 Branching İş Axınları 3.5 Uzaq Branch’lar 3.6 Rebasing 3.7 Qısa Məzmun 4. Server’də Git 4.1 Protokollar 4.2 Serverdə Git Əldə Etmək 4.3 Sizin öz SSH Public Key’nizi yaratmaq 4.4 Server qurmaq 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Tərəf Seçimləri 4.10 Qısa Məzmun 5. Paylanmış Git 5.1 Distribyutorluq İş Axınları 5.2 Layihəyə Töhfə vermək 5.3 Layihənin Saxlanılması 5.4 Qısa Məzmun 6. GitHub 6.1 Hesab Qurma və Konfiqurasiya 6.2 Bir Layihəyə Töhfə Vermək 6.3 Bir Layihənin Saxlanılması 6.4 Bir Təşkilatı Idarə Etmək 6.5 GitHub Skriptləmə 6.6 Qısa Məzmun 7. Git Alətləri 7.1 Reviziya Seçimi 7.2 Interaktiv Səhnələşdirmə 7.3 Stashing və Təmizləmə 7.4 İşinizin İmzalanması 7.5 Axtarış 7.6 Tarixi Yenidən Yazmaq 7.7 Reset Demystified 7.8 İnkişaf etmiş Birləşmə 7.9 Rerere 7.10 Git ilə Debugging 7.11 Alt Modullar 7.12 Bundling 7.13 Dəyişdirmək 7.14 Etibarlı Yaddaş 7.15 Qısa Məzmun 8. Git’i Fərdiləşdirmək 8.1 Git Konfiqurasiyası 8.2 Git Atributları 8.3 Git Hook’ları 8.4 Git-Enforced Siyasət Nümunəsi 8.5 Qısa Məzmun 9. Git və Digər Sistemlər 9.1 Git Müştəri kimi 9.2 Git’ə Miqrasiya 9.3 Qısa Məzmun 10. Git’in Daxili İşləri 10.1 Plumbing və Porcelain 10.2 Git Obyektləri 10.3 Git Referansları 10.4 Packfile’lar 10.5 Refspec 10.6 Transfer Protokolları 10.7 Maintenance və Məlumatların Bərpası 10.8 Mühit Dəyişənləri 10.9 Qısa Məzmun A1. Appendix A: Digər Mühitlərdə Git A1.1 Qrafik interfeyslər A1.2 Visual Studio’da Git A1.3 Visual Studio Code’da Git A1.4 Eclipse’də Git A1.5 Sublime Text’də Git A1.6 Bash’da Git A1.7 Zsh’də Git A1.8 PowerShell’də Git A1.9 Qısa Məzmun A2. Appendix B: Proqramlara Git Daxil Etmək A2.1 Əmr-sətri Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Əmrləri A3.1 Quraşdırma və Konfiqurasiya A3.2 Layihələrin Alınması və Yaradılması A3.3 Sadə Snapshotting A3.4 Branching və Birləşmə A3.5 Layihələrin Paylaşılması və Yenilənməsi A3.6 Yoxlama və Müqayisə A3.7 Debugging A3.8 Patching A3.9 E-poçt A3.10 Xarici Sistemlər A3.11 İdarəetmə A3.12 Plumbing Əmrləri 2nd Edition 6.6 GitHub - Qısa Məzmun Qısa Məzmun İndi bir GitHub istifadəçisisiniz. Bir hesab yaratmağı, bir təşkilatı necə idarə etməyi, depo’lar yaratmağı və push etməyi, başqalarının layihələrinə töhfə verməyi və başqalarının töhfələrini qəbul etməyi bilirsiniz. Növbəti fəsildə, sizi həqiqətən Git ustası halına gətirəcək mürəkkəb vəziyyətlərlə məşğul olmaq üçün daha güclü alətlər və tövsiyələr öyrənəcəksiniz. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/hyroai-adaptive-communications-platform/?trk=products_seo_search
Responsible AI Agents Platform | LinkedIn Skip to main content LinkedIn Hyro in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Responsible AI Agents Platform Live Chat Software by Hyro See who's skilled in this Add as skill Get started Report this product About Omni-channel platform, featuring call centers, websites, SMS and mobile applications, that automates interactions using voice and text AI. 1. Ingest - Hyro seamlessly ingests data from sources such as webpages, CSVs, databases, APIs and more into their knowledge graph 2. Converse - Hyro pumps those relationships and attributes through their natural language engine, so that all conversations now have context and a high understanding rate 3. Inform - Hyro provides valuable insights that reveal trends, keywords and previously unknown issues to help teams optimize their digital channels This product is intended for Chief Information Officer Chief Digital Officer Chief Marketing Officer Chief Technology Officer Vice President Operations Chief Operating Officer Director of Information Technology Director of Digital Services Head of Digital Director of Web Services Media Products media viewer No more previous content Adaptive Communications for Real Estate 1. Ingest Information 2. Make it Conversational 3. Generate Insights No more next content Featured customers of Responsible AI Agents Platform Roper St. Francis Healthcare Hospitals and Health Care 20,641 followers Weill Cornell Medicine Hospitals and Health Care 136,102 followers Hartford HealthCare Hospitals and Health Care 56,117 followers Hackensack Meridian Health Hospitals and Health Care 113,076 followers Inova Health Hospitals and Health Care 129,024 followers Baptist Health Hospitals and Health Care 52,393 followers Summa Health Hospitals and Health Care 43,453 followers Emory Healthcare Hospitals and Health Care 124,375 followers Montefiore Health System Hospitals and Health Care 89,063 followers Bon Secours Mercy Health Hospitals and Health Care 45,259 followers Tampa General Hospital Hospitals and Health Care 77,291 followers Piedmont Hospitals and Health Care 122,976 followers Intermountain Health Hospitals and Health Care 133,460 followers Sutter Health Hospitals and Health Care 213,645 followers Show more Show less Similar products Freshchat Freshchat Live Chat Software ChatR ChatR Live Chat Software ICE Chat ICE Chat Live Chat Software Webchat Webchat Live Chat Software Microsoft Teams Microsoft Teams Live Chat Software LiveChat LiveChat Live Chat Software Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Branchen-in-Git-Branch-workflows
Git - Branch workflows About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 3.4 Branchen in Git - Branch workflows Branch workflows Nu je de basis van het branchen en mergen onder de knie hebt, wat kan je of zou je daarmee kunnen doen? In deze deel gaan we een aantal veel voorkomende workflows die deze lichtgewicht branches mogelijk maken behandelen, zodat je kunt besluiten of je ze wilt toepassen in je eigen ontwikkelcyclus. Langlopende branches branches, long-running) Omdat Git gebruik maakt van een eenvoudige drieweg-merge, is het meerdere keren mergen vanuit een branch in een andere gedurende een langere periode over het algemeen eenvoudig te doen. Dit houdt in dat je meerdere branches kunt hebben, die altijd open staan en die je voor verschillende fases van je ontwikkelcyclus gebruikt; je kunt regelmatig vanuit een aantal mergen in andere. Veel Git-ontwikkelaars hebben een workflow die deze aanpak omarmt, zoals het hebben van alleen volledig stabiele code in hun master -branch — mogelijk alleen code die is of zal worden vrijgegeven. Ze hebben een andere parallelle branch develop of next genaamd waarop ze werken of die ze gebruiken om stabiliteit te testen — het is niet noodzakelijkerwijs altijd stabiel, maar zodra het in een stabiele status verkeert, kan het worden gemerged in master . Deze wordt gebruikt om topic branches (branches met een korte levensduur, zoals jou eerdere iss53 -branch) te pullen zodra die klaar zijn, om zich ervan te overtuigen dat alle tests slagen en er geen fouten worden geïntroduceerd. Feitelijk praten we over verwijzingen die worden verplaatst over de lijn van de commits die je maakt. De stabiele branches zijn verder stroomafwaarts in je commit-historie, en de splinternieuwe branches zijn verder naar voren in de historie. Figuur 26. Een lineare kijk op progressief-stabiliteits branchen Ze zijn misschien makkelijker voor te stellen als silo’s, waar sets van commits stapsgewijs naar een meer stabiele silo worden gepromoveerd als ze volledig getest zijn. Figuur 27. Een “silo” kijk op progressief-stabiliteits branchen Je kunt dit blijven doen voor elk niveau van stabiliteit. Sommige grotere projecten hebben ook een proposed of pu (proposed updates) branch die branches geïntegreerd heeft die wellicht nog niet klaar zijn om in de next of master -branch te gaan. Het idee erachter is dat de branches op verschillende niveaus van stabiliteit zitten. Zodra ze een stabieler niveau bereiken, worden ze in de branch boven hen gemerged. Nogmaals, het hebben van meerdere langlopende branches is niet noodzakelijk, maar het helpt vaak wel; in het bijzonder als je te maken hebt met zeer grote of complexe projecten. Topic branches Topic branches zijn nuttig in projecten van elke grootte. Een topic branch is een kortlopende branch die je maakt en gebruikt om een specifieke functie te realiseren of daaraan gerelateerd werk te doen. Dit is iets wat je waarschijnlijk nooit eerder met een VCS gedaan hebt, omdat het over het algemeen te duur is om branches aan te maken en te mergen. Maar in Git is het niet ongebruikelijk om meerdere keren per dag branches aan te maken, daarop te werken, en ze te verwijderen. Je zag dit in de vorige paragraaf met de iss53 en hotfix -branches die je gemaakt had. Je hebt een aantal commits op ze gedaan en ze meteen verwijderd nadat je ze gemerged had in je hoofd-branch. Deze techniek stelt je in staat om snel en volledig van context te veranderen — omdat je werk is onderverdeeld in silo’s waar alle wijzigingen in die branch te maken hebben met dat onderwerp, is het makkelijker te zien wat er is gebeurd tijdens een code review en dergelijke. Je kunt de wijzigingen daar minuten-, dagen- of maandenlang bewaren, en ze mergen als ze er klaar voor zijn, ongeacht de volgorde waarin ze gemaakt zijn of er aan gewerkt is. Neem als voorbeeld een situatie waarbij wat werk gedaan wordt (op master ), er wordt een branche gemaakt voor een probleem ( iss91 ) en daar wordt wat aan gewerkt, er wordt een tweede branch gemaakt om op een andere manier te proberen hetzelfde op te lossen ( iss91v2 ); weer even wordt teruggegaan naar de master branch om daar een tijdje te werken, en dan vanaf daar wordt gebrancht om wat werk te doen waarvan je niet zeker weet of het wel zo’n slim idee is ( dumbidea -branch). Je commit-historie zal eruitzien als volgt: Figuur 28. Meerdere topic branches Laten we zeggen dat je besluit dat je de tweede oplossing voor je probleem het beste vindt ( iss91v2 ), en je hebt de dumbidea -branch aan je collega’s laten zien en het blijkt geniaal te zijn. Je kunt dan de oorspronkelijke iss91 weggooien (waardoor je commits C5 en C6 kwijt raakt), en de andere twee mergen. Je historie ziet er dan uit volgt: Figuur 29. Historie na het mergen van dumbidea en iss91v2 We zullen in meer detail behandelen wat de verschillende mogelijke workflows zijn voor jouw Git project in Gedistribueerd Git , dus voordat je besluit welk branching schema je voor jouw volgende project wilt gebruiken, zorg dat je dat hoofdstuk gelezen hebt. Het is belangrijk om te beseffen dat tijdens al deze handelingen, al deze branches volledig lokaal zijn. Als je aan het branchen of mergen bent, dan wordt alles alleen in jouw Git repository gedaan — dus er vindt geen server communicatie plaats. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/ru/v2/%d0%92%d0%b2%d0%b5%d0%b4%d0%b5%d0%bd%d0%b8%d0%b5-%d0%9a%d0%be%d0%bc%d0%b0%d0%bd%d0%b4%d0%bd%d0%b0%d1%8f-%d1%81%d1%82%d1%80%d0%be%d0%ba%d0%b0
Git - Командная строка About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Введение 1.1 О системе контроля версий 1.2 Краткая история Git 1.3 Что такое Git? 1.4 Командная строка 1.5 Установка Git 1.6 Первоначальная настройка Git 1.7 Как получить помощь? 1.8 Заключение 2. Основы Git 2.1 Создание Git-репозитория 2.2 Запись изменений в репозиторий 2.3 Просмотр истории коммитов 2.4 Операции отмены 2.5 Работа с удалёнными репозиториями 2.6 Работа с тегами 2.7 Псевдонимы в Git 2.8 Заключение 3. Ветвление в Git 3.1 О ветвлении в двух словах 3.2 Основы ветвления и слияния 3.3 Управление ветками 3.4 Работа с ветками 3.5 Удалённые ветки 3.6 Перебазирование 3.7 Заключение 4. Git на сервере 4.1 Протоколы 4.2 Установка Git на сервер 4.3 Генерация открытого SSH ключа 4.4 Настраиваем сервер 4.5 Git-демон 4.6 Умный HTTP 4.7 GitWeb 4.8 GitLab 4.9 Git-хостинг 4.10 Заключение 5. Распределённый Git 5.1 Распределённый рабочий процесс 5.2 Участие в проекте 5.3 Сопровождение проекта 5.4 Заключение 6. GitHub 6.1 Настройка и конфигурация учётной записи 6.2 Внесение собственного вклада в проекты 6.3 Сопровождение проекта 6.4 Управление организацией 6.5 Создание сценариев GitHub 6.6 Заключение 7. Инструменты Git 7.1 Выбор ревизии 7.2 Интерактивное индексирование 7.3 Припрятывание и очистка 7.4 Подпись 7.5 Поиск 7.6 Перезапись истории 7.7 Раскрытие тайн reset 7.8 Продвинутое слияние 7.9 Rerere 7.10 Обнаружение ошибок с помощью Git 7.11 Подмодули 7.12 Создание пакетов 7.13 Замена 7.14 Хранилище учётных данных 7.15 Заключение 8. Настройка Git 8.1 Конфигурация Git 8.2 Атрибуты Git 8.3 Хуки в Git 8.4 Пример принудительной политики Git 8.5 Заключение 9. Git и другие системы контроля версий 9.1 Git как клиент 9.2 Переход на Git 9.3 Заключение 10. Git изнутри 10.1 Сантехника и Фарфор 10.2 Объекты Git 10.3 Ссылки в Git 10.4 Pack-файлы 10.5 Спецификации ссылок 10.6 Протоколы передачи данных 10.7 Обслуживание репозитория и восстановление данных 10.8 Переменные окружения 10.9 Заключение A1. Приложение A: Git в других окружениях A1.1 Графические интерфейсы A1.2 Git в Visual Studio A1.3 Git в Visual Studio Code A1.4 Git в Eclipse A1.5 Git в IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.6 Git в Sublime Text A1.7 Git в Bash A1.8 Git в Zsh A1.9 Git в PowerShell A1.10 Заключение A2. Приложение B: Встраивание Git в ваши приложения A2.1 Git из командной строки A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Приложение C: Команды Git A3.1 Настройка и конфигурация A3.2 Клонирование и создание репозиториев A3.3 Основные команды A3.4 Ветвление и слияния A3.5 Совместная работа и обновление проектов A3.6 Осмотр и сравнение A3.7 Отладка A3.8 Внесение исправлений A3.9 Работа с помощью электронной почты A3.10 Внешние системы A3.11 Администрирование A3.12 Низкоуровневые команды 2nd Edition 1.4 Введение - Командная строка Командная строка Есть много различных способов использования Git. Помимо оригинального клиента, имеющего интерфейс командной строки, существует множество клиентов с графическим пользовательским интерфейсом, в той или иной степени реализующих функциональность Git. В рамках данной книги мы будем использовать Git в командной строке. С одной стороны, командная строка — это единственное место, где вы можете запустить все команды Git, так как большинство клиентов с графическим интерфейсом реализуют для простоты только некоторую часть функциональности Git. Если вы знаете, как выполнить какое-либо действие в командной строке, вы, вероятно, сможете выяснить, как то же самое сделать и в GUI-версии, а вот обратное не всегда верно. Кроме того, в то время, как выбор графического клиента — это дело личного вкуса, инструменты командной строки доступны всем пользователям сразу после установки Git. Поэтому мы предполагаем, что вы знаете, как открыть терминал в Mac или командную строку, или PowerShell в Windows. Если вам не понятно, о чем мы здесь говорим, то вам, возможно, придётся ненадолго прерваться и изучить эти вопросы, чтобы вы могли понимать примеры и пояснения из этой книги. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/sv/v2/Git-p%c3%a5-servern-GitLab
Git - GitLab About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Kom igång 1.1 Om versionshantering 1.2 En kort historik av Git 1.3 Vad är Git? 1.4 Kommandoraden 1.5 Installera Git 1.6 Använda Git för första gången 1.7 Få hjälp 1.8 Sammanfattning 2. Grunder i Git 2.1 Skaffa ett Git-förvar 2.2 Spara ändringar till förvaret 2.3 Visa historiken 2.4 Ångra saker 2.5 Jobba med fjärrförvar 2.6 Taggning 2.7 Git alias 2.8 Sammanfattning 3. Git förgreningar 3.1 Grenar i ett nötskal 3.2 Grundläggande förgrening och sammanslagning 3.3 Hantera grenar 3.4 Arbetsflöde med grenar 3.5 Fjärrgrenar 3.6 Grenflytt 3.7 Sammanfattning 4. Git på servern 4.1 Protokollen 4.2 Skaffa Git på en server 4.3 Generera din publika SSH-nyckel 4.4 Konvigurera servern 4.5 Git Daemonen 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Alternativ tillhandahållna av tredje part 4.10 Sammanfattning 5. Distribuerade Git 5.1 Distribuerade arbetsflöden 5.2 Medverka i ett projekt 5.3 Underhålla ett projekt 5.4 Sammanfattning 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Bilaga A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in PowerShell A1.7 Summary A2. Bilaga B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bilaga C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.8 Git på servern - GitLab GitLab GitWeb är dock ganska primitivt. Om du är ute efter en mer modern, fullutrustad Gitserver finns det några öppen källkods-alternativ där ute som du kan installera istället. Eftersom GitLab är en av de mer populära, behandlar vi installation och användning av det som exempel. Detta är lite mer komplicerat än GitWeb-alternativet och kräver troligtvis mer underhåll, men det är ett mycket mer utrustat alternativ. Installation GitLab är en databasbackad webbapplikation, så dess installation är lite mer krävande än andra Gitservrar. Lyckligtvis är denna processen väldigt väldokumenterad och underhållen. Det finns ett antal vägar du kan ta för att installera GitLab. För att få igång något snabbt, kan du ladda ner en virtuell maskin-avbild eller ett en-klicks-installationsprogram från https://bitnami.com/stack/gitlab , och modifiera konfigurationen för att passa din miljö. Ett trevligt som Bitnami har inkluderat är inloggningsskärmen (som man kommer åt med alt+→); den ger dig IP-adressen och standardanvändarnamnet och lösenordet för det installerade GitLab. Figur 50. Inloggningsskärmen till Bitnamis virtuella maskin. För allt annat, följ guiden i Readme för GitLab Community Edition, vilken du finner på https://gitlab.com/gitlab-org/gitlab-ce/tree/master . Där hittar du också hjälp för att installera GitLab via Chef-recept, en virtuell maskin på Digital Ocean, och RPM- samt DEB-paket (vilket, vid tiden för författandet, är i betaversion). Det finns också “inofficiella” guider för hur man får igång GitLab med ovanliga operativsystem och databaser, ett fullt manuellt installationsskript, och många andra saker. Administration GitLabs administrationsgränssnitt nås över webben. Via din webbläsare, gå till datornamnet eller IP-adressen för den dator där GitLab är installerat och logga in som en admin-användare. Standardanvändarnamnet är admin@local.host och standardlösenordet är 5iveL!fe (som du kommer bli påbjuden att ändra så fort du skriver in det). När du väl loggat in, klicka på menyalternativet “Admin area” i menyn uppe till höger. Figur 51. Menyalternativet “Admin area” i GitLab menyn. Användare Användare i GitLab är konton som motsvarar personer. Användarkonton är inte så komplicerade; i huvudsak är det en samling personlig information bunden till inloggningsdata. Varje användarkonto har en namnrymd , som är en logisk gruppering av projekt som tillhör den användaren. Om användaren jane har ett projekt som heter project så är det projektets URL http://server/jane/project . Figur 52. Användaradministrationsskärmen i GitLab. Man kan ta bort en användare på två sätt. “Blockering” av en användare förhindrar att de loggar in på GitLabinstansen, men all data under den användarens namnrymd finns bevarad, och versioner som är signerade med den användarens epostadress länkar fortfarande till deras profil. “Radering” av en användare å andra sidan, tar bort dem från både databasen och filsystemet. Alla projekt och data i deras namnrymd tas bort, och alla grupper de äger kommer också att tas bort. Detta är således en mycket mer permanent och destruktiv handling, och används därför sällan. Grupper En GitLab-grupp är en samling av projekt tillsammans med data om hur användare kan nå de projekten. Varje grupp har en projektnamnrymd (på samma sätt som användare), så om gruppen training har ett projekt materials , kommer dess URL att vara http://server/training/materials . Figur 53. Gruppadministrationsskärmen i GitLab. Varje grupp är associerad med ett antal användare, som var och en har en nivå av rättigheter för gruppens projekt och gruppen i sig. Dessa spänner från “Guest” (enbart arbetspaket och chat) till “Owner” (full kontroll över gruppen, dess medlemmar och dess projekt). De olika rättigheterna är för många för att lista här, men GitLab har en hjälpsam länk på administrationsskärmen. Projekt Ett GitLabprojekt motsvarar grovt mot ett enskilt Gitrepo. Varje projekt tillhör en enda namnrymd, antingen en användare eller en grupp. Om projektet tillhör en en användare, har ägaren av projektet direkt kontroll över vem som har åtkomst till projektet; om projektet tillhör en grupp, spelar även gruppens användarnivåbehörigheter roll. Varje projekt har en synlighetsnivå, som kontrollerar vem som har läsåtkomst till projektets sidor och repo. Om ett projekt är Private , måste projektets ägare explicit ge åtkomst till specifika användare. Ett projekt som är Internal är synligt för alla inloggade anbvändare, medan projekt som är Public är synliga för alla. Notera att denna kontrollerar både git fetch -åtkomst såväl som åtkomst till webbgränssnittet för det projektet. Krokar GitLab har stöd för krokar både på projekt- och systemnivå. Oavsett vilken kommer GitLabservern att utföra ett HTTP POST-anrop med en beskrivande JSON när en relevant händelse inträffar. Detta är ett förträffligt sätt att koppla dina Gitrepon och GitLabinstansen till resten av din utvecklingsautomation, såsom CI-servrar, chatrum eller distributionsverktyg. Grundläggande användning Det första du kommer vilja göra med GitLab är att skapa ett nytt projekt. Detta gör du genom att klicka på ikonen “+” i verktygsfältet. Du kommer bli ombedd att ange projektets namn, vilken namnrymd det skall tillhöra och vad dess synlighetsnivå skall vara. Det mesta du anger här är inte permanent och kan justeras senare genom inställningsgränssnittet. Klicka på “Create Project”, och sedan är du klar. Så fort projektet finns så vill du säkert ansluta det med ett lokalt Gitrepo. Varje projekt är nåbart över HTTPS eler SSH, som båda kan användas för att konfigurera ett fjärrepo. URL:erna är synliga vid toppen av projektets hemsida. För ett existerande lokalt repo, kommer följande kommando att skapa ett repo benämnt gitlab till värdplatsen: $ git remote add gitlab https://server/namespace/project.git Om du inte har en lokal kopia av repot, kan du helt enkelt göra såhär: $ git clone https://server/namespace/project.git Webbgränssnittet tillhandahåller åtkomst till flera användbara vyer av repot självt. På varje projekts hemsida visas senaste aktivitet, och länkar i toppen leder dig till vyer över peojektets filer och versionslogg. Arbeta tillsammans Det enklaste sättet att arbeta tillsammans på ett GitLabprojekt är genom att ge en annan användare direkt skrivrättighet till Gitrepot. Du kan lägga till en användare till ett projekt genom att gå till delen “Members” i projektet inställningar och associera den nya användaren med enn åtkomstnivå (skillnaden mellan åtkomstnivåer diskuteras i Grupper ). Genom att ge en användare åtkomstnivån “Developer” eller över, kan den användaren villkorslöst skicka upp versioner och grenar direkt till repot. Ett annat, mer frikopplat sätt att samarbeta är att använda sammanslagningsbegäran. Denna funktion ger vilken användare som helst som kan se projektet möjlighet att bidra till det på ett kontrollerat sätt. Användare med direkt åtkomst kan helt enkelt skapa en gren, skicka versioner till den och öppna en sammanslagningsbegäran från sin gren tillbaks in till master eller någon annan gren. Användare som inte har skrivrättigheter fölr ett repo kan “klyva” repot (skapa sin egen kopia), skicka versioner till den kopian, och sedan öppna en sammanslagningsbegäran från deran gren tillbaks in till huvudprojektet. Denna modellen ger ägaren möjlighet att ständight ha full kontroll av vad som kommer in i repot och när, samtidigt som man tillåter bidrag från opålitliga användare. Sammanslagningsbegäran och felrapporter är huvuddelarna i långlivade diskussioner i GitLab. Varje sammanslagningsbegäran tillåter rad-för-rad-diskussion av den föreslagna ändringen (som stödjer en enkel form av kodgranskning), så väl som generell översiktlig disussion. Båda kan tilldelas användare, eller organiseras i milstolpar. Detta avsnitt fokuserade huvudsakligen på Git-relaterade funktioner av GitLab, men som ett moget projekt har den många andra verktyg för att hjälpa ditt team att jobba tillsammans såsom projektwikisidor och sytemunderhållningsverktyg. En fördel med GitLa är att när väl Servern är uppe och kör, behöver du sällan modifiera en konfigurationsfil eller läsrättigheter till servern via SSH; den mesta administrationen och generell användning kan ske via webbläsarfönstret. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://materiais.resultadosdigitais.com.br/novo-bate-papo-hr-crm?trk=products_details_guest_secondary_call_to_action
Aumente suas vendas com o RD Station CRM Personalize, analise e acelere seu funil de vendas Personalize, analise e acelere seu funil de vendas Com o RD Station CRM você tem liberdade para evoluir: permissões personalizadas, automações flexíveis, dashboards customizados e IA para otimizar a rotina e vender mais. Com o RD Station CRM você tem liberdade para evoluir: permissões personalizadas, automações flexíveis, dashboards customizados e IA para otimizar a rotina e vender mais. Deixe seus dados para entrarmos em contato com você: Email* Email* Empresa* Qual o tamanho do time comercial? Selecione 1 vendedor 2 vendedores 3-5 vendedores 6-10 vendedores 11-20 vendedores 21-50 vendedores 50+ vendedores 1 + 1 = ? RECEBER CONTATO Mais de 20 mil empresas já escolheram o RD Station CRM! Tudo o que você precisa para ter alta performance em vendas Controle e acelere seu processo comercial Com o RD Station CRM, você centraliza seu processo de vendas, automatiza tarefas e ajusta o fluxo de trabalho conforme o crescimento da sua empresa, garantindo que tudo funcione de forma eficiente e organizada. Gerencie e analise seu time de vendas Saiba exatamente como sua equipe está performando e acompanhe cada venda ao vivo. Com relatórios detalhados, você tem uma visão clara dos resultados e pode ajustar suas estratégias com rapidez. Tenha flexibilidade para personalizar O RD Station CRM se adapta ao seu processo comercial: configure permissões personalizadas, crie automações flexíveis para agilizar tarefas, tenha dashboards customizados com as métricas mais relevantes e conte com IA para otimizar a rotina e vender mais rápido. Cresça junto dos melhores do Brasil A RD Station é líder em soluções de tecnologia para Marketing Digital, Vendas e Atendimento na América Latina, e entrega um conjunto de ferramentas completas para impulsionar negócios. Atualmente, temos 50 mil clientes em mais de 60 países utilizando diariamente nossas soluções. Por que escolher o RD Station CRM? Implementação simples O RD Station CRM é intuitivo. Depois de contratar, você implementa a ferramenta rapidamente e em poucos dias toda a sua equipe estará adaptada, sem nenhuma complicação! Apoio Educativo e estratégico A RD Station disponibiliza diversos níveis de apoio educativo, com materiais e tutoriais para facilitar o uso da ferramenta no dia a dia, além do suporte técnico de especialistas para implementação. Ferramenta Brasileira Entendemos o contexto das empresas nacionais e levamos isso em consideração no desenvolvimento e na evolução dos nossos produtos. Cobramos em moeda local, sem sustos na hora da fatura! Integrações e parceiros para ir além A RD Station possui um ecossistema com mais de 2 mil parceiros entre agências, consultores e soluções integradas para atender às suas necessidades mais estratégicas. Tudo pronto para levar suas vendas ao próximo nível? Conheça agora o RD Station CRM! RECEBER CONTATO Copyright © 2025 RD Station - Política de Privacidade
2026-01-13T09:29:19
https://www.netscout.com/product/arbor-cloud
Cloud DDoS Protection Service - On Demand DDoS Protection Skip to main content myNETSCOUT Login Solutions ← Back Enterprise Communication Service Provider By Industry Technology Partners Enterprise Observability nGenius Solutions for Observability Network Performance Management Application Performance Management Remote Site Observability Cloud Performance Monitoring VPN/VDI SaaS AIOps AIOps for Enterprises Collaboration UC&C and UCaaS Contact Centers Cybersecurity Network Detection and Response NetOps and SecOps Collaboration DDoS Protection Adaptive DDoS Protection Cybersecurity Stack Optimization DNS Infrastructure Protection 5G Transformation 5G for Enterprise Private 4G/5G Communication Service Provider AIOps AIOps for Communication Service Providers GeoMarketing Mobile Networks 5G 5G Fixed Wireless Access LTE/VoLTE/VoWi-Fi RAN Planning, Optimization, and Automation Cybersecurity & DDoS Protection Arbor DDoS Protection Carpet Bombing DDoS Protection DNS Infrastructure Protection Mobile Network Security Network Detection and Response Cloud and Digital Transformation Cloud Visibility Internet of Things (IoT) Network Function Virtualization (NFV) OTT Visibility Cable/MSO and Fixed Networks Cable/MSO and Fixed Networks Fiber (FTTx) and DAA Monitoring Analytics Automated Analytics Business Analytics By Industry Banking Capital Markets Education Government Healthcare Insurance Manufacturing Pharmaceutical Retail Telecommunications Transportation Utilities Featured Industry: Healthcare Optimize productivity, enhance the patient care experience, and improve bottom-line with powerful healthcare IT solutions. Learn more Technology Partners AWS Cisco Systems Citrix Dell Technologies F5 Google Cloud Microsoft Oracle Palo Alto Networks Red Hat ServiceNow Splunk VMware Products ← Back Enterprise Network Security Arbor DDoS Protection Communication Service Provider AIOps Enterprise nGeniusONE for Enterprise Enable real-time visibility into any enterprise network infrastructure environment to improve performance and user experience. InfiniStreamNG Capture, store, and analyze packet-level data in real-time to help troubleshoot issues and optimize performance across IT infrastructures. nGenius Edge Sensor Help improve user experience at remote sites and other network edges to achieve holistically strong network and application performance. vSTREAM Improve visibility into virtualized and cloud environments with deep packet inspection to further performance management and security. nGeniusPULSE Continuously test and track the performance of applications, network services, and user experience across all IT infrastructures. nGenius Packet Switches Packet flow switch architectures can combine with other NETSCOUT solutions to enable pervasive visibility across network environments. Network Security Omnis Cyber Intelligence NDR Leverage packet-level insights to power network detection and response (NDR) solutions and cyber threat hunting. Omnis CyberStream Detect, investigate, and respond to cyberthreats in real-time by leveraging packet-level data to identify suspicious activity across IT environments. nGenius Decryption Appliance Decrypt SSL and TLS-encrypted traffic to enable visibility into hidden threats within network traffic. Arbor DDoS Protection Arbor Sightline Leverage network intelligence to detect, analyze, and respond to DDoS threats across the most complex service provider and enterprise networks. Arbor Sightline Mobile Safeguard mobile services and ensure network performance and security with a DDoS protection solution specifically designed for mobile networks. Arbor Threat Mitigation System Detect, mitigate, and prevent DDoS attacks with real-time traffic analysis and automated attack mitigation. Arbor Edge Defense (AED) Powerful AI & ML powered DDoS protection backed by unmatched actionable threat intelligence maintains availability to critical services and applications. Arbor Global DDoS Threat Intelligence NETSCOUT global DDoS threat intelligence provides insights into the most current cyberthreats and their tactics, enhancing products and solutions. Arbor Cloud Cloud DDoS protection that defends against large-scale, volumetric attacks. Arbor Managed Services Round-the-clock support and assistance from Arbor product experts is available with managed services. Communication Service Provider nGenius Business Analytics Drive better business decisions at scale with efficient data collection, enrichment, and third-party application feeds. nGenius Session Analyzer Isolate and resolve congestion and interference problems contributing to 70% of all subscriber issues. nGeniusONE for Carrier Assure a premium user experience with end-to-end visibility across network domains, service enablers, and devices. InfiniStreamNG Access real-time views of RAN control plane for end-to-end visibility across physical and virtual network environments. ISNG-RAN Prioritize the mobile subscriber experience with scalable, cost-effective and cloud-ready RAN network monitoring. Omnis RAN Stay ahead of capacity and mobility targets with distinct automation modules to address growing RAN complexities. TrueCall Improve your understanding of network capacity and complexity with true user performance insights. AIOps Omnis AI Sensor Unparalleled telemetry insights at enterprise speed and scale across your entire IT ecosystem. Omnis AI Streamer Build data pipelines that sift through the noise to improve the efficiency, reliability, and scalability of IT operations and AI/ML workflows. Support & Resources ← Back Blogs, Resources, & Webinars Intelligence & Reports Support & Training Blogs, Resources, & Webinars Blog Stay updated with the latest insights, tips, and news on our blog. Resource Library Explore the latest case studies, whitepapers, eBooks, and more to discover real-world applications of our products and solutions. Learning Center Quickly learn the definitions of key phrases in networking, security, and DDoS. Webinars Gain insights from product and industry experts through our live & on-demand webinars. NETSCOUT Blog Whether you're a seasoned network performance professional or a novice in the industry, you're guaranteed to learn something new from the NETSCOUT blog. Read the blog Intelligence & Reports ASERT Threat Intelligence Team Meet the ASERT team, the leading experts on DDoS and cyber threat intelligence. DDoS Threat Intelligence Report Learn about the latest DDoS trends and statistics powered by NETSCOUT’s industry-leading global internet traffic visibility. Cyber Threat Horizon Look at live DDoS attack maps to discover what attacks are occurring in real-time. DDoS Threat Intelligence Report NETSCOUT’s latest DDoS Threat Intelligence dives into the most recent trends in DDoS adversarial tactics. Get the report Support & Training Overview Get support for your products and solutions to maximize the value of your investment. My.NETSCOUT Existing customers, log in to MyNETSCOUT to access additional assets, resources, and support channels. NETSCOUT University Our mission is to help you and your team to successfully implement NETSCOUT’s solutions so that your network functions seamlessly and effectively. NO LATENCY Explore tech tips, use cases, and more in our NO LATENCY newsletter. MasterCare Support NETSCOUT's award-winning global support service. Learn more Company ← Back About Us News & Events Partners About Us About Us Board of Directors Careers Customer Reviews Environmental, Social, and Governance Executive Team Industry Analyst Perspectives Investor Relations The Future of Intelligence The quality of your AI outcomes hinges on the data available. With NETSCOUT's Smart Data, ensure you have the network visibility you need to overcome the challenges of tomorrow. NETSCOUT provides the Data that Drives You. Watch the video News & Events Events Newsroom Press Releases Read the latest NETSCOUT news to stay informed on the latest updates, press, and more. Read the latest news Partners Partner Portal Become a Partner Learn about our NETSCOUT Connect 360 partner program. Search Under Attack? Contact Us Under Attack? Home Products Arbor Cloud DDoS Protection Services Arbor Cloud DDoS Protection Services Fully Managed, Intelligently Automated On-Demand DDoS Protection Overview Overview Benefits Features Resources Related Pages Proven Cloud DDoS Mitigation Provider NETSCOUT’s Arbor hybrid DDoS protection solution using Arbor Cloud employs an adaptive DDoS attack mitigation mechanism that operates both from data centers and the cloud. This approach is designed to ensure comprehensive protection against various DDoS threats. These services specifically counter high-volume attacks targeting bandwidth, "low and slow" attacks aimed at applications and infrastructure, and simultaneous multi-vector attacks. With Arbor Cloud, inbound traffic is routed via BGP or DNS through 16 DDoS scrubbing centers with over 15Tbps capacity. Even the largest attacks are well mitigated, and only clean traffic is returned to the internet access links and servers. Benefits Purpose Built Global DDoS Protection Arbor Cloud provides over 15 Tbps of DDoS attack mitigation capacity via 16 worldwide scrubbing centers located in Asia, Europe and The Americas. Seamless Hybrid Protection to Industry Best Practices On-premise protection guards against state-exhausting and application attacks aimed at your infrastructure in tandem with Arbor Cloud. It protects against attacks that target Firewalls, Intrusion Prevention Systems (IPS) and business-critical applications. Expertise The Arbor Cloud DDoS protection service is supported by NETSCOUT’s 24x7 ASERT Team. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Sub Minute Mitigation SLA With the intelligent automation capabilities of Arbor Cloud, we can provide mitigation initiation within 60-seconds of detecting an attack via cloud signaling from your AED, via flow-detection or as part of your always-on DDoS solution. Click to enlarge image Arbor Cloud Managed DDoS Protection Service for Enterprises Arbor Cloud combines on-premise DDoS defense with cloud-based traffic scrubbing services that are tightly integrated via an automated cloud signal. This multi-layered, adaptive DDoS protection approach is a proven industry best practice and is the only way to mitigate today's full spectrum of DDoS threats for both Service Providers and Enterprises, all from a single cloud DDoS protection provider. DDoS Scrubbing Capabilities NETSCOUT's Arbor Cloud is fueled by 16 DDoS scrubbing centers worldwide, boasting a capacity of more than 15 Tbps. These scrubbing centers are located across the globe to best protect against attacks originating from any corner of the world. This industry-leading capability helps monitor and maintain availability in the face of large-scale volumetric DDoS attacks . Solution Brief Why NETSCOUT's Arbor DDoS Attack Protection Solution is Better NETSCOUT Arbor has the industry’s broadest portfolio of DDoS attack protection products and services that enable organizations of any size to customize a solution to meet their technical and financial requirements of today – and the future. Learn more Features Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Using stateless packet-processing technology and/or cloud-based IP flow analysis, DDoS attacks can be automatically detected and routed to Arbor Cloud global DDoS scrubbing centers for mitigation. Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Click to enlarge image Arbor Cloud brings the availability of your network, services, and applications back under your control.  Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Click to enlarge image It can also be employed in an intelligent combination of in-cloud and on-premise hybrid DDoS protection, providing you the flexibility to design a comprehensive DDoS mitigation solution that fits your environment. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Click to enlarge image ATLAS Security Engineering & Response Team (ASERT) delivers world-class network security research and analysis for the benefit of today’s enterprise and network operators. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Using automated malware analysis pipelines, sinkholes, scanners, honeypots, open-source intelligence data sets and ASERT analysis, we can provide a unique view in the threat landscape.  Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Data Sheet NETSCOUT Arbor Cloud for Service Providers Augment your local mitigation capacity and staff with Arbor Cloud for Service Providers. Learn more Data Sheet NETSCOUT Arbor Cloud DDoS Protection for Enterprises NETSCOUT Arbor Cloud combines on-premise AED DDoS defense with global cloud-based traffic scrubbing services for best-of-breed, hybrid, DDoS attack protection. Read More Featured Resources Case Study Globe Telecom Protects Critical Infrastructure and Customer Experience with NETSCOUT Arbor DDoS Attack Protection Globe Telecom struggled with DDoS attacks due to a lack of real-time systems for detection and prevention. This caused multiple issues for the company. Report SPARK Matrix for Distributed Denial of Service (DDoS) Mitigation, Q3 2025 NETSCOUT Named both a Technology Leader and Ace Performer in DDoS Mitigation for 2025 White Paper IDC - Security Requirements Driving Comprehensive DDoS Protection Security buyers must maintain a vigilant approach toward assessing the state of their defenses—an incomplete strategy can be as costly of a mistake as a reactive strategy. White Paper EMA PRISM Report – DDoS Mitigation Solutions NETSCOUT Solution Profile Quick Look Healthcare Organizations are Increasingly Targets of DDoS Attacks Healthcare Organization Uses NETSCOUT Solution To Solve Cybersecurity Challenges by Stopping DDoS Attacks Impacting Availability of Services Solution Brief On-Premise Protection is the Best First Step Against DDoS and Cyberattacks for Academic Institutions Due to the complexity of DDoS attacks, Educational Institutions need to understand the risk reduction that on-premise edge protection can provide to their on line learning... Use Case Using Arbor Cloud to Protect Multi-Cloud and Hybrid Environments Against DDoS Attacks Arbor Cloud provides a superior means for protecting critical assets hosted in a single or multi-cloud environment, offering several advantages over less capable protection... Use Case Defending Government Agencies Against High-Volume Multi-Vector DDoS Attacks Attacks are often waged against national security infrastructure. NETSCOUT has observed a massive increase from the previous half year in attacks against the US national security... View more resources Related Pages Product ATLAS Intelligence Feed (AIF) ATLAS Intelligence Feed (AIF) provides up-to-date threat intelligence on the latest DDoS threats to our DDoS protection products. Automate defenses against trending attacks with AIF. Learn more Product Arbor Edge Defense (AED) Arbor Edge Defense (AED) is an inline security appliance deployed at the network perimeter, between the internet router and firewall. This AI and ML-powered solution provides stateless, on-premise DDoS protection. Learn more Product Arbor Sightline DDoS Attack Detection Solution Whether you are a service provider or have a complex enterprise network, Arbor Sightline helps monitor and identify networking and security issues at any network scale with AI and ML-powered insights. Learn more Product Arbor Threat Mitigation System for DDoS Attacks Filter out malicious traffic while allowing legitimate traffic to pass through, at any scale, with Arbor Threat Mitigation System (TMS). Learn more Contact NETSCOUT Today To learn more about how NETSCOUT can help your company, speak to one of our highly experienced subject matter experts. Contact Us Guardians of the Connected World ® Discover Our Data Observability Solutions Service Provider Service Assurance AIOps Omnis Network Security Arbor DDoS Protection All Solutions All Products Learn Blog Resource Library Learning Center MyNETSCOUT Events Webinars About About Us Careers Environmental, Social, and Governance For Investors Newsroom Partners Connect 1-888-357-7667 Contact Us Under DDoS Attack? © 2026 NETSCOUT Terms & Conditions Website Terms of Use Data Privacy and Trust Center Transparency in Supply Chains Privacy Policy Vendor Portal Do Not Sell or Share My Personal Information California Privacy Notice UK Disclosure
2026-01-13T09:29:19
https://git-scm.com/book/ru/v2/%d0%92%d0%b2%d0%b5%d0%b4%d0%b5%d0%bd%d0%b8%d0%b5-%d0%9a%d1%80%d0%b0%d1%82%d0%ba%d0%b0%d1%8f-%d0%b8%d1%81%d1%82%d0%be%d1%80%d0%b8%d1%8f-Git
Git - Краткая история Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Введение 1.1 О системе контроля версий 1.2 Краткая история Git 1.3 Что такое Git? 1.4 Командная строка 1.5 Установка Git 1.6 Первоначальная настройка Git 1.7 Как получить помощь? 1.8 Заключение 2. Основы Git 2.1 Создание Git-репозитория 2.2 Запись изменений в репозиторий 2.3 Просмотр истории коммитов 2.4 Операции отмены 2.5 Работа с удалёнными репозиториями 2.6 Работа с тегами 2.7 Псевдонимы в Git 2.8 Заключение 3. Ветвление в Git 3.1 О ветвлении в двух словах 3.2 Основы ветвления и слияния 3.3 Управление ветками 3.4 Работа с ветками 3.5 Удалённые ветки 3.6 Перебазирование 3.7 Заключение 4. Git на сервере 4.1 Протоколы 4.2 Установка Git на сервер 4.3 Генерация открытого SSH ключа 4.4 Настраиваем сервер 4.5 Git-демон 4.6 Умный HTTP 4.7 GitWeb 4.8 GitLab 4.9 Git-хостинг 4.10 Заключение 5. Распределённый Git 5.1 Распределённый рабочий процесс 5.2 Участие в проекте 5.3 Сопровождение проекта 5.4 Заключение 6. GitHub 6.1 Настройка и конфигурация учётной записи 6.2 Внесение собственного вклада в проекты 6.3 Сопровождение проекта 6.4 Управление организацией 6.5 Создание сценариев GitHub 6.6 Заключение 7. Инструменты Git 7.1 Выбор ревизии 7.2 Интерактивное индексирование 7.3 Припрятывание и очистка 7.4 Подпись 7.5 Поиск 7.6 Перезапись истории 7.7 Раскрытие тайн reset 7.8 Продвинутое слияние 7.9 Rerere 7.10 Обнаружение ошибок с помощью Git 7.11 Подмодули 7.12 Создание пакетов 7.13 Замена 7.14 Хранилище учётных данных 7.15 Заключение 8. Настройка Git 8.1 Конфигурация Git 8.2 Атрибуты Git 8.3 Хуки в Git 8.4 Пример принудительной политики Git 8.5 Заключение 9. Git и другие системы контроля версий 9.1 Git как клиент 9.2 Переход на Git 9.3 Заключение 10. Git изнутри 10.1 Сантехника и Фарфор 10.2 Объекты Git 10.3 Ссылки в Git 10.4 Pack-файлы 10.5 Спецификации ссылок 10.6 Протоколы передачи данных 10.7 Обслуживание репозитория и восстановление данных 10.8 Переменные окружения 10.9 Заключение A1. Приложение A: Git в других окружениях A1.1 Графические интерфейсы A1.2 Git в Visual Studio A1.3 Git в Visual Studio Code A1.4 Git в Eclipse A1.5 Git в IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.6 Git в Sublime Text A1.7 Git в Bash A1.8 Git в Zsh A1.9 Git в PowerShell A1.10 Заключение A2. Приложение B: Встраивание Git в ваши приложения A2.1 Git из командной строки A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Приложение C: Команды Git A3.1 Настройка и конфигурация A3.2 Клонирование и создание репозиториев A3.3 Основные команды A3.4 Ветвление и слияния A3.5 Совместная работа и обновление проектов A3.6 Осмотр и сравнение A3.7 Отладка A3.8 Внесение исправлений A3.9 Работа с помощью электронной почты A3.10 Внешние системы A3.11 Администрирование A3.12 Низкоуровневые команды 2nd Edition 1.2 Введение - Краткая история Git Краткая история Git Как и многие вещи в жизни, Git начинался с капелькой творческого хаоса и бурных споров. Ядро Linux — это достаточно большой проект с открытым исходным кодом. Большую часть времени разработки ядра Linux (1991–2002 гг.) изменения передавались между разработчиками в виде патчей и архивов. В 2002 году проект ядра Linux начал использовать проприетарную децентрализованную систему контроля версий BitKeeper. В 2005 году отношения между сообществом разработчиков ядра Linux и коммерческой компанией, которая разрабатывала BitKeeper, прекратились, и бесплатное использование утилиты стало невозможным. Это сподвигло сообщество разработчиков ядра Linux (а в частности Линуса Торвальдса — создателя Linux) разработать свою собственную утилиту, учитывая уроки, полученные при работе с BitKeeper. Некоторыми целями, которые преследовала новая система, были: Скорость Простая архитектура Хорошая поддержка нелинейной разработки (тысячи параллельных веток) Полная децентрализация Возможность эффективного управления большими проектами, такими как ядро Linux (скорость работы и разумное использование дискового пространства) С момента своего появления в 2005 году, Git развился в простую в использовании систему, сохранив при этом свои изначальные качества. Он удивительно быстр, эффективен в работе с большими проектами и имеет великолепную систему веток для нелинейной разработки (см. главу Ветвление в Git ). prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/bg/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d0%b8-%d0%bd%d0%b0-Git-%d0%a2%d0%b0%d0%b3%d0%be%d0%b2%d0%b5-%d0%b2-Git
Git - Тагове в Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Начало 1.1 За Version Control системите 1.2 Кратка история на Git 1.3 Какво е Git 1.4 Конзолата на Git 1.5 Инсталиране на Git 1.6 Първоначална настройка на Git 1.7 Помощна информация в Git 1.8 Обобщение 2. Основи на Git 2.1 Създаване на Git хранилище 2.2 Запис на промени в хранилището 2.3 Преглед на историята на действията 2.4 Възстановяване на направени действия 2.5 Работа с отдалечени хранилища 2.6 Тагове в Git 2.7 Псевдоними в Git 2.8 Обобщение 3. Клонове в Git 3.1 Накратко за разклоненията 3.2 Основи на клоновете код и сливането 3.3 Управление на клонове 3.4 Стратегии за работа с клонове код 3.5 Отдалечени клонове 3.6 Управление на проект 3.7 Обобщение 4. GitHub 4.1 Създаване и настройка на акаунт 4.2 Как да сътрудничим в проект 4.3 Управление на проект 4.4 Управление на организация 4.5 Автоматизиране с GitHub 4.6 Обобщение 5. Git инструменти 5.1 Избор на къмити 5.2 Интерактивно индексиране 5.3 Stashing и Cleaning 5.4 Подписване на вашата работа 5.5 Търсене 5.6 Манипулация на историята 5.7 Мистерията на командата Reset 5.8 Сливане за напреднали 5.9 Rerere 5.10 Дебъгване с Git 5.11 Подмодули 5.12 Пакети в Git (Bundling) 5.13 Заместване 5.14 Credential Storage система 5.15 Обобщение 6. Настройване на Git 6.1 Git конфигурации 6.2 Git атрибути 6.3 Git Hooks 6.4 Примерна Git-Enforced политика 6.5 Обобщение 7. Git и други системи 7.1 Git като клиент 7.2 Миграция към Git 7.3 Обобщение 8. Git на ниско ниво 8.1 Plumbing и Porcelain команди 8.2 Git обекти 8.3 Git референции 8.4 Packfiles 8.5 Refspec спецификации 8.6 Транспортни протоколи 8.7 Поддръжка и възстановяване на данни 8.8 Environment променливи 8.9 Обобщение 9. Приложение A: Git в други среди 9.1 Графични интерфейси 9.2 Git във Visual Studio 9.3 Git във Visual Studio Code 9.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 9.5 Git в Sublime Text 9.6 Git в Bash 9.7 Git в Zsh 9.8 Git в PowerShell 9.9 Обобщение 10. Приложение B: Вграждане на Git в приложения 10.1 Git от команден ред 10.2 Libgit2 10.3 JGit 10.4 go-git 10.5 Dulwich A1. Приложение C: Git команди A1.1 Настройки и конфигурация A1.2 Издърпване и създаване на проекти A1.3 Snapshotting A1.4 Клонове и сливане A1.5 Споделяне и обновяване на проекти A1.6 Инспекция и сравнение A1.7 Дебъгване A1.8 Patching A1.9 Email команди A1.10 Външни системи A1.11 Административни команди A1.12 Plumbing команди 2nd Edition 2.6 Основи на Git - Тагове в Git Тагове в Git Подобно на повечето VCS системи, Git позволява да маркирате (тагвате) специфични точки от историята на хранилището като важни. Обикновено това се използва за маркиране на различни версии на проекта (v1.0, v2.0 и т.н.). В тази секция, ще научите как да показвате наличните тагове, да създавате и премахвате тагове и да ги различавате по тип. Показване на таговете Показването на налични тагове в Git е просто. Просто напишете git tag (с опционални параметри -l или --list ): $ git tag v1.0 v2.0 Тази команда отпечатва таговете по азбучен ред, редът по който са изобразени няма реално значение. Можете също да търсите тагове по определен стринг. Хранилището на Git например, съдържа повече от 500 тага. Ако се интересувате само от версиите 1.8.5, можете да изпълните следното: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Забележка Използването на wildcards изисква параметъра -l или --list Ако просто искате целия списък тагове, изпълнението на командата git tag без параметри изрично подразбира, че желаете списък и го показва, в този случай използването на -l или --list е по желание Ако обаче подадете wildcard маска за търсене на имена на тагове, тогава -l или --list са задължителни. Създаване на тагове Git поддържа два основни типа тагове: lightweight и annotated . Lightweight тагът прилича на branch, който не се променя — това е просто указател към специфичен къмит. Annotated таговете обаче, се съхраняват като пълни обекти в базата данни - те съдържат имейла и името на тагващия, дата, описателно съобщение и дори могат да се подписват и проверяват с GNU Privacy Guard (GPG). Хубаво е да се създават annotated тагове, защото тогава съхранявате всичката описана информация за тях, но ако искате временен такъв или по някаква причина не искате да пазите подробните описания, lightweight таговете също са вариант. Annotated тагове Създаването на анотиран таг в Git е лесно. Най-лесният начин е да подадете флага -a , когато пускате командата tag : $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 Флагът -m указва съобщението за тага, което ще се съхранява заедно с него. Ако не укажете такова, Git ще стартира редактора ви, така че да можете да го напишете, точно както при къмитите. Можете да разгледате данните за тага с къмита, който е бил тагнат с командата git show : $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Тя показва информация за тагващия разработчик, датата на която къмитът е бил тагнат и съобщението на тага преди да покаже информацията за самия къмит. Lightweight тагове Другият начин да тагвате даден къмит е с lightweight таг. Това най-просто казано е чексумата на къмита записана във файл - не се съхранява друга информация. За да създадете такъв олекотен таг, не подавайте флаговете -a , -s , или -m - просто укажете името на тага: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Този път, ако пуснете git show за тага, няма да виждате допълнителна информация за него. Командата просто показва къмита: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Тагване на предишни къмити Можете да тагвате къмити и след като сте приключили с тях. Представете си, че историята на вашите къмити изглежда така: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme Да предположим, че сте забравили да тагнете проекта като версия v1.2, която е трябвало да бъде маркирана в къмита регистриран като “Updated rakefile”. Можете да го направите и в по-късен момент. За да тагнете този стар къмит, подайте чексумата му (или част от нея) в края на командата: $ git tag -a v1.2 9fceb02 Сега можете да проверите дали сте тагнали къмита успешно: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... Споделяне на тагове По подразбиране, git push не изпраща таговете към отдалечените сървъри. Ще трябва ръчно да ги изпратите след като сте ги създали. Този процес е точно като споделяне на отдалечени клонове — можете да изпълните git push origin <tagname> . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Ако имате много тагове и искате да ги изпратите наведнъж, подайте на командата флага --tags . Това ще трансферира всички ваши тагове, които не са били налични на сървъра наведнъж. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Сега, ако някой клонира или издърпва от вашето хранилище, ще получи и таговете ви. Забележка git push публикува и двата вида тагове Публикуването на тагове с git push <remote> --tags изпраща и lightweight и annotated таговете. В момента няма начин за изпращане само на lightweight тагове, но ако използвате командата git push <remote> --follow-tags , тогава към отдалеченото хранилище ще бъдат изпратени само annotated таговете. Изтриване на тагове За да изтриете таг от локалното си хранилище, може да използвате командата git tag -d <tagname> . Например, можем да изтрием lightweight тага отгоре така: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Това обаче не изтрива тага от никой отдалечен сървър. Съществуват два начина за изтриване на таг от отдалечен сървър. Първият е да използвате git push <remote> :refs/tags/<tagname> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw Начинът да изтълувате горното странно изписване е да го възприемете като изпращане на нулева стойност преди двуеточието към името на отдалечения таг, което ефективно го изтрива. Вторият (и по-интуитивен) начин е с команда като тази: $ git push origin --delete <tagname> Извличане по тагове Ако искате да видите версиите на файловете, към които сочи даден таг, можете да направите git checkout на този таг, въпреки че това ще постави хранилището в статус “detached HEAD”, което има някои неприятни странични ефекти: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image В режим “detached HEAD”, ако направите промени и след това създадете къмит, тагът ще остане същия, но новия ви къмит няма да принадлежи към никой клон и няма да бъде достъпен освен по точния хеш на къмита. Затова, ако трябва да правите промени, например да поправите грешка в стара версия например — вероятно ще искате да създадете клон: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Разбира се, ако направите това и направите къмит, вашият version2 клон ще бъде леко различен от тага v2.0.0 , защото ще се премести напред с вашите промени, така че бъдете внимателни. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/fr/v2/D%c3%a9marrage-rapide-Installation-de-Git
Git - Installation de Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Démarrage rapide 1.1 À propos de la gestion de version 1.2 Une rapide histoire de Git 1.3 Rudiments de Git 1.4 La ligne de commande 1.5 Installation de Git 1.6 Paramétrage à la première utilisation de Git 1.7 Obtenir de l’aide 1.8 Résumé 2. Les bases de Git 2.1 Démarrer un dépôt Git 2.2 Enregistrer des modifications dans le dépôt 2.3 Visualiser l’historique des validations 2.4 Annuler des actions 2.5 Travailler avec des dépôts distants 2.6 Étiquetage 2.7 Les alias Git 2.8 Résumé 3. Les branches avec Git 3.1 Les branches en bref 3.2 Branches et fusions : les bases 3.3 Gestion des branches 3.4 Travailler avec les branches 3.5 Branches de suivi à distance 3.6 Rebaser (Rebasing) 3.7 Résumé 4. Git sur le serveur 4.1 Protocoles 4.2 Installation de Git sur un serveur 4.3 Génération des clés publiques SSH 4.4 Mise en place du serveur 4.5 Démon (Daemon) Git 4.6 HTTP intelligent 4.7 GitWeb 4.8 GitLab 4.9 Git hébergé 4.10 Résumé 5. Git distribué 5.1 Développements distribués 5.2 Contribution à un projet 5.3 Maintenance d’un projet 5.4 Résumé 6. GitHub 6.1 Configuration et paramétrage d’un compte 6.2 Contribution à un projet 6.3 Maintenance d’un projet 6.4 Gestion d’un regroupement 6.5 Écriture de scripts pour GitHub 6.6 Résumé 7. Utilitaires Git 7.1 Sélection des versions 7.2 Indexation interactive 7.3 Remisage et nettoyage 7.4 Signer votre travail 7.5 Recherche 7.6 Réécrire l’historique 7.7 Reset démystifié 7.8 Fusion avancée 7.9 Rerere 7.10 Déboguer avec Git 7.11 Sous-modules 7.12 Empaquetage (bundling) 7.13 Replace 7.14 Stockage des identifiants 7.15 Résumé 8. Personnalisation de Git 8.1 Configuration de Git 8.2 Attributs Git 8.3 Crochets Git 8.4 Exemple de politique gérée par Git 8.5 Résumé 9. Git et les autres systèmes 9.1 Git comme client 9.2 Migration vers Git 9.3 Résumé 10. Les tripes de Git 10.1 Plomberie et porcelaine 10.2 Les objets de Git 10.3 Références Git 10.4 Fichiers groupés 10.5 La refspec 10.6 Les protocoles de transfert 10.7 Maintenance et récupération de données 10.8 Les variables d’environnement 10.9 Résumé A1. Annexe A: Git dans d’autres environnements A1.1 Interfaces graphiques A1.2 Git dans Visual Studio A1.3 Git dans Visual Studio Code A1.4 Git dans IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git dans Sublime Text A1.6 Git dans Bash A1.7 Git dans Zsh A1.8 Git dans PowerShell A1.9 Résumé A2. Annexe B: Embarquer Git dans vos applications A2.1 Git en ligne de commande A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Commandes Git A3.1 Installation et configuration A3.2 Obtention et création des projets A3.3 Capture d’instantané basique A3.4 Création de branches et fusion A3.5 Partage et mise à jour de projets A3.6 Inspection et comparaison A3.7 Débogage A3.8 Patchs A3.9 Courriel A3.10 Systèmes externes A3.11 Administration A3.12 Commandes de plomberie 2nd Edition 1.5 Démarrage rapide - Installation de Git Installation de Git Avant de commencer à utiliser Git, il faut qu’il soit disponible sur votre ordinateur. Même s’il est déjà installé, c’est probablement une bonne idée d’utiliser la dernière version disponible. Vous pouvez l’installer soit comme paquet ou avec un installateur, soit en téléchargeant le code et en le compilant par vous-même. Note Ce livre a été écrit en utilisant Git version 2.8.0 . Bien que la plupart des commandes utilisées fonctionnent vraisemblablement encore avec d’anciennes version de Git, certaines peuvent agir différemment. Comme Git est particulièrement excellent pour préserver les compatibilités amont, toute version supérieure à 2.8 devrait fonctionner sans différence. Installation sur Linux Si vous voulez installer les outils basiques de Git sur Linux via un installateur binaire, vous pouvez généralement le faire au moyen de l’outil de gestion de paquet fourni avec votre distribution. Sur Fedora (ou toute distribution parente basée sur RPM, telle que RHEL ou CentOS), vous pouvez utiliser dnf  : $ sudo dnf install git-all Sur une distribution basée sur Debian, telle que Ubuntu, essayez apt  : $ sudo apt install git-all Pour plus d’options, des instructions d’installation sur différentes versions Unix sont disponibles sur le site web de Git, à https://git-scm.com/download/linux . Installation sur macOS Il existe plusieurs méthodes d’installation de Git sur un Mac. La plus facile est probablement d’installer les Xcode Command Line Tools . Sur Mavericks (10.9) ou postérieur, vous pouvez simplement essayer de lancer git dans le terminal la première fois. $ git --version S’il n’est pas déjà installé, il vous demandera de le faire. Si vous souhaitez une version plus à jour, vous pouvez aussi l’installer à partir de l’installateur binaire. Un installateur de Git pour macOS est maintenu et disponible au téléchargement sur le site web de Git à https://git-scm.com/download/mac . Figure 7. Installateur de Git pour macOS Vous pouvez aussi l’installer comme sous-partie de l’installation de GitHub pour macOS. Leur outil Git graphique a une option pour installer les outils en ligne de commande. Vous pouvez télécharger cet outil depuis le site web de GitHub pour macOS, à https://desktop.github.com . Installation sur Windows Il existe aussi plusieurs manières d’installer Git sur Windows. L’application officielle est disponible au téléchargement sur le site web de Git. Rendez-vous sur https://git-scm.com/download/win et le téléchargement démarrera automatiquement. Notez que c’est un projet nommé Git for Windows (appelé aussi msysGit), qui est séparé de Git lui-même ; pour plus d’information, rendez-vous à https://msysgit.github.io/ . Pour obtenir une installation automatisée, vous pouvez utiliser le paquet Chocolatey Git . Notez que le paquet Chocolatey est maintenu par la communauté. Une autre méthode facile pour installer Git est d’installer Github for Windows . L’installateur inclut une version en ligne de commande avec l’interface graphique. Elle fonctionne aussi avec PowerShell et paramètre correctement les caches d’authentification et les réglages CRLF. Nous en apprendrons plus sur ces sujets plus tard, mais il suffit de savoir que ces options sont très utiles. Vous pouvez télécharger ceci depuis le site de Github for Windows , à l’adresse https://windows.github.com . Installation depuis les sources Certains peuvent plutôt trouver utile d’installer Git depuis les sources car on obtient la version la plus récente. Les installateurs de version binaire tendent à être un peu en retard, même si Git a gagné en maturité ces dernières années, ce qui limite les évolutions. Pour installer Git, vous avez besoin des bibliothèques suivantes : autotools, curl, zlib, openssl, expat, libiconv. Par exemple, si vous avez un système d’exploitation qui utilise dnf (tel que Fedora) ou apt-get (tel qu’un système basé sur Debian), vous pouvez utiliser l’une des commandes suivantes pour installer les dépendances minimales pour compiler et installer les binaires Git : $ sudo dnf install dh-autoreconf curl-devel expat-devel gettext-devel \ openssl-devel perl-devel zlib-devel $ sudo apt-get install dh-autoreconf libcurl4-gnutls-dev libexpat1-dev \ gettext libz-dev libssl-dev Pour pouvoir ajouter la documentation dans différents formats (doc, html, info), ces dépendances suppplémentaires sont nécessaires : $ sudo dnf install asciidoc xmlto docbook2X $ sudo apt-get install asciidoc xmlto docbook2x Note Les utilisateurs de RHEL ou dérivés tel que CentOS et Scientific Linux devront activer le dépôt EPEL pour télécharger le paquet docbooc2X . Si vous utilisez une distribution basée sur Debian (Debian/Ubuntu/dérivés d’Ubuntu), vous avez aussi besoin du paquet install-info  : $ sudo apt-get install install-info Si vous utilisez une distribution basée sur RPM (Fedora/RHEL/dérivés de RHEL), vous avez aussi besoin du paquet getopt (qui est déjà installé sur les distributions basées sur Debian) : $ sudo dnf install getopt De plus, si vous utilisez Fedora/RHEL/dérivé de RHEL, vous devez faire ceci : $ sudo ln -s /usr/bin/db2x_docbook2texi /usr/bin/docbook2x-texi à cause des différences de nom des binaires. Quand vous avez toutes les dépendances nécessaires, vous pouvez poursuivre et télécharger la dernière version de Git depuis plusieurs sites. Vous pouvez l’obtenir via Kernel.org, à https://www.kernel.org/pub/software/scm/git , ou sur le miroir sur le site web GitHub à https://github.com/git/git/releases . L’indication de la dernière version est généralement plus claire sur la page GitHub, mais la page kernel.org a également des signatures de version si vous voulez vérifier votre téléchargement. Puis, compilez et installez : $ tar -zxf git-2.8.0.tar.gz $ cd git-2.8.0 $ make configure $ ./configure --prefix=/usr $ make all doc info $ sudo make install install-doc install-html install-info Après ceci, vous pouvez obtenir Git par Git lui-même pour les mises à jour : $ git clone git://git.kernel.org/pub/scm/git/git.git prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/kaspersky-ddos-protection/?trk=products_seo_search
Kaspersky DDoS Protection | LinkedIn Skip to main content LinkedIn Kaspersky in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Kaspersky DDoS Protection DDoS Protection Software by Kaspersky See who's skilled in this Add as skill Learn more Report this product About Distributed Denial of Service (DDoS) can cause financial costs for your business. Keep your service protected with Kaspersky DDoS-Protection! Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less Kaspersky products Kaspersky Container Security Kaspersky Container Security Container Security Software Kaspersky Digital Footprint Intelligence Kaspersky Digital Footprint Intelligence Digital Forensics Software Kaspersky Industrial Cybersecurity Kaspersky Industrial Cybersecurity Enterprise Architecture Software Kaspersky Next Kaspersky Next Endpoint Detection & Response (EDR) Software Kaspersky Password Manager Kaspersky Password Manager Password Management Software Kaspersky Premium Kaspersky Premium Antivirus Software Kaspersky Safe Kids Kaspersky Safe Kids Antivirus Software Kaspersky Threat Intelligence Kaspersky Threat Intelligence Threat Intelligence Platforms Kaspersky VPN Kaspersky VPN Virtual Private Network (VPN) Software KasperskyOS KasperskyOS Operating Systems Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/az/v2/Git-Al%c9%99tl%c9%99ri-Reviziya-Se%c3%a7imi
Git - Reviziya Seçimi About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlanğıc 1.1 Versiyaya Nəzarət Haqqında 1.2 Git’in Qısa Hekayəsi 1.3 Git Nədir? 1.4 Əmr Sətiri 1.5 Git’i Quraşdırmaq 1.6 İlk Dəfə Git Quraşdırması 1.7 Kömək Almaq 1.8 Qısa Məzmun 2. Git’in Əsasları 2.1 Git Deposunun Əldə Edilməsi 2.2 Depoda Dəyişikliklərin Qeyd Edilməsi 2.3 Commit Tarixçəsinə Baxış 2.4 Ləğv Edilən İşlər (Geri qaytarılan) 2.5 Uzaqdan İşləmək 2.6 Etiketləmə 2.7 Git Alias’lar 2.8 Qısa Məzmun 3. Git’də Branch 3.1 Nutshell’də Branch’lar 3.2 Sadə Branching və Birləşdirmə 3.3 Branch İdarəedilməsi 3.4 Branching İş Axınları 3.5 Uzaq Branch’lar 3.6 Rebasing 3.7 Qısa Məzmun 4. Server’də Git 4.1 Protokollar 4.2 Serverdə Git Əldə Etmək 4.3 Sizin öz SSH Public Key’nizi yaratmaq 4.4 Server qurmaq 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Tərəf Seçimləri 4.10 Qısa Məzmun 5. Paylanmış Git 5.1 Distribyutorluq İş Axınları 5.2 Layihəyə Töhfə vermək 5.3 Layihənin Saxlanılması 5.4 Qısa Məzmun 6. GitHub 6.1 Hesab Qurma və Konfiqurasiya 6.2 Bir Layihəyə Töhfə Vermək 6.3 Bir Layihənin Saxlanılması 6.4 Bir Təşkilatı Idarə Etmək 6.5 GitHub Skriptləmə 6.6 Qısa Məzmun 7. Git Alətləri 7.1 Reviziya Seçimi 7.2 Interaktiv Səhnələşdirmə 7.3 Stashing və Təmizləmə 7.4 İşinizin İmzalanması 7.5 Axtarış 7.6 Tarixi Yenidən Yazmaq 7.7 Reset Demystified 7.8 İnkişaf etmiş Birləşmə 7.9 Rerere 7.10 Git ilə Debugging 7.11 Alt Modullar 7.12 Bundling 7.13 Dəyişdirmək 7.14 Etibarlı Yaddaş 7.15 Qısa Məzmun 8. Git’i Fərdiləşdirmək 8.1 Git Konfiqurasiyası 8.2 Git Atributları 8.3 Git Hook’ları 8.4 Git-Enforced Siyasət Nümunəsi 8.5 Qısa Məzmun 9. Git və Digər Sistemlər 9.1 Git Müştəri kimi 9.2 Git’ə Miqrasiya 9.3 Qısa Məzmun 10. Git’in Daxili İşləri 10.1 Plumbing və Porcelain 10.2 Git Obyektləri 10.3 Git Referansları 10.4 Packfile’lar 10.5 Refspec 10.6 Transfer Protokolları 10.7 Maintenance və Məlumatların Bərpası 10.8 Mühit Dəyişənləri 10.9 Qısa Məzmun A1. Appendix A: Digər Mühitlərdə Git A1.1 Qrafik interfeyslər A1.2 Visual Studio’da Git A1.3 Visual Studio Code’da Git A1.4 Eclipse’də Git A1.5 Sublime Text’də Git A1.6 Bash’da Git A1.7 Zsh’də Git A1.8 PowerShell’də Git A1.9 Qısa Məzmun A2. Appendix B: Proqramlara Git Daxil Etmək A2.1 Əmr-sətri Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Əmrləri A3.1 Quraşdırma və Konfiqurasiya A3.2 Layihələrin Alınması və Yaradılması A3.3 Sadə Snapshotting A3.4 Branching və Birləşmə A3.5 Layihələrin Paylaşılması və Yenilənməsi A3.6 Yoxlama və Müqayisə A3.7 Debugging A3.8 Patching A3.9 E-poçt A3.10 Xarici Sistemlər A3.11 İdarəetmə A3.12 Plumbing Əmrləri 2nd Edition 7.1 Git Alətləri - Reviziya Seçimi İndiyə qədər mənbə kodu nəzarəti üçün Git deposunu idarə etməli və ya saxlamağınız lazım olan gündəlik əmrlərin və iş axınlarının çoxunu öyrəndiniz. Faylları tracking və committing əsas tapşırıqlarını yerinə yetirdiniz və quruluş sahəsinin gücünü və yüngül mövzunun branching’ni və birləşməsini istifadə etdiniz. İndi Git’in gündəlik olaraq istifadə edə bilməyəcəyiniz, ancaq bir anda ehtiyacınız ola biləcəyi çox güclü şeyləri araşdıracaqsınız. Reviziya Seçimi Git, bir sıra commit-lərə, commit-lər dəstinə və ya commit-lərə istinad etməyə imkan verir. Bunlar mütləq açıq deyil, bilmək faydalıdır. Tək Reviziyalar Tamamilə 40 xarakterli SHA-1 hash ilə hər hansı bir commit-ə istinad edə bilərsiniz, lakin commit-ləri ifadə etmənin daha çox insan dostu yolları var. Bu bölüm, hər hansı bir commit-ə istinad edə biləcəyiniz müxtəlif yolları əks etdirir. Qısa SHA-1 Git, SHA-1 hash-nın ilk bir neçə simvolunu verdiyiniz təqdirdə nəyi nəzərdə tutduğunuzu başa düşmək üçün kifayət qədər ağıllıdır, qismən qarışıq ən azı dörd simvol uzun və birmənalıdır; Başqa sözlə, obyekt verilənlər bazasındakı heç bir obyektdə eyni prefikslə başlayan bir hash ola bilməz. Məsələn, müəyyən bir funksionallıq əlavə etdiyinizi bildiyiniz xüsusi bir commit-i araşdırmaq üçün əvvəlcə commit-i tapmaq üçün git log əmrini işə sala bilərsiniz: $ git log commit 734713bc047d87bf7eac9674765ae793478c50d3 Author: Scott Chacon <schacon@gmail.com> Date: Fri Jan 2 18:32:33 2009 -0800 Fix refs handling, add gc auto, update tests commit d921970aadf03b3cf0e71becdaab3147ba71cdef Merge: 1c002dd... 35cfb2b... Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' commit 1c002dd4b536e7479fe34593e72e6c6c1819e53b Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 14:58:32 2008 -0800 Add some blame and merge stuff Bu vəziyyətdə, hash 1c002dd... ilə başlayan commit-lə maraqlandığınızı varsayaq. Aşağıdakı git show varyasyonlarından hər hansı biri ilə əlaqəli olanı yoxlaya bilərsiniz (daha qısa versiyaların birmənalı olduğunu düşünərək): $ git show 1c002dd4b536e7479fe34593e72e6c6c1819e53b $ git show 1c002dd4b536e7479f $ git show 1c002d Git, SHA-1 dəyərləriniz üçün qısa, bənzərsiz bir qısaltmanı müəyyən edə bilər. --abbrev-commit git log əmrinə keçsəniz, çıxış daha qısa dəyərlərdən istifadə edəcək, lakin onları unikal saxlayır; yeddi simvol istifadə etmək üçün standartdır, lakin SHA-1-in birmənalı olması üçün onları daha uzun edir: $ git log --abbrev-commit --pretty=oneline ca82a6d Change the version number 085bb3b Remove unnecessary test code a11bef0 Initial commit Ümumiyyətlə, səkkiz-on simvol bir proyektdə bənzərsiz olmaq üçün kifayətdir. Məsələn, 2019-cu ilin fevral ayından etibarən, Linux kernelinin (olduqca əhəmiyyətli bir layihədir) 875.000-dən çox commit-i və obyekt bazasında təxminən yeddi milyon obyekti var, ilk 12 simvolda SHA-1’ləri eyni olan iki obyekt yoxdur. Note SHA-1 HAQQINDA QISA QEYD Bir çox insan təsadüfi bir şəkildə, eyni SHA-1 dəyərinə qarışan depolarında iki fərqli obyektə sahib olacaqlarından bir anda narahat olurlar. Bəs onda nə etmək lazımdır? Əgər deponuzdakı əvvəlki fəərqli obyekti ilə eyni SHA-1 dəyərinə bərabər olan bir obyekt törətmisinizsə, Git əvvəlki obyekti Git verilənlər bazanızda görəcək, artıq yazıldığını düşünün və sadəcə yenidən istifadə edin. Bir nöqtədə yenidən həmin obyekti yoxlamağa çalışsanız, həmişə ilk obyektin məlumatlarını əldə edəcəksiniz. Bununla birlikdə, bu ssenarinin nə qədər gülünc bir şəkildə ehtimal olunmadığının fərqində olmalısınız. SHA-1 həcmi 20 bayt və ya 160 bitdir. Tək bir toqquşma ehtimalının 50% olmasını təmin etmək üçün lazım olan təsadüfi yığılmış obyektlərin sayı təxminən 2 80 -dir (toqquşma ehtimalını müəyyənləşdirmək üçün düstur p = (n(n-1)/2) * (1/2^160)) . 2 80 1.2 x 10 24 təşkil edir və ya 1 milyon milyard. Bu, yer üzündə qum dənələrinin sayından 1,200 dəfə çoxdur. Burada SHA-1 toqquşması üçün nə lazım olduğunu düşünmək üçün bir nümunə var. Yer üzündəki 6,5 milyard insanın hamısı proqramlaşdırma aparsaydı və hər saniyədə hər biri bütün Linux nüvə tarixinə (6.5 milyon Git obyekt) bərabər olan bir kod istehsal etsəydi və onu böyük bir Git deposuna salsaydı, təxminən 2 il çəkərdi. Bu depoda bir SHA-1 obyektinin toqquşma ehtimalı 50% -ə çatacaq qədər obyekt var qədər. Beləliklə, SHA-1 toqquşması, proqramlaşma komandanızın hər bir üzvünün eyni gecədə əlaqəsi olmayan hadisələrdə canavarların hücumuna məruz qalması və öldürülməsi ehtimalı daha azdır. Branch Referansları Müəyyən bir commit-ə istinad etməyin bir sadə yolu branch-ın ucundakı commit-in olmasıdır; bu halda, sadəcə bir commit-ə istinad gözləyən hər hansı bir Git əmrində branch adını istifadə edə bilərsiniz. Məsələn, bir branch-dakı son commit obyektini araşdırmaq istəyirsinizsə, aşağıdakı mövzuda əmrlər ekvivalentdir ki, topic1 branch-ının ca82a6d... işarə etdiyini göstərir: $ git show ca82a6dff817ec66f44342007202690a93763949 $ git show topic1 Bir branch-ın hansı konkret SHA-1-ə işarə etdiyini görmək və ya bu nümunələrdən hər hansı birinin SHA-1-lər baxımından nəyə bənzədiyini görmək istəyirsinizsə, rev-parse adlı Git santexnika alətindən istifadə edə bilərsiniz. Santexnika alətləri haqqında daha çox məlumat üçün Git’in Daxili İşləri -ə baxa bilərsiniz; əsasən, rev-parse aşağı səviyyəli əməliyyatlar üçün mövcuddur və gündəlik əməliyyatlarda istifadə üçün nəzərdə tutulmayıb. Ancaq bəzən həqiqətən nələrin baş verdiyini görmək lazım olduqda faydalı ola bilər. Burada branch-ınızda rev-parse işlədə bilərsiniz. $ git rev-parse topic1 ca82a6dff817ec66f44342007202690a93763949 RefLog Qısa Adları Gitin arxada işləyərkən arxa planda gördüyü işlərdən biri də reflog saxlamaqdır - HEAD və branch istinadlarınızın son bir neçə ayda olduğu bir qeyd. Reflogunuzu git reflog istifadə edərək görə bilərsiniz: $ git reflog 734713b HEAD@{0}: commit: Fix refs handling, add gc auto, update tests d921970 HEAD@{1}: merge phedders/rdocs: Merge made by the 'recursive' strategy. 1c002dd HEAD@{2}: commit: Add some blame and merge stuff 1c36188 HEAD@{3}: rebase -i (squash): updating HEAD 95df984 HEAD@{4}: commit: # This is a combination of two commits. 1c36188 HEAD@{5}: rebase -i (squash): updating HEAD 7e05da5 HEAD@{6}: rebase -i (pick): updating HEAD Branch ucunuz hər hansı bir səbəbdən yeniləndikdə Git bu məlumatları sizin üçün bu müvəqqəti tarixdə saxlayır. Köhnə commit-lərə də istinad etmək üçün reflog məlumatlarınızı istifadə edə bilərsiniz. Məsələn, deposunuzun HEAD-in əvvəlki beşinci dəyərini görmək istəyirsinizsə, reflog çıxışında gördüyünüz @{5} istinadından istifadə edə bilərsiniz: $ git show HEAD@{5} Bu sintaksisdən branch-ın müəyyən bir müddət əvvəl harada olduğunu görmək üçün də istifadə edə bilərsiniz. Məsələn, dünən master branch-ınızın harada olduğunu görmək üçün yaza bilərsiniz: $ git show master@{yesterday} Bu, dünən " master branch-ınızın ucunun harada olduğunu göstərəcəkdir. Bu texnika yalnız hələ də qeydlərinizdə olan məlumatlar üçün işləyir, buna görə də bir neçə aydan yuxarı commit-lər axtarmaq üçün istifadə edə bilməzsiniz. Reflog məlumatlarını git log çıxışı kimi formatlanmış şəkildə görmək üçün git log -g işlədə bilərsiniz: $ git log -g master commit 734713bc047d87bf7eac9674765ae793478c50d3 Reflog: master@{0} (Scott Chacon <schacon@gmail.com>) Reflog message: commit: Fix refs handling, add gc auto, update tests Author: Scott Chacon <schacon@gmail.com> Date: Fri Jan 2 18:32:33 2009 -0800 Fix refs handling, add gc auto, update tests commit d921970aadf03b3cf0e71becdaab3147ba71cdef Reflog: master@{1} (Scott Chacon <schacon@gmail.com>) Reflog message: merge phedders/rdocs: Merge made by recursive. Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' Qeyd etmək vacibdir ki, reflog məlumatları ciddi şəkildə localdır - bu, yalnız deponuzda etdiyiniz işlərin bir qeydidir. Referanslar başqasının deposunun kopyasında eyni olmayacaq; ayrıca, əvvəlcə bir deponu klonladıqdan dərhal sonra, depoda hələ heç bir fəaliyyət baş vermədiyi üçün boş bir refloqa sahib olacaqsınız. git show HEAD@{2.months.ago} -ı işə salmaq, sizə yalnız ən azı iki ay əvvəl layihəni klonlaşdırdığınız təqdirdə uyğunlaşma commit-ini göstərəcəkdir - daha yaxınlarda klonlaşdırsanız, yalnız ilk local commit-i görəcəksiniz. Tip Reflogu Git-in shell tarixinin versiyası kimi düşünün UNIX və ya Linux arxa planınız varsa, reflog-u Git-in shell tarixinin versiyası olaraq düşünə bilərsiniz, burada olanların yalnız sizin və sizin “sessiyanız” üçün açıq şəkildə əlaqəli olduğunu vurğulayan və eyni maşında işləyə başqa heç kimlə ilə əlaqəsi yoxdur. Ancestry Referansları Bir commit-i müəyyənləşdirməyin digər əsas yolu əcdadı ilə bağlıdır. Bir referansın sonunda bir ^ (caret) qoysanız, Git, bu commit-in valideynini ifadə etmək üçün onu təhlil edir. Tutaq ki, layihənizin tarixinə nəzər yetirdiniz: $ git log --pretty=format:'%h %s' --graph * 734713b Fix refs handling, add gc auto, update tests * d921970 Merge commit 'phedders/rdocs' |\ | * 35cfb2b Some rdoc changes * | 1c002dd Add some blame and merge stuff |/ * 1c36188 Ignore *.gem * 9b29157 Add open3_detach to gemspec file list Daha sonra, “the parent of HEAD” mənasını verən HEAD^ göstərərək əvvəlki commit-i görə bilərsiniz: $ git show HEAD^ commit d921970aadf03b3cf0e71becdaab3147ba71cdef Merge: 1c002dd... 35cfb2b... Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' Note Craet-i Windows-dan xilas etmək Windows-da cmd.exe , ^ xüsusi bir xarakter daşıyır və fərqli davranılmalıdır. Ya ikiqat edə bilərsiniz, ya da commit arayışını quote-lara daxil edə bilərsiniz: $ git show HEAD^ # will NOT work on Windows $ git show HEAD^^ # OK $ git show "HEAD^" # OK İstədiyiniz valideynin hansı olduğunu müəyyən etmək üçün ^ -dən sonra bir rəqəm də göstərə bilərsiniz; məsələn, d921970^2 “d921970-in ikinci valideynidir.” deməkdir. Bu sintaksis yalnız birdən çox valideynə sahib olan birləşmə commit-ləri üçün faydalıdır - birləşdirmə commit-inin birinci valideyni birləşdikdə olduğunuz branch-dan (tez-tez master ), birləşmə commit-nin ikinci valideyn hissəsi isə birləşdirilmiş branch-dan ( topic deyək): $ git show d921970^ commit 1c002dd4b536e7479fe34593e72e6c6c1819e53b Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 14:58:32 2008 -0800 Add some blame and merge stuff $ git show d921970^2 commit 35cfb2b795a55793d7cc56a6cc2060b4bb732548 Author: Paul Hedderly <paul+git@mjr.org> Date: Wed Dec 10 22:22:03 2008 +0000 Some rdoc changes Digər əsas əcdad spesifikasiyası ~ (tilde)-dir. Bu da birinci valideynə aiddir, buna görə HEAD~ və HEAD^ bərabərdir. Fərq bir rəqəm göstərdiyiniz zaman aydın olur. HEAD~2 , “ilk valideynin birinci valideyni” və ya “nənə və baba” deməkdir - ilk valideynlərə göstərdiyiniz vaxt keçir. Məsələn, əvvəllər sadalanan tarixdə HEAD~3 : $ git show HEAD~3 commit 1c3618887afb5fbcbea25b7c013f4e2114448b8d Author: Tom Preston-Werner <tom@mojombo.com> Date: Fri Nov 7 13:47:59 2008 -0500 Ignore *.gem Yenidən ilk valideynin ilk valideyninin ilk valideyni olan HEAD~~~ yazıla bilər: $ git show HEAD~~~ commit 1c3618887afb5fbcbea25b7c013f4e2114448b8d Author: Tom Preston-Werner <tom@mojombo.com> Date: Fri Nov 7 13:47:59 2008 -0500 Ignore *.gem Bu sintaksisləri də birləşdirə bilərsiniz - əvvəlcədən istinadın ikinci əsas hissəsini (birləşdirmə əmri olduğunu düşünərək) HEAD~3^2 və s. istifadə edərək əldə edə bilərsiniz. Commit Aralıqları İndi fərdi commit-lər təyin edə bildiyinizə görə, commit-lərin hüdudlarını necə təyin edəcəyimizə baxaq. Bu, branch-larınızı idarə etmək üçün xüsusilə faydalıdır - çox sayda branch-ınız varsa, “Bu branch-da hələ əsas branch-a birləşdirmədiyim hansı iş var?” kimi suallara cavab vermək üçün spesifikasiyalardan istifadə edə bilərsiniz. Cüt nöqtə Ən geniş yayılmış spesifikasiya cüt nöqtəli sintaksisdir. Bu, əsasən Git-dən bir commit-dən əldə edilə bilən, digərinə çatmayan bir sıra commit-ləri həll etməsini xahiş edir. Məsələn, Aralıq seçimi üçün nümunə tarixçəsi kimi görünən commit tarixçəniz olduğunu söyləyin. Figure 137. Aralıq seçimi üçün nümunə tarixçəsi experiment branch-ınızda hələ master branch-ınıza birləşdirilməyənləri görmək istədiyinizi söyləyin. Git-dən sizə yalnız master..experiment ilə işləyənlərin bir jurnalını göstərməsini xahiş edə bilərsiniz - bu, ‘` master -dən əldə edilə bilməyən experiment -dən əldə edilə bilən bütün commit-lər’' deməkdir. Bu nümunələrdə qısalıq və aydınlıq üçün, diaqramdakı commit obyektlərinin hərfləri göstərəcəkləri qaydada həqiqi log çıxışı yerinə istifadə olunur: $ git log master..experiment D C Digər tərəfdən bunun əksini görmək istəyirsinizsə - bütün commit-lər experiment -də olmayan master -də işləyirsə - branch adlarını tərsinə çevirə bilərsiniz. experiment..master sizə experiment -dən əlçatmaz olan hər şeyi master -də göstərir: $ git log experiment..master F E Bu, experiment branch-ını yeniləmək və birləşdirmək istədiklərinizi önizləmək istəsəniz faydalıdır. Bu sintaksisin başqa bir tez-tez istifadəsi uzaq məsafəyə nəyi push edəcəyinizi görməkdir: $ git log origin/master..HEAD Bu əmr sizə cari branch-ınızdakı origin remote-dakı master branch-ında olmayan hər hansı bir commit-i göstərir. Bir git push işə salırsınızsa və mövcud branch-ınız origin/master izləyirsə, git log origin/master..HEAD tərəfindən sadalanan commit-lər serverə ötürülən commit-lərdir. Git-in HEAD olduğunu qəbul etməsi üçün sintaksisin bir tərəfini də tərk edə bilərsiniz. Məsələn, git log origin/master.. yazaraq əvvəlki nümunədəki ilə eyni nəticələr əldə edə bilərsiniz - bir tərəfi yoxdursa, HEAD əvəzlə. Birdən Çox Pal İkili nöqtəli sintaksis shorthand kimi faydalıdır, lakin bəlkə də hazırda olduğunuz branch-da olmayan bir neçə branch-dan birinin nə olduğunu görmək kimi düzəlişlərinizi göstərmək üçün ikidən çox branch göstərmək istəyirsiniz. Git, əlçatan commit-lər görmək istəmədiyiniz hər hansı bir istinaddan əvvəl ^ simvolunu və ya --not istifadə edərək bunu etməyə imkan verir. Beləliklə, aşağıdakı üç əmr bərabərdir: $ git log refA..refB $ git log ^refA refB $ git log refB --not refA Bu çox yaxşıdır, çünki bu sintaksislə sorğunuzda ikiqat nöqtəli sintaksis ilə edə bilməyəcəyiniz ikidən çox istinad daxil edə bilərsiniz. Məsələn, refA ya da refB -dən əldə edilə bilən, ancaq refC -dən edilməyən, bütün commit-ləri görmək istəyirsinizsə, aşağıdakılardan birini istifadə edə bilərsiniz: $ git log refA refB ^refC $ git log refA refB --not refC Bu, branch-larınızda nə olduğunu anlamanıza kömək edəcək çox güclü bir revizyon sorğu sistemi yaradır. Üçqat Nöqtə Son böyük aralıq seçmə sintaksisi, hər ikisindən deyil, iki istinadın hər ikisi tərəfindən də əldə edilə bilən bütün commit-ləri təyin edən üç nöqtəli sintaksisdir. Aralıq seçimi üçün nümunə tarixçəsi -dakı commit tarixçəsinə baxın. master və ya experiment -də olanları görmək istəsəniz, lakin ümumi istinadları yox görmək istəməsəniz, işlədə bilərsiniz: $ git log master...experiment F E D C Yenə də, bu sizə normal bir log çıxışı verir, ancaq ənənəvi commit tarixi sifarişində görünən yalnız bu dörd commit üçün commit məlumatlarını göstərir. Bu halda log əmri ilə istifadə olunan ümumi bir keçid, hər bir commit aralığın hansı tərəfində olduğunu göstərən --left-right -dır. Bu, nəticənin daha faydalı olmasına kömək edir: $ git log --left-right master...experiment < F < E > D > C Bu vasitələrlə Git-ə nəyi yoxlamaq istədiyinizi və ya commit-lərinizi daha asanlıqla bildirə bilərsiniz. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/es/v2/Fundamentos-de-Git-Trabajar-con-Remotos
Git - Trabajar con Remotos About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Inicio - Sobre el Control de Versiones 1.1 Acerca del Control de Versiones 1.2 Una breve historia de Git 1.3 Fundamentos de Git 1.4 La Línea de Comandos 1.5 Instalación de Git 1.6 Configurando Git por primera vez 1.7 ¿Cómo obtener ayuda? 1.8 Resumen 2. Fundamentos de Git 2.1 Obteniendo un repositorio Git 2.2 Guardando cambios en el Repositorio 2.3 Ver el Historial de Confirmaciones 2.4 Deshacer Cosas 2.5 Trabajar con Remotos 2.6 Etiquetado 2.7 Alias de Git 2.8 Resumen 3. Ramificaciones en Git 3.1 ¿Qué es una rama? 3.2 Procedimientos Básicos para Ramificar y Fusionar 3.3 Gestión de Ramas 3.4 Flujos de Trabajo Ramificados 3.5 Ramas Remotas 3.6 Reorganizar el Trabajo Realizado 3.7 Recapitulación 4. Git en el Servidor 4.1 Los Protocolos 4.2 Configurando Git en un servidor 4.3 Generando tu clave pública SSH 4.4 Configurando el servidor 4.5 El demonio Git 4.6 HTTP Inteligente 4.7 GitWeb 4.8 GitLab 4.9 Git en un alojamiento externo 4.10 Resumen 5. Git en entornos distribuidos 5.1 Flujos de trabajo distribuidos 5.2 Contribuyendo a un Proyecto 5.3 Manteniendo un proyecto 5.4 Resumen 6. GitHub 6.1 Creación y configuración de la cuenta 6.2 Participando en Proyectos 6.3 Mantenimiento de un proyecto 6.4 Gestión de una organización 6.5 Scripting en GitHub 6.6 Resumen 7. Herramientas de Git 7.1 Revisión por selección 7.2 Organización interactiva 7.3 Guardado rápido y Limpieza 7.4 Firmando tu trabajo 7.5 Buscando 7.6 Reescribiendo la Historia 7.7 Reiniciar Desmitificado 7.8 Fusión Avanzada 7.9 Rerere 7.10 Haciendo debug con Git 7.11 Submódulos 7.12 Agrupaciones 7.13 Replace 7.14 Almacenamiento de credenciales 7.15 Resumen 8. Personalización de Git 8.1 Configuración de Git 8.2 Git Attributes 8.3 Puntos de enganche en Git 8.4 Un ejemplo de implantación de una determinada política en Git 8.5 Recapitulación 9. Git y Otros Sistemas 9.1 Git como Cliente 9.2 Migración a Git 9.3 Resumen 10. Los entresijos internos de Git 10.1 Fontanería y porcelana 10.2 Los objetos Git 10.3 Referencias Git 10.4 Archivos empaquetadores 10.5 Las especificaciones para hacer referencia a…​ (refspec) 10.6 Protocolos de transferencia 10.7 Mantenimiento y recuperación de datos 10.8 Variables de entorno 10.9 Recapitulación A1. Apéndice A: Git en otros entornos A1.1 Interfaces gráficas A1.2 Git en Visual Studio A1.3 Git en Eclipse A1.4 Git con Bash A1.5 Git en Zsh A1.6 Git en Powershell A1.7 Resumen A2. Apéndice B: Integrando Git en tus Aplicaciones A2.1 Git mediante Línea de Comandos A2.2 Libgit2 A2.3 JGit A3. Apéndice C: Comandos de Git A3.1 Configuración A3.2 Obtener y Crear Proyectos A3.3 Seguimiento Básico A3.4 Ramificar y Fusionar A3.5 Compartir y Actualizar Proyectos A3.6 Inspección y Comparación A3.7 Depuración A3.8 Parcheo A3.9 Correo Electrónico A3.10 Sistemas Externos A3.11 Administración A3.12 Comandos de Fontanería 2nd Edition 2.5 Fundamentos de Git - Trabajar con Remotos Trabajar con Remotos Para poder colaborar en cualquier proyecto Git, necesitas saber cómo gestionar repositorios remotos. Los repositorios remotos son versiones de tu proyecto que están hospedadas en Internet o en cualquier otra red. Puedes tener varios de ellos, y en cada uno tendrás generalmente permisos de solo lectura o de lectura y escritura. Colaborar con otras personas implica gestionar estos repositorios remotos enviando y trayendo datos de ellos cada vez que necesites compartir tu trabajo. Gestionar repositorios remotos incluye saber cómo añadir un repositorio remoto, eliminar los remotos que ya no son válidos, gestionar varias ramas remotas, definir si deben rastrearse o no y más. En esta sección, trataremos algunas de estas habilidades de gestión de remotos. Ver Tus Remotos Para ver los remotos que tienes configurados, debes ejecutar el comando git remote . Mostrará los nombres de cada uno de los remotos que tienes especificados. Si has clonado tu repositorio, deberías ver al menos origin (origen, en inglés) - este es el nombre que por defecto Git le da al servidor del que has clonado: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin También puedes pasar la opción -v , la cual muestra las URLs que Git ha asociado al nombre y que serán usadas al leer y escribir en ese remoto: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Si tienes más de un remoto, el comando los listará todos. Por ejemplo, un repositorio con múltiples remotos para trabajar con distintos colaboradores podría verse de la siguiente manera. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Esto significa que podemos traer contribuciones de cualquiera de estos usuarios fácilmente. Es posible que también tengamos permisos para enviar datos a algunos, aunque no podemos saberlo desde aquí. Fíjate que estos remotos usan distintos protocolos; hablaremos sobre ello más adelante, en Configurando Git en un servidor . Añadir Repositorios Remotos En secciones anteriores hemos mencionado y dado alguna demostración de cómo añadir repositorios remotos. Ahora veremos explícitamente cómo hacerlo. Para añadir un remoto nuevo y asociarlo a un nombre que puedas referenciar fácilmente, ejecuta git remote add [nombre] [url] : $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) A partir de ahora puedes usar el nombre pb en la línea de comandos en lugar de la URL entera. Por ejemplo, si quieres traer toda la información que tiene Paul pero tú aún no tienes en tu repositorio, puedes ejecutar git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit La rama maestra de Paul ahora es accesible localmente con el nombre pb/master - puedes combinarla con alguna de tus ramas, o puedes crear una rama local en ese punto si quieres inspeccionarla. (Hablaremos con más detalle acerca de qué son las ramas y cómo utilizarlas en [ch03-git-branching] .) Traer y Combinar Remotos Como hemos visto hasta ahora, para obtener datos de tus proyectos remotos puedes ejecutar: $ git fetch [remote-name] El comando irá al proyecto remoto y se traerá todos los datos que aun no tienes de dicho remoto. Luego de hacer esto, tendrás referencias a todas las ramas del remoto, las cuales puedes combinar e inspeccionar cuando quieras. Si clonas un repositorio, el comando de clonar automáticamente añade ese repositorio remoto con el nombre “origin”. Por lo tanto, git fetch origin se trae todo el trabajo nuevo que ha sido enviado a ese servidor desde que lo clonaste (o desde la última vez que trajiste datos). Es importante destacar que el comando git fetch solo trae datos a tu repositorio local - ni lo combina automáticamente con tu trabajo ni modifica el trabajo que llevas hecho. La combinación con tu trabajo debes hacerla manualmente cuando estés listo. Si has configurado una rama para que rastree una rama remota (más información en la siguiente sección y en [ch03-git-branching] ), puedes usar el comando git pull para traer y combinar automáticamente la rama remota con tu rama actual. Es posible que este sea un flujo de trabajo mucho más cómodo y fácil para ti; y por defecto, el comando git clone le indica automáticamente a tu rama maestra local que rastree la rama maestra remota (o como se llame la rama por defecto) del servidor del que has clonado. Generalmente, al ejecutar git pull traerás datos del servidor del que clonaste originalmente y se intentará combinar automáticamente la información con el código en el que estás trabajando. Enviar a Tus Remotos Cuando tienes un proyecto que quieres compartir, debes enviarlo a un servidor. El comando para hacerlo es simple: git push [nombre-remoto] [nombre-rama] . Si quieres enviar tu rama master a tu servidor origin (recuerda, clonar un repositorio establece esos nombres automáticamente), entonces puedes ejecutar el siguiente comando y se enviarán todos los commits que hayas hecho al servidor: $ git push origin master Este comando solo funciona si clonaste de un servidor sobre el que tienes permisos de escritura y si nadie más ha enviado datos por el medio. Si alguien más clona el mismo repositorio que tú y envía información antes que tú, tu envío será rechazado. Tendrás que traerte su trabajo y combinarlo con el tuyo antes de que puedas enviar datos al servidor. Para información más detallada sobre cómo enviar datos a servidores remotos, véase [ch03-git-branching] . Inspeccionar un Remoto Si quieres ver más información acerca de un remoto en particular, puedes ejecutar el comando git remote show [nombre-remoto] . Si ejecutas el comando con un nombre en particular, como origin , verás algo como lo siguiente: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) El comando lista la URL del repositorio remoto y la información del rastreo de ramas. El comando te indica claramente que si estás en la rama maestra y ejecutas el comando git pull , automáticamente combinará la rama maestra remota con tu rama local, luego de haber traído toda la información de ella. También lista todas las referencias remotas de las que ha traído datos. Ejemplos como este son los que te encontrarás normalmente. Sin embargo, si usas Git de forma más avanzada, puede que obtengas mucha más información de un git remote show : $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Este comando te indica a cuál rama enviarás información automáticamente cada vez que ejecutas git push , dependiendo de la rama en la que estés. También te muestra cuáles ramas remotas no tienes aún, cuáles ramas remotas tienes que han sido eliminadas del servidor, y varias ramas que serán combinadas automáticamente cuando ejecutes git pull . Eliminar y Renombrar Remotos Si quieres cambiar el nombre de la referencia de un remoto puedes ejecutar git remote rename . Por ejemplo, si quieres cambiar el nombre de pb a paul , puedes hacerlo con git remote rename : $ git remote rename pb paul $ git remote origin paul Es importante destacar que al hacer esto también cambias el nombre de las ramas remotas. Por lo tanto, lo que antes estaba referenciado como pb/master ahora lo está como paul/master . Si por alguna razón quieres eliminar un remoto - has cambiado de servidor o no quieres seguir utilizando un mirror o quizás un colaborador ha dejado de trabajar en el proyecto - puedes usar git remote rm : $ git remote rm paul $ git remote origin prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/fr/v2/Les-bases-de-Git-%c3%89tiquetage
Git - Étiquetage About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Démarrage rapide 1.1 À propos de la gestion de version 1.2 Une rapide histoire de Git 1.3 Rudiments de Git 1.4 La ligne de commande 1.5 Installation de Git 1.6 Paramétrage à la première utilisation de Git 1.7 Obtenir de l’aide 1.8 Résumé 2. Les bases de Git 2.1 Démarrer un dépôt Git 2.2 Enregistrer des modifications dans le dépôt 2.3 Visualiser l’historique des validations 2.4 Annuler des actions 2.5 Travailler avec des dépôts distants 2.6 Étiquetage 2.7 Les alias Git 2.8 Résumé 3. Les branches avec Git 3.1 Les branches en bref 3.2 Branches et fusions : les bases 3.3 Gestion des branches 3.4 Travailler avec les branches 3.5 Branches de suivi à distance 3.6 Rebaser (Rebasing) 3.7 Résumé 4. Git sur le serveur 4.1 Protocoles 4.2 Installation de Git sur un serveur 4.3 Génération des clés publiques SSH 4.4 Mise en place du serveur 4.5 Démon (Daemon) Git 4.6 HTTP intelligent 4.7 GitWeb 4.8 GitLab 4.9 Git hébergé 4.10 Résumé 5. Git distribué 5.1 Développements distribués 5.2 Contribution à un projet 5.3 Maintenance d’un projet 5.4 Résumé 6. GitHub 6.1 Configuration et paramétrage d’un compte 6.2 Contribution à un projet 6.3 Maintenance d’un projet 6.4 Gestion d’un regroupement 6.5 Écriture de scripts pour GitHub 6.6 Résumé 7. Utilitaires Git 7.1 Sélection des versions 7.2 Indexation interactive 7.3 Remisage et nettoyage 7.4 Signer votre travail 7.5 Recherche 7.6 Réécrire l’historique 7.7 Reset démystifié 7.8 Fusion avancée 7.9 Rerere 7.10 Déboguer avec Git 7.11 Sous-modules 7.12 Empaquetage (bundling) 7.13 Replace 7.14 Stockage des identifiants 7.15 Résumé 8. Personnalisation de Git 8.1 Configuration de Git 8.2 Attributs Git 8.3 Crochets Git 8.4 Exemple de politique gérée par Git 8.5 Résumé 9. Git et les autres systèmes 9.1 Git comme client 9.2 Migration vers Git 9.3 Résumé 10. Les tripes de Git 10.1 Plomberie et porcelaine 10.2 Les objets de Git 10.3 Références Git 10.4 Fichiers groupés 10.5 La refspec 10.6 Les protocoles de transfert 10.7 Maintenance et récupération de données 10.8 Les variables d’environnement 10.9 Résumé A1. Annexe A: Git dans d’autres environnements A1.1 Interfaces graphiques A1.2 Git dans Visual Studio A1.3 Git dans Visual Studio Code A1.4 Git dans IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git dans Sublime Text A1.6 Git dans Bash A1.7 Git dans Zsh A1.8 Git dans PowerShell A1.9 Résumé A2. Annexe B: Embarquer Git dans vos applications A2.1 Git en ligne de commande A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Commandes Git A3.1 Installation et configuration A3.2 Obtention et création des projets A3.3 Capture d’instantané basique A3.4 Création de branches et fusion A3.5 Partage et mise à jour de projets A3.6 Inspection et comparaison A3.7 Débogage A3.8 Patchs A3.9 Courriel A3.10 Systèmes externes A3.11 Administration A3.12 Commandes de plomberie 2nd Edition 2.6 Les bases de Git - Étiquetage Étiquetage À l’instar de la plupart des VCS, Git donne la possibilité d’étiqueter un certain état dans l’historique comme important. Généralement, les gens utilisent cette fonctionnalité pour marquer les états de publication ( v1.0 et ainsi de suite). Dans cette section, nous apprendrons comment lister les différentes étiquettes ( tag en anglais), comment créer de nouvelles étiquettes et les différents types d’étiquettes. Lister vos étiquettes Lister les étiquettes existantes dans Git est très simple. Tapez juste git tag  : $ git tag v0.1 v1.3 Cette commande liste les étiquettes dans l’ordre alphabétique. L’ordre dans lequel elles apparaissent n’a aucun rapport avec l’historique. Vous pouvez aussi rechercher les étiquettes correspondant à un motif particulier. Par exemple, le dépôt des sources de Git contient plus de 500 étiquettes. Si vous souhaitez ne visualiser que les séries 1.8.5, vous pouvez lancer ceci : $ git tag -l 'v1.8.5*' v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Note Lister les étiquettes avec des jokers nécessite les options -l ou --list Si vous voulez juste la liste complète des étiquettes, la commande git tag considère implicitement que vous souhaitez une liste et elle la fournit ; l’utilisation de -l ou --list est optionnelle dans ce cas. Cependant, si vous fournissez un motif joker pour filtrer les d’étiquettes, l’utilisation de -l ou --list est obligatoire. Créer des étiquettes Git utilise deux types principaux d’étiquettes : légères et annotées. Une étiquette légère ressemble beaucoup à une branche qui ne change pas, c’est juste un pointeur sur un commit spécifique. Les étiquettes annotées, par contre, sont stockées en tant qu’objets à part entière dans la base de données de Git. Elles ont une somme de contrôle, contiennent le nom et l’adresse e-mail du créateur, la date, un message d’étiquetage et peuvent être signées et vérifiées avec GNU Privacy Guard (GPG). Il est généralement recommandé de créer des étiquettes annotées pour générer toute cette information mais si l’étiquette doit rester temporaire ou l’information supplémentaire n’est pas désirée, les étiquettes légères peuvent suffire. Les étiquettes annotées Créer des étiquettes annotées est simple avec Git. Le plus simple est de spécifier l’option -a à la commande tag  : $ git tag -a v1.4 -m 'ma version 1.4' $ git tag v0.1 v1.3 v1.4 L’option -m permet de spécifier le message d’étiquetage qui sera stocké avec l’étiquette. Si vous ne spécifiez pas de message en ligne pour une étiquette annotée, Git lance votre éditeur pour pouvoir le saisir. Vous pouvez visualiser les données de l’étiquette à côté du commit qui a été marqué en utilisant la commande git show  : $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 ma version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Cette commande affiche le nom du créateur, la date de création de l’étiquette et le message d’annotation avant de montrer effectivement l’information de validation. Les étiquettes légères Une autre manière d’étiqueter les commits est d’utiliser les étiquettes légères. Celles-ci se réduisent à stocker la somme de contrôle d’un commit dans un fichier, aucune autre information n’est conservée. Pour créer une étiquette légère, il suffit de n’utiliser aucune des options -a , -s ou -m  : $ git tag v1.4-lg $ git tag v0.1 v1.3 v1.4 v1.4-lg v1.5 Cette fois-ci, en lançant git show sur l’étiquette, on ne voit plus aucune information complémentaire. La commande ne montre que l’information de validation : $ git show v1.4-lg commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Étiqueter après coup Vous pouvez aussi étiqueter des commits plus anciens. Supposons que l’historique des commits ressemble à ceci : $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Fusion branche 'experimental' a6b4c97498bd301d84096da251c98a07c7723e65 Début de l'écriture support 0d52aaab4479697da7686c15f77a3d64d9165190 Un truc de plus 6d52a271eda8725415634dd79daabbc4d9b6008e Fusion branche 'experimental' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc ajout d'une fonction de validation 4682c3261057305bdd616e23b64b0857d832627b ajout fichier a_faire 166ae0c4d3f420721acbb115cc33848dfcc2121a début de l'écriture support 9fceb02d0ae598e95dc970b74767f19372d61af8 mise à jour rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc validation a_faire 8a5cbc430f1a9c3d00faaeffd07798508422908a mise à jour lisezmoi Maintenant, supposons que vous avez oublié d’étiqueter le projet à la version v1.2 qui correspondait au commit « mise à jour rakefile ». Vous pouvez toujours le faire après l’évènement. Pour étiqueter ce commit , vous spécifiez la somme de contrôle du commit (ou une partie) en fin de commande : $ git tag -a v1.2 9fceb02 Le commit a été étiqueté : $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lg v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... Partager les étiquettes Par défaut, la commande git push ne transfère pas les étiquettes vers les serveurs distants. Il faut explicitement pousser les étiquettes après les avoir créées localement. Ce processus s’apparente à pousser des branches distantes — vous pouvez lancer git push origin [nom-du-tag] . $ git push origin v1.5 Décompte des objets: 14, fait. Delta compression using up to 8 threads. Compression des objets: 100% (12/12), fait. Écriture des objets: 100% (14/14), 2.05KiB | 0 bytes/s, fait. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Si vous avez de nombreuses étiquettes que vous souhaitez pousser en une fois, vous pouvez aussi utiliser l’option --tags avec la commande git push . Ceci transférera toutes les nouvelles étiquettes vers le serveur distant. $ git push origin --tags Décompte des objets: 1, fait. Écriture des objets: 100% (1/1), 160 bytes | 0 bytes/s, fait. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lg -> v1.4-lg À présent, lorsqu’une autre personne clone ou tire depuis votre dépôt, elle obtient aussi les étiquettes. Note git push pousse les deux types d’étiquettes git push <distant> --tags pousse à la fois les étiquettes légères et annotées. Il n’y a actuellement aucune option pour pousser seulement les étiquettes légères, mais si vous utilisez git push <distant> --follow-tags , seules les étiquettes annotées seront poussées sur le serveur distant. Supprimer les étiquettes Pour supprimer une étiquette de votre dépôt local, vous pouvez utiliser git tag -d <nom-d-etiquette> . Par example, nous pourrions supprimer notre étiquette légère ci-dessus, comme ceci : $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Notez que ceci ne supprime pas l’étiquette sur aucun serveur distant. Il y a deux méthodes communes pour supprimer une étiquette d’un serveur distant. La première est git push <distant> :refs/tags/<nom-d-etiquette> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw Ceci s’interprète comme une valeur nulle devant les deux points qui est envoyée sur le nom d’étiquette distante, ce qui l’efface. La seconde manière (et la plus intuitive) pour supprimer une étiquette distante utilise l’option --delete  : $ git push origin --delete <nom-d-etiquette> Extraire une étiquette Si vous souhaitez voir les versions de fichiers qu’une étiquette pointe, vous pouvez faire un git checkout de cette étiquette, bien que cela positionne votre dépôt dans un état « HEAD détachée », ce qui a quelques effets de bords malheureux : git checkout v2.29.2 Note : basculement sur 'v2.29.2'. Vous êtes dans l'état « HEAD détachée ». Vous pouvez visiter, faire des modifications expérimentales et les valider. Il vous suffit de faire un autre basculement pour abandonner les commits que vous faites dans cet état sans impacter les autres branches Si vous voulez créer une nouvelle branche pour conserver les commits que vous créez, il vous suffit d'utiliser l'option -c de la commande switch comme ceci : git switch -c <nom-de-la-nouvelle-branche> Ou annuler cette opération avec : git switch - Désactivez ce conseil en renseignant la variable de configuration advice.detachedHead à false HEAD est maintenant sur 898f80736c Git 2.29.2 $ git checkout v2.29.1 La position précédente de HEAD était sur 898f80736c Git 2.29.2 HEAD est maintenant sur b927c80531 Git 2.29.1 Dans l’état « HEAD détachée », si vous modifiez puis créez un commit, l’étiquette restera identique, mais votre nouveau commit n’appartiendra à aucune branche et sera non joignable, à part avec son empreinte de commit exacte. Ainsi, si vous avez besoin de faire des modifications — disons que vous corrigez un bogue d’une ancienne version, par exemple — vous voudrez généralement créer une branche : $ git checkout -b v2.9.X Basculement sur la nouvelle branche 'v2.9.X' Si vous faites ceci et que vous faites un commit, votre branche V2.9.X sera légèrement différente de votre étiquette v2.9.1 puisqu’elle aura avancé avec les modifications que vous y aurez intégrées, donc faites attention. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://l.facebook.com/l.php?u=https%3A%2F%2Fwww.instagram.com%2F&h=AT2Od7sSHri2WsTQOIvSnShCby5NptWdK78Pu-XcuAdwLs5Ne4OnT7-VVZRLdhEywVbz5jA2DZBx5-vWrIiBFuExWanot1SiAVerc6685pg467O7NXMMbSUyWYSrZhL6HqrUSIvSenfeJMQt
Facebook Facebook 이메일 또는 휴대폰 비밀번호 계정을 잊으셨나요? 새 계정 만들기 일시적으로 차단됨 일시적으로 차단됨 회원님의 이 기능 사용 속도가 너무 빠른 것 같습니다. 이 기능 사용에서 일시적으로 차단되었습니다. Back 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch 가입하기 로그인 Messenger Facebook Lite 동영상 Meta Pay Meta 스토어 Meta Quest Ray-Ban Meta Meta AI Meta AI 콘텐츠 더 보기 Instagram Threads 투표 정보 센터 개인정보처리방침 개인정보 보호 센터 정보 광고 만들기 페이지 만들기 개발자 채용 정보 쿠키 AdChoices 이용 약관 고객 센터 연락처 업로드 및 비사용자 설정 활동 로그 Meta © 2026
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Gedistribueerd-Git-Samenvatting
Git - Samenvatting About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 5.4 Gedistribueerd Git - Samenvatting Samenvatting Je zou je nu redelijk op je gemak moeten voelen om aan een project bij te dragen met Git, maar ook om je eigen project te beheren of de bijdragen van andere gebruikers te integreren. Gefeliciteerd, je bent nu een effectieve Git ontwikkelaar! In het volgende hoofdstuk zullen we je vertellen hoe de grootste en meest populaire Git hosting service te gebruiken: GitHub. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/workbook/part-I-foundations/
Google SRE - Tools for Building Reliable Systems Part I - Foundations Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Part I. Foundations Every implementation guide needs to start with a common base from which to build. In this case, the basic foundations of SRE include SLOs, monitoring, alerting, toil reduction, and simplicity . Getting these basics right will set you up well to succeed on your SRE journey. The following chapters explore techniques for turning these core principles into concrete practices for your organization. Previous Chapter 1 - How SRE Relates to DevOps Next Chapter 2 - Implementing SLOs Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://git-scm.com/book/uz/v2/Git-%d0%b0%d1%81%d0%be%d1%81%d0%bb%d0%b0%d1%80%d0%b8-Git-%d0%be%d0%bc%d0%b1%d0%be%d1%80%d0%b8%d0%bd%d0%b8-%d1%8f%d1%80%d0%b0%d1%82%d0%b8%d1%88
Git - Git омборини яратиш About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Иш бошланиши 1.1 Талқинларни бошқариш ҳақида 1.2 Git нинг қисқача тарихи 1.3 Git асоси 1.4 Командалар сатри 1.5 Git ни ўрнатиш 1.6 Git да биринчи созлашлар 1.7 Қандай ёрдам олиш мумкин? 1.8 Хулосалар 2. Git асослари 2.1 Git омборини яратиш 2.2 Ўзгаришларни омборга ёзиш 2.3 Фиксирлашлар тарихини кўриш 2.4 Ўзгаришларни бекор қилиш 2.5 Узоқ масофадаги омборлар билан ишлаш 2.6 Тамғалаш 2.7 Git да таҳаллуслар 2.8 Хулоса 3. Git да тармоқланиш 3.1 Тармоқланиш ҳақида икки оғиз сўз 3.2 Тармоқланиш ва бирлашиш асослари 3.3 Тармоқларни бошқариш 3.4 Иш жараёнларини тармоқлаш 3.5 Узоқ масофадаги тармоқлар 3.6 Қайта асосланиш 3.7 Хулосалар 4. Git серверда 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Sizning SSH ochiq (public) kalitingizni generatsiyalash 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Хулосалар 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Qism modullar (Submodule) 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.1 Git асослари - Git омборини яратиш Агар сиз фақат битта бўлимни ўқиб Git билан ишлашни бошламоқчи бўлсангиз у ҳолда бу бўлим айнан сиз учун. Бу ерда Git билан ишлаш жараёнида келиб чиқадиган кўпчилик муаммоларни ҳал қилиш учун керак бўладиган барча асосий командалар кўриб чиқилган. Бу бўлимни ўқиб чиққач сиз омборни созлаш ва қийматлашни, файлларнинг талқинларини бошқаришни бошлаш ва тугатишни, ўзгаришларни тайёрлаш ва фиксирлашни билиб оласиз. Биз шунингдек Git да алоҳида файл ёки уларнинг гуруҳларини қандай қилиб ташлаб кетишни, қандай қилиб тез ва оддийгина йўл билан нотўғри ўзгаришларни бекор қилишни, қандай қилиб лойиҳангиз ва алоҳида юборилганлар (commit) орасидаги тарихни кўришни, қандай қилиб узоқ масофадаги омборга(дан) ўзгаришларни қўшиш(push) (ва олиш(pull))ларни намойиш қилиб берамиз. Git омборини яратиш Git омборини яратиш учун иккита асосий ёндашув мавжуд. Биринчи ёндашув – мавжуд кталог ёки лойиҳани импорт қилиш. Иккинчиси – серверда мавжуд бўлган омборни клонлаштириш. Мавжуд бўлган каталогда омборни яратиш Агар сиз Git ни мавжд бўлган лойиҳа учун ишлатмоқчи бўлсангиз у ҳолда сизга лойиҳа каталогига ўтиб ва командалар сатрида қуйидагини териш зарур $ git init Ушбу команда жорий каталогда ўз ичига Git омборига асосланган – омбор учун керак бўладиган янги .git номли қисм каталогини яратади. Бу босқичда сизнинг лойиҳангиз ҳали – ҳамон талқинлар бошқарувида эмас. ( Git Internals бўлимда хозиргина яратилган .git каталог ичидаги файллар ҳақида батафсил ёзилган.) Агар сиз талқинлар бошқарувига мавжуд бўлган файлни қўшмоқчи бўлсангиз сиз ушбу файларни индекслаб ва биринчи навбатда ўзгаришларни фиксирлашингизга тўғри келади. Буни амалга ошириш учун git add да индексланадиган файлларини кўрсатиб, сўнгра юборишни ( commit ) кўрсатувчи бир қанча командалардан фойдаланишингиз мумкин: $ git add *.c $ git add LICENSE $ git commit -m 'initial project version' Биз ушбу командалар нима иш қилаётганини сал кейинроқ таҳлил этамиз. Айни вақтдаги босқичда файллар қўшилган ва бошланғич коммитли сизнинг Git омборингиз мавжуд. Мавжуд омборни клонлаштириш Агар сиз мавжуд бўлган омборни нусхасини олишни хоҳласангиз, масалан лойиҳани олишни хоҳласангиз, у ҳолда сизга git clone керак бўлади. Агар сиз Subversion каби бошқа талқинларни бошқариш тизими билан таниш бўлсангиз шунга эътибор берингки команда checkout эмас балки clone. Бу муҳим фарқ – Git амалда сервердаги барча маълумотлар нусхасини олади. git clone – бажарилганда ҳар бир файлнинг ҳар бир талқини сервердаги лойиҳа тарихидан олинади(pulled). Агар сервер диски ишдан чиқса сиз клиентлардаги ихтиёрий клонлардан фойдаланиб, серверни клонлаштириш вақтидаги ҳолатига қайтаришингиз мумкин бўлади (сиз сервернинг ушлаб олгичларининг қисмларини (server-side hooks) йўқотишингиз мумкин, лекин ҳаммасини эмас. Талқинлар бошқарувига қўшилганлари сақланиб қолади. Бу ҳақда Getting Git on a Server бўлимга қаранг.). Омборни клонлаштириш git clone [url] командаси орқали амалга оширилади. Масалан, агар сиз libgit2 кутубхонасини клонлаштирмоқчи бўлсангиз уни қуйидагича кўринишда бажаришингиз мумкин: $ git clone https://github.com/libgit2/libgit2 Ушбу буйруқ “libgit2” номли каталогни яратади, унинг ичида .git каталогини яратиб унда яратиш ва қийматлаш ишларини бажаради (инициализациялайди), ушбу омбор учун барча маълумотларни юклаб олади ва омбор учун охирги талқинни яратади (checks out). Агар сиз янги libgit2 каталогига кирсангиз сиз унда ишлатиш ва қўллаш учун ярайдиган лойиҳа файлларини кўрасиз. Агар сиз омборни libgit2 дан бошқа каталогга клонлаштиришни хоҳласангиз буни командалар қаторида қуйидагича кўрсатиш мумкин: $ git clone https://github.com/libgit2/libgit2 mylibgit Ушбу команда бундан олдинги команда билан бир хил иш бажаради ва улар орасидаги фарқ шуки, бунда натижавий каталог номи - mylibgit . Git сиз ишлатишингиз мумкин бўлган бир қанча транспорт протоколларидан фойдалана олади. Аввалги мисолда https://`протоколи ишлатилди. Сиз шунингдек SSH узатиш протоколидан фойдаланувчи `git:// ёки user@server:path/to/repo.git ларни ҳам учратишингиз мумкин. Getting Git on a Server бўлимда сизнинг омборингизга ҳуқуқ яратиш учун серверни созлашнинг ҳамма вариантлари билан биргаликда уларнинг камчиликлари ҳақида маълумотлар келтирилган. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/GitHub-Een-project-onderhouden
Git - Een project onderhouden About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 6.3 GitHub - Een project onderhouden Een project onderhouden Nu we ons op ons gemak voelen met bij te dragen aan een project, laten we het eens van de andere kant bekijken: je eigen project aanmaken, onderhouden en beheren. Een nieuwe repository aanmaken Laten we eens een nieuwe repository aanmaken waarmee we de code van ons project delen. Begin met het klikken op de “New repository” knop aan de rechterkant van het dashboard, of vanaf de {plus} knop in de bovenste toolbar naast je gebruikersnaam zoals te zien is in De “New repository” dropdown. . Figuur 110. Het “Your repositories” gebied. Figuur 111. De “New repository” dropdown. Dit leidt je naar het “new repository” formulier: Figuur 112. Het “new repository” formulier. Alles wat je echt moet doen is je project een naam geven, de overige velden zijn volledig optioneel. Voor nu klik je gewoon de “Create Repository” knop en boem - je hebt een nieuwe repository op GitHub, genaamd <gebruiker>/<project_naam> . Omdat er nog geen code is, zal GitHub je aanwijzigen geven hoe je een gloednieuwe Git repository maakt, of verbindt met een bestaand Git project. We zullen het hier nog niet uitwerken, als je een opfrisser nodig hebt kijk dan nog eens naar Git Basics . Nu je project op GitHub gehost wordt, kan je het URL aan iedereen geven waarmee je je project wilt delen. Elk project op GitHub is toegankelijk via HTTP als https://github.com/<gebruiker>/<project_naam> , en via SSH als git@github.com:<gebruiker>/<project_naam> . Git kan fetchen van en pushen naar beide URLs, maar toegang wordt bepaald op basis van de gebruikersgegevens van de gebruiker die ze benadert. Noot Voor openbare projecten wordt vaak de voorkeur gegeven aan het delen middels de HTTP URL, omdat de gebruiker daarvoor niet perse een GitHub account nodig heeft om het te klonen. Gebruikers moeten wel een account hebben en een geüploade SSH sleutel om je project te benaderen als je ze het SSH URL geeft. De HTTPS variant is exact dezelfde URL die ze in hun browser zouden plakken als ze het project daar zouden willen bekijken. Medewerkers toevoegen Als je met andere mensen werkt die je commit toegang wilt geven, moet je ze als “collaborators” toevoegen. Als Ben, Jeff en Louise allemaal GitHub accounts aanvragen, en je wilt ze push toegang geven op jouw repository, kan je ze toevoegen aan je project. Als je dit doet geeft je ze “push” toegang, wat inhoudt dat ze zowel lees als schrijf toegang hebben tot het project en de Git repository. Klik op de “Settings” link onderaan in de rechter zijkolom. Figuur 113. De repository settings link. Kies daarna “Collaborators” op het menu aan de linker kant. Daarna type je gewoon een gebruikersnaam in het veld en klikt “Add collaborator.” Je kunt dit zo vaak herhalen als je toegang wilt verlenen aan iedereen die je maar wilt. Als je toegang wilt ontzeggen, klik je gewoon de “X” rechts van de betreffende rij. Figuur 114. Repository medewerkers. Pull Requests beheren Nu je een project hebt met wat code erin en misschien zelfs een paar medewerkers die ook push toegang hebben, laten we eens behandelen wat je moet doen als je zelf een Pull Request ontvangt. Pull Requests kunnen komen van een branch in een fork van jouw repository of ze kunnen komen van een andere branch in dezelfde repository. Het enige verschil is dat ze in een fork vaak van mensen komen waar jij niet naar hun branch kan pushen en zij niet naar de jouwe, terwijl bij interne Pull Requests beide partijen over het algemeen de branch kunnen benaderen. Voor de volgende voorbeelden, laten we aannemen dat jij “tonychacon” bent en dat je een nieuwe Arduino code project genaamd “fade” gemaakt hebt. Email berichten Er is nu iemand die een wijziging op je code gemaakt heeft en die je een Pull Request stuurt. Je zou een email bericht moeten krijgen die je over het nieuwe Pull Request bericht en die er ongeveer zo uitziet als Email bericht van een nieuwe Request. . Figuur 115. Email bericht van een nieuwe Request. Er zijn een aantal dingen te zien aan deze email. Het geeft je een kleine diffstat — een lijst met bestanden die gewijzigd zijn in de Pull Request en hoezeer ze zijn gewijzigd. Het geeft je een link naar de Pull Request op GitHub. Het geeft je ook een aantal URLs die je vanaf de commando regel kunt gebruiken. Als je de regel bekijkt waar git pull <url> patch-1 staat, is dit een eenvoudige manier om een remote branch te mergen zonder een remote toe te voegen. We hebben dit kort behandeld in Remote branches uitchecken . Als je wilt, kan je een topic branch maken en ernaar switchen en dit commando aanroepen om de Pull Request wijzigingen erin te mergen. De andere interessante URLs zijn de .diff en .patch URLs, die zoals je zult vermoeden, de unified diff en patch versies van de Pull Request geven. Je zou technisch gesproken de Pull Request in je werk mergen met zoiets als dit: $ curl https://github.com/tonychacon/fade/pull/1.patch | git am Samenwerken op basis van de Pull Request Zoals behandeld in De GitHub flow , kan je nu een conversatie voeren met de persoon die de Pull Request geopend heeft. Je kunt commentaar geven op specifieke regels code, commentaar geven op hele commits of commentaar geven over de gehele Pull Request; waarbij de GitHub smaak van Markdown overal gebruikt kan worden. Elke keer als iemand anders commentaar geeft op de Pull Request blijf je e-mails ontvangen zodat je weet dat er activiteit plaatsvindt. Alle betrokkenen krijgen een link naar de Pull Request waar de activiteit op plaatsvindt en je kunt ook direct de mail beantwoorden (reply-to) om commentaar op de Pull Request conversatie te geven. Figuur 116. Antwoorden op de emails worden in de converstatie betrokken. Zodra de code goed genoeg is en je het wilt mergen, kan je ofwel de code lokaal pullen en mergen met de git pull <url> <branch> syntax die we eerder gezien hebben, of door de fork als remote toe te voegen en dan te fetchen en mergen. Als de merge triviaal is, kan je ook gewoon de “Merge” knop op de GitHub site klikken. Dit zal een “non-fast-forward” merge uitvoeren, waardoor een merge commit wordt gemaakt zelfs als er een fast-forward merge mogelijk was. Dit houdt in dat hoe dan ook, elke keer als je de merge knop klikt een merge commit gemaakt wordt. Zoals je kunt zien in Merge knop en instructies hoe je een Pull Request handmatig merged. , geeft GitHub je al deze informatie als je de hint link klikt. Figuur 117. Merge knop en instructies hoe je een Pull Request handmatig merged. Als je besluit dat je het niet wilt mergen, kan je ook gewoon het Pull Request sluiten en de persoon die het geopend heeft krijgt een berichtje. Pull Request Refs (Pull Request referenties) Als je te maken hebt met veel Pull Requests en niet elke keer een aantal remotes wilt toevoegen of eenmalige pulls wilt doen, is er een aardige truuk die GitHub je toestaat te doen. Dit is een beetje een truuk voor gevorderden en we zullen de details beter behandelen in De Refspec , maar het kan behoorlijk handig zijn. GitHub presenteert de Pull Request branches voor een repository als een soort van pseudo-branches op de server. Standaard zal je ze niet krijgen als je kloont, maar ze zijn er wel op een verborgen manier en je kunt ze op een redelijk eenvoudige wijze benaderen. Om dit te demonstreren, zullen we een low-level (laag niveau) commando (waar vaak aan wordt gerefereerd als een “sanitaire voorziening” (plumbing) commando, waar we meer over zullen lezen in Binnenwerk en koetswerk (plumbing and porcelain) ) genaamd ls-remote gebruiken. Dit commando wordt over het algemeen niet gebruikt in de dagelijks Git operaties maar het is handig om te laten zien welke referenties er zich op de server bevinden. Als we dit commando gebruiken met de “blink” repository die we eerder zagen, krijgen we een lijst van alle branches, tags en andere refernties in de repository. $ git ls-remote https://github.com/schacon/blink 10d539600d86723087810ec636870a504f4fee4d HEAD 10d539600d86723087810ec636870a504f4fee4d refs/heads/master 6a83107c62950be9453aac297bb0193fd743cd6e refs/pull/1/head afe83c2d1a70674c9505cc1d8b7d380d5e076ed3 refs/pull/1/merge 3c8d735ee16296c242be7a9742ebfbc2665adec1 refs/pull/2/head 15c9f4f80973a2758462ab2066b6ad9fe8dcf03d refs/pull/2/merge a5a7751a33b7e86c5e9bb07b26001bb17d775d1a refs/pull/4/head 31a45fc257e8433c8d8804e3e848cf61c9d3166c refs/pull/4/merge Natuurlijk, als je in je repository bent en je gebruikt git ls-remote origin of welke remote je ook wilt controleren, zal het je iets vergelijkbaars laten zien. Als de repository op GitHub is en er zijn Pull Requests die open staan, zal je deze referenties krijgen die een prefix refs/pull/ hebben. Dit zijn eigenlijk gewoon branches, maar omdat deze niet onder refs/heads/ vermeld staan krijg je ze normaalgesproken niet als je van de server kloont of fetcht — het proces van fetchen negeert ze gewoonlijk. Er zijn twee referenties per Pull Request - degene die op /head eindigt wijst naar precies dezelfde commit als de laatste commit in de Pull Request branch. Dus als iemand in onze repository een Pull Request opent en zijn branch heeft de naam bug-fix` en deze wijst naar commit a5a775 , dan zal in onze repository geen branch bug-fix aanwezig zijn (omdat deze in hun fork zit), maar we hebben wel pull/<pr#>/head die wijst naar a5a775 . Dit betekent dat we relatief eenvoudig in één keer elke Pull Request branch kunnen pullen zonder daarvoor een heleboel remotes hoeven toe te voegen. Nu kan je iets doen wat lijkt op het direct fetchen van de referentie. $ git fetch origin refs/pull/958/head From https://github.com/libgit2/libgit2 * branch refs/pull/958/head -> FETCH_HEAD Dit zegt tegen Git, “Maak contact met de origin remote, en download de ref genaamd refs/pull/958/head .” Git gehoorzaamt trouw, en zal alles downloaden wat je nodig hebt om die ref samen te stellen, en zal een verwijzing naar de commit die je wilt onder .git/FETCH_HEAD zetten. Je kunt daarna een git merge FETCH_HEAD doen in een branch waar je dit in wilt testen, maar dat merge commit bericht zal er wat vreemd uitzien. Daarnaast, als je veel pull requests zal reviewen, wordt dit erg vervelend. Er is ook een manier om alle pull requests te fetchen en ze up-to-date te houden elke keer als je contact maakt met de remote. Open .git/config in je favoriete editor, en ga op zoek naar de origin remote. Dat zou er zo ongeveer uit moeten zien: [remote "origin"] url = https://github.com/libgit2/libgit2 fetch = +refs/heads/*:refs/remotes/origin/* De regel die begint met fetch = is een “refspec.” Het is een manier om namen op de remote te mappen met namen in je lokale .git directory. Deze specifieke zegt tegen Git: "de spullen op de remote die onder refs/heads staan moeten in mijn lokale repository onder refs/remotes/origin worden geplaatst." Je kan deze sectie wijzigen door een andere refspec toe te voegen: [remote "origin"] url = https://github.com/libgit2/libgit2.git fetch = +refs/heads/*:refs/remotes/origin/* fetch = +refs/pull/*/head:refs/remotes/origin/pr/* Die laatste regel vertelt Git, “Alle refs die er als refs/pull/123/head uitzien moeten lokaal worden opgeslagen als refs/remotes/origin/pr/123 .” Als je dit bestand nu opslaat en een git fetch uitvoert: $ git fetch # … * [new ref] refs/pull/1/head -> origin/pr/1 * [new ref] refs/pull/2/head -> origin/pr/2 * [new ref] refs/pull/4/head -> origin/pr/4 # … Nu worden alle remote pull request lokaal vertegenwoordigt met refs die zich vrijwel hetzelfde gedragen als tracking branches: ze zijn alleen-lezen, en ze worden geüpdatet als je een fetch doet. Dat maakt het enorm makkelijk om de code van een pull request lokaal uit te checken: $ git checkout pr/2 Checking out files: 100% (3769/3769), done. Branch pr/2 set up to track remote branch pr/2 from origin. Switched to a new branch 'pr/2' De oplettende lezers tussen jullie zullen de head aan het eind van het remote deel van de refspec hebben opgemerkt. Er is ook een refs/pull/#/merge ref aan de GitHub kant, welke de commit vertegenwoordigt die het resultaat zou zijn als je op de “merge” knop op de site zou klikken. Dit stelt je in staat de merge te testen zelfs voordat je de knop klikt. Pull Requests op Pull Requests Je kunt niet alleen Pull Requests openen die de main of master -branch betreffen, je kunt zelfs een Pull Request openen die betrekking heeft op elke willekeurige branch in het netwerk. Je kunt zelfs een Pull Request openen op een andere Pull Request. Als je een Pull Request ziet die de juiste kant op gaat en je hebt een idee voor een verandering die daarvan afhankelijk is of je weet niet zeker of het een goed idee is, of je hebt domweg geen push-toegang op de doelbranch, kan je een Pull Request direct op de Pull Request openen. Als je een Pull Request opent, is er een invoerveld bovenaan op de pagina die aangeeft naar welke branch je wilt laten pullen en welke je het van wilt laten pullen. Als je de “Edit” knop rechts van dat invoerveld klikt kan je niet alleen de branches wijzigen, maar ook welke fork. Figuur 118. Handmatig de Pull Request doelfork en -branch wijzigen. Hier kan je redelijk eenvoudig aangeven om jouw nieuwe branch in een andere Pull Request te mergen of in een andere fork van het project. Vermeldingen en meldingen GitHub heeft ook een redelijk aardig ingebouwde meldingen systeem die handig kan worden als je vragen hebt of terugkoppeling wilt van specifiek individuen of teams. In elk commentaar kan je een @ teken typen en een automatische aanvulling met de namen en gebruikersnamen van mensen die medewerkers of bijdragers in het project zal beginnen. Figuur 119. Begin @ te typen om iemand te vermelden. Je kunt ook een gebruiker vermelden die niet in die dropdown staat, maar vaak kan de automatische aanvuller het sneller maken. Als je eenmaal een commentaar hebt gepost met een gebruikersvermelding, zal die gebruiker een berichtje krijgen. Dat houdt in dat dit een erg effectieve manier is om iemand in een discussie te betrekken in plaats van ze actief hierop te laten controleren. Het gebeurt heel vaak op GitHub dat mensen via Pull Requests anderen in hun team of bedrijf erbij betrekken om een Issue of Pull Request te laten reviewen. Als iemand vermeld wordt in een Pull Request of Issue, worden ze erop “geabonneerd” en blijven meldingen krijgen voor elke activiteit die erop plaatsvindt. Je wordt ook geabonneerd op iets wat je geopend hebt, als je een repository volgt of als je ergens commentaar op geeft. Als je niet langer deze meldingen wilt krijgen, is er een “Unsubscribe” (Afmeld) knop op de pagina die je kunt klikken om de meldingen te stoppen. Figuur 120. Afmelden van een Issue of Pull Request. De meldingen pagina Als we hier over “meldingen” spreken in de context van GitHub, bedoelen we de specifieke manier waarop GitHub probeert om met je in contact te blijven als er gebeurtenissen zijn en er zijn een aantal manieren waarop je dit kunt configureren. Als je naar de “Notification center” tab van de instellingen pagina gaat, kan je een aantal opties zien die je hebt. Figuur 121. Notification center opties. De twee keuzes zijn om de berichten via “Email” en via “Web” te ontvangen en je kunt een van twee, geen of beide selecteren als je actief deelneemt aan zaken en voor activiteiten op repositories die je volgt. Web Meldingen Web meldingen bestaan alleen binnen GitHub en je kunt ze alleen op GitHub controleren. Als je deze optie geselecteerd hebt in je voorkeuren en een bericht wordt voor je gemaakt, zie je een kleine blauwe stip boven je meldingen ikoon boven in je scherm zoals getoond in Notification center. . Figuur 122. Notification center. Als je hierop klikt, zal je een lijst met alle items zien waar je voor wordt bericht, gegroepeerd per project. Je kunt de meldingen van een specifiek project filtreren door op de naam te klikken in de kolom links. Je kunt ook de meldingen bevestigen door het vink-ikoon te klikken naast elke melding, of alle meldingen in een project bevestigen door de vink boven in de groep te klikken. Er is ook een demp (mute) knop naast elke vinkteken die je kunt klikken om geen enkel bericht meer voor dat bericht te ontvangen. Al deze instrumenten zijn erg handig voor het afhandelen van grote aantallen meldingen. Veel gevorderde GitHub gebruikers zullen eenvoudigweg alle email berichten uitschakelen en al hun meldingen via dit scherm afhandelen. E-mail meldingen E-mail meldingen zijn de andere manier waarop je berichten kunt afhandelen via GitHub. Als je deze ingeschakeld hebt, krijg je per melding een e-mail. We hebben hiervan voorbeelden gezien in Email berichten en Email bericht van een nieuwe Request. . De e-mails zullen ook juist geketend (threaded) worden, wat handig is als je een zgn. threading e-mail client gebruikt. Er zit ook een behoorlijke hoeveelheid metadata in de headers van de e-mails die GitHub je stuurt, deze kunnen heel handig zijn voor het inrichten van zelfgemaakte filters en regels. Bijvoorbeeld, als we een kijkje nemen naar de daadwerkelijke e-mail headers die aan Tony zijn gestuurd in de e-mail in Email bericht van een nieuwe Request. , zullen we hetvolgende zien in de gestuurde gegevens: To: tonychacon/fade <fade@noreply.github.com> Message-ID: <tonychacon/fade/pull/1@github.com> Subject: [fade] Wait longer to see the dimming effect better (#1) X-GitHub-Recipient: tonychacon List-ID: tonychacon/fade <fade.tonychacon.github.com> List-Archive: https://github.com/tonychacon/fade List-Post: <mailto:reply+i-4XXX@reply.github.com> List-Unsubscribe: <mailto:unsub+i-XXX@reply.github.com>,... X-GitHub-Recipient-Address: tchacon@example.com Hier zijn een aantal interessante dingen. Als je e-mails wilt vlaggen of routeren naar dit specifieke project of zelfs Pull Request, geeft de informatie in Message-ID je alle gegevens in <user>/<project>/<type>/<id> formaat. Als dit bijvoorbeeld een issue zou zijn zou het <type> veld “issues” zijn niet “pull”. De List-Post en List-Unsubscribe velden houden in dat als je een mail client hebt die deze velden begrijpt, je eenvoudig een bericht kan sturen naar de thread of ervan kunt “Unsubscriben”. Dat zou effectief hetzelfde zijn als de “mute” knop klikken op de web versie van de melding of “Unsubscribe” op het Issue of Pull Request pagina zelf. Het is ook de moeite om te vermelden dat als je zowel e-mail als web meldingen aan hebt staan en je de e-mail versie van de melding gelezen hebt, de webversie ook als gelezen wordt gemarkeerd als je het laden van plaatjes toestaat in je mail client. Speciale Bestanden Er zijn een aantal speciale bestanden die GitHub zal opmerken als ze in je repository staan. README Het eerste is de README file, die kan ongeveer elk formaat hebben die GitHub herkent als vrije tekst. Bijvoorbeeld het zou README , README.md , README.asciidoc , etc. kunnen zijn. Als GitHub een README bestand in je broncode vindt, zal het dit op de ingangspagina van het project tonen. Veel teams gebruiken dit bestand om alle relevante project informatie te bevatten ter informatie voor iemand die nieuw is binnen de repository of het project. Meestal bevat het zaken als: Waartoe dient het project Hoe deze te configureren en te installeren Een voorbeeld hoe het te gebruiken of het aan te lopen te krijgen De licentie waaronder het project wordt aangeboden Hoe er aan bij te dragen Omdat GitHub dit bestand toont, kan je plaatjes of links er in opnemen voor het verhogen van het begrip. CONTRIBUTING Het andere speciale bestand dat GitHub herkent is het CONTRIBUTING bestand. Als je een bestand genaamd CONTRIBUTING met een willekeurige extentie, zal GitHub Een Pull Request openen als er een CONTRIBUTING file bestaan. tonen als iemand een Pull Request gaat openen. Figuur 123. Een Pull Request openen als er een CONTRIBUTING file bestaan. Het idee hierachter is dat je specifieke zaken kunt aangeven die je wel of niet wilt in een Pull Request die aan je project wordt gestuurd. Op deze manier zouden mensen de richtlijnen misschien echt lezen voordat ze een Pull Request openen. Project Beheer Over het algemeen zijn er niet veel beheersmatige zaken die je kunt doen met een enkel project, maar er zijn een aantal dingen die van belang kunnen zijn. De standaard branch wijzigen Als je een branch anders dan “master” gebruikt als je standaard branch waarvan je wilt dat mensen er Pull Requests op openen of die ze standaard zien, kan je dat in de instellingen pagina van je repository wijzigen onder de “Options” tab. Figuur 124. De default branch van een project wijzigen. Eenvoudigweg de standaard branch in de dropdown wijzigen en dat wordt vanaf dat moment de standaard-branch voor alle belangrijke handelingen, inclusief welke branch er standaard uitgechecked wordt als iemand de repository kloont. Een project overdragen Als je een project wilt overdragen aan een andere gebruiker of organisatie in GitHub, is er een “Transfer ownership” optie onderaan dezelfde “Options” tab van de instellingen pagina van je repository die je dat kan laten doen. Figuur 125. Draag een project over aan een andere GitHub gebruiker of organisatie. Dit is handig als je een project verlaat en iemand anders wilt het overnemen, of je project wordt groter en je wilt het in een organisatie onderbrengen. Niet alleen verplaatst dit de repostory met al zijn volgers en sterren naar een andere plaats, het richt ook een redirect (doorverwijzing) van jouw URL naar de nieuwe plaats. Het zal ook de clones en fetches van Git doorverwijzen, niet alleen de web-aanvragen. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/tr/v2/Bir-Sunucuda-Git-Kurma-GitLab
Git - GitLab About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlangıç 1.1 Sürüm Denetimi 1.2 Git’in Kısa Tarihçesi 1.3 Git Nedir? 1.4 Komut Satırı 1.5 Git’i Yüklemek 1.6 Git’i İlk Defa Kurmak 1.7 Yardım Almak 1.8 Özet 2. Git Temelleri 2.1 Bir Git Reposu Oluşturma/Kopyalama 2.2 Değişikliklerin Repoya Kaydedilmesi 2.3 Katkı Geçmişini Görüntüleme 2.4 Değişiklikleri Geri Alma 2.5 Uzak Repo ile Çalışmak 2.6 Etiketleme 2.7 Komut Kısayolu (Alias) Ayarlama 2.8 Özet 3. Git Dalları 3.1 Dallar 3.2 Kısaca Dallandırma ve Birleştirme Temelleri 3.3 Dal Yönetimi 3.4 İş Akışı Dallandırması 3.5 Uzak Dallar 3.6 Yeniden Temelleme (rebase) 3.7 Özet 4. Bir Sunucuda Git Kurma 4.1 İletişim Kuralları (Protocols) 4.2 Bir Sunucuda Git Kurma 4.3 SSH Ortak Anahtarınızı Oluşturma 4.4 Sunucu Kurma 4.5 Git Cini (Daemon) 4.6 Akıllı HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Taraf Barındırma (Hosting) Seçenekleri 4.10 Özet 5. Dağıtık Git 5.1 Dağıtık İş Akışları 5.2 Projenin Gelişiminde Rol Almak 5.3 Bir Projeyi Yürütme 5.4 Özet 6. GitHub 6.1 Bir Projeye Katkıda Bulunmak 6.2 Proje Bakımı 6.3 Kurumsal Yönetim 6.4 GitHub’ı otomatikleştirme 6.5 Özet 7. Git Araçları 7.1 Düzeltme Seçimi 7.2 Etkileşimli İzlemleme (Staging) 7.3 Saklama ve Silme 7.4 Çalışmanızı İmzalama 7.5 Arama 7.6 Geçmişi Yeniden Yazma 7.7 Reset Komutunun Gizemleri 7.8 İleri Seviye Birleştirme 7.9 Rerere 7.10 Git’le Hata Ayıklama 7.11 Alt Modüller 7.12 Demetleme (Bundling) 7.13 Git Nesnesini Değiştirme 7.14 Kimlik Bilgisi Depolama 7.15 Özet 8. Git’i Özelleştirmek 8.1 Git Yapılandırması 8.2 Git Nitelikleri 8.3 Git Kancaları (Hooks) 8.4 Bir Örnek: Mecburi Git Politikası 8.5 Özet 9. Git ve Diğer Sistemler 9.1 İstemci Olarak Git 9.2 Git’e Geçiş 9.3 Özet 10. Dahili Git Ögeleri 10.1 Tesisat ve Döşeme (Plumbing ve Porcelain) 10.2 Git Nesneleri 10.3 Git Referansları 10.4 Packfiles 10.5 Refspec 10.6 Transfer Protokolleri 10.7 Bakım ve Veri Kurtarma 10.8 Ortam Değişkenleri 10.9 Özet A1. Ek bölüm A: Diğer Ortamlarda Git A1.1 Görsel Arayüzler A1.2 Visual Studio ile Git A1.3 Visual Studio Code ile Git A1.4 Eclipse ile Git A1.5 Sublime Text ile Git A1.6 Bash ile Git A1.7 Zsh ile Git A1.8 PowerShell ile Git A1.9 Özet A2. Ek bölüm B: Git’i Uygulamalarınıza Gömmek A2.1 Git Komut Satırı A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Ek bölüm C: Git Komutları A3.1 Kurulum ve Yapılandırma Komutları A3.2 Proje Oluşturma Komutları A3.3 Kısaca Poz (Snapshot) Alma A3.4 Dallandırma ve Birleştirme Komutları A3.5 Projeleri Paylaşma ve Güncelleme Komutları A3.6 İnceleme ve Karşılaştırma Komutları A3.7 Hata Ayıklama (Debugging) Komutları A3.8 Yamalama (Patching) A3.9 E-Posta Komutları A3.10 Harici Sistemler A3.11 Yönetim A3.12 Tesisat (Plumbing) Komutları 2nd Edition 4.8 Bir Sunucuda Git Kurma - GitLab GitLab GitWeb oldukça basit olmasına rağmen, daha modern ve tam özellikli bir Git sunucusu arıyorsanız, onun yerine kurabileceğiniz birkaç açık kaynak çözümü bulunmaktadır. GitLab popüler olanlardan biri olduğu için, onu örnek olarak kurmayı ve kullanmayı ele alacağız. GitWeb seçeneğinden biraz daha karmaşık olmasına ve muhtemelen daha fazla bakım gerektirmesine karşın, çok daha kapsamlı bir seçenektir. Kurulum GitLab, veritabanı destekli bir ağ uygulaması olduğundan, kurulumu diğer bazı Git sunucularından biraz daha karmaşıktır. Neyse ki, bu süreç çok iyi belgelendirilmiştir ve desteklenmektedir. GitLab’ı kurmak için takip edebileceğiniz birkaç yöntem bulunmaktadır. Çalışan bir şeyi hızlı bir şekilde kurmak için, bir sanal makine görüntüsü indirebilir veya https://bitnami.com/stack/gitlab adresinden tek tıkla kurabilir ve özel ortamınıza uygun şekilde yapılandırabilirsiniz. Bitnami’nin eklediği hoş bir özellik, giriş ekranıdır (alt+→ tuşlarını yazarak erişilir). Bu, yüklenmiş GitLab’ın IP adresini ve varsayılan kullanıcı ad ve şifresini size bildirir. Görsel 50. Bitnami GitLab sanal makine giriş ekranı. Başka bir şey için, GitLab Community Edition readme (beni-oku) dosyasındaki yönergeleri takip edin (bu dosyayı https://gitlab.com/gitlab-org/gitlab-ce/tree/master adresinde bulabilirsiniz). Buradaki şef tarifleri, Digital Ocean üzerinde bir sanal makine ile RPM ve DEB paketleri kullanarak (yazıldığı tarih itibariyle beta sürümünde olanlar) size GitLab’ı nasıl kuracağınıza dair yardım sulabilir. Ayrıca, standart olmayan işletim sistemleri ve veritabanlarıyla GitLab’ı çalıştırma, tamamen manuel bir kurulum betiği ve birçok başka konuyla ilgili "resmi olmayan" yönergeler bulabilirsiniz. Yönetim GitLab’ın yönetim arayüzüne ağ üzerinden erişilir. Sadece tarayıcınızı GitLab’ın kurulu olduğu ana bilgisayar adına veya IP adresine yönlendirin ve bir yönetici kullanıcısı olarak oturum açın. Varsayılan kullanıcı adı admin@local.host ve varsayılan şifre 5iveL!fe şeklindedir (şifreyi girdikten hemen sonra değiştirmeniz istenecektir). Oturum açtıktan sonra, sağ üst köşede bulunan menüdeki "Yönetici Alanı" simgesine tıklayın. Görsel 51. GitLab menüsündeki "Yönetici Alanı" ögesi. Kullanıcılar GitLab’deki kullanıcılar, insanlara karşılık gelen hesaplardır. Kullanıcı hesapları çok karmaşık değildir: genellikle, giriş verilerine eklenmiş kişisel bilgilerin bir koleksiyonudur. Her kullanıcı hesabı, o kullanıcıya ait projelerin mantıksal bir gruplaması olan bir adalanı (namespace) ile gelir. Eğer mehmet kullanıcısının projem adında bir projesi olsaydı, o proje’nin URL’si http://sunucu/mehmet/projem olurdu. Görsel 52. GitLab kullanıcı yönetim ekranı. Bir kullanıcıyı kaldırmak iki şekilde yapılabilir: Bir kullanıcıyı "engellemek", onların GitLab örneğine giriş yapmasını engeller, ancak o kullanıcının adalanı altındaki tüm veriler korunur ve o kullanıcının e-posta adresiyle imzalanan katkılar hala profiliyle bağlantılıdır. Öte yandan, bir kullanıcıyı "silmek", onu tamamen veritabanından ve dosya sisteminden kaldırır. Adalanı altındaki tüm projeler ve veriler, sahip oldukları tüm gruplarla birlikte kaldırılır. Bu açıkça çok daha kalıcı ve yıkıcı bir işlem olduğu için kullanımı nadirdir. Gruplar Bir GitLab grubu bir projeler topluluğudur ve kullanıcıların bu projelere nasıl erişebileceği hakkında veriler içerir. Her grup (kullanıcılar gibi) bir proje adalanına sahiptir, bu nedenle training adında bir grup materials adında bir projeye sahipse, URL’si http://sunucu/training/materials olur. Görsel 53. GitLab grup yönetim ekranı. Her bir grup, grubun kendisine ve projelerine, belli bir düzeyde erişim iznine sahip, bir dizi kullanıcı ile ilişkilidir. Bu izinler, "Misafir" (yalnızca konu ve sohbet) ile "Sahip" (grubun, üyelerin ve projelerin tam denetimi) arasında değişmektedir. İzin türleri burada listelenmek için fazla sayıda, ancak GitLab’ın yönetim ekranında yardımcı bir bağlantı bulunmaktadır. Projeler Bir GitLab projesi yaklaşık olarak tek bir Git reposuna karşılık gelir. Her proje, bir kullanıcı veya bir grup olmak üzere tek bir adalanına aittir. Proje bir kullanıcıya aitse, projenin sahibi projeye kimin erişebileceğini doğrudan kontrol eder. Eğer proje bir gruba aitse, grup düzeyindeki kullanıcı izinleri de etkili olacaktır. Her proje ayrıca bir görünürlük seviyesine sahiptir, bu da projenin sayfalarına ve reposuna kimlerin okuma erişimine sahip olduğunu denetler. Bir proje Private (özel) ise, proje sahibi belirli kullanıcılara erişimi açıkça vermelidir. Internal (iç) bir proje sadece oturum açmış her kullanıcı tarafından görülebilir, Public (açık) bir proje ise herkese görünür. Unutmayın ki, bu hem git fetch erişimini hem de o projenin web arayüzüne erişimi kontrol eder. Kancalar (Hooks) GitLab, hem proje ve hem de sistem düzeyinde kancaları destekler. Her ikisi için de, ilgili olaylar meydana geldiğinde GitLab sunucusu açıklamacı JSON ile bir HTTP POST gerçekleştirir. Bu, Git repolarınızı ve GitLab örneğinizi, CI sunucuları, sohbet odaları veya dağıtım araçları gibi geliştirme otomasyonunuzun geri kalanına bağlamanın harika bir yoludur. Basit Kullanım GitLab’da yapmak isteyeceğiniz ilk şey yeni bir proje oluşturmaktır. Bunu, araç çubuğundaki ``+`` simgesine tıklayarak yapabilirsiniz. Projenin adı, hangi adalanına ait olması gerektiği ve görünürlük seviyesinin ne olması gerektiği sorulacaktır. Burada belirttiğiniz çoğu şey kalıcı değildir ve daha sonra ayarlar arayüzü üzerinden yeniden ayarlanabilir. ``Create Project``e (proje oluştur) tıklayın ve işlem tamamdır. Proje oluştuktan sonra, muhtemelen yerel bir Git reposuyla bağlantı kurmak isteyeceksiniz. Her proje HTTPS veya SSH üzerinden erişilebilirdir ve her ikisi de bir Git uzak sunucusu yapılandırmak için kullanılabilir. URL’ler, proje ana sayfasının üstünde görünür. Bu komut mevcut bir yerel repo için barındırılan konuma gitlab adında bir uzak sunucu oluşturacaktır: $ git remote add gitlab https://server/namespace/project.git Eğer reponuzun yerel bir kopyası yoksa, bunu sadece şu şekilde yapabilirsiniz: $ git clone https://server/namespace/project.git Ağ arayüzü reponun kendisine ait bazı kullanışlı görüntülere erişim sağlar. Her proje ana sayfası, son etkinliği gösterir ve üst kısımdaki bağlantılar, proje dosyalarının ve işlem günlüğün görüntülerine yönlendirir. Birlikte Çalışma Bir GitLab projesinde birlikte çalışmanın en basit yolu, başka bir kullanıcıya Git reposuna doğrudan itme erişimi vermektedir. Bunu yapmak için, o projenin ayarlarında "Üyeler" bölümüne giderek yeni kullanıcıyı bir erişim seviyesi ile ilişkilendirebilirsiniz (farklı erişim seviyeleri konusuna Gruplar bölümünde biraz girilmiştir). "Geliştirici" veya daha üst bir erişim seviyesi verilen bir kullanıcı o repoya gelişigüzel olarak doğrudan işlemler ve dallar ekleyebilir. Daha bağımsız bir işbirliği yöntemi de "birleştirme isteklerini" kullanmaktır. Bu özellik, bir projeyi görebilen herhangi bir kullanıcının, denetimli bir şekilde katkıda bulunmasını sağlar. Doğrudan erişimi olan kullanıcılar basitçe bir dal oluşturabilir, işlemleri ona iter ve dallarından master veya başka bir dala birleştirme isteği açabilir. Depoya itme izinleri olmayan kullanıcılar onu "çatallayabilir" (kendi kopyalarını oluşturabilir), işlemleri bu kopyaya iter ve kendi çatalından ana projeye birleştirme isteği açabilir. Bu model "güvenilmeyen kullanıcılardan" katkı alınmasına olanak tanırken, sahibin (owner) de repoya neyin, ne zaman gireceği konusunda tam bir kontrol sahibi olmasını sağlar. Birleştirme istekleri ve konular, GitLab’da uzun süreli tartışmanın ana birimleridir. Her birleştirme isteği, önerilen değişikliğin satır bazında tartışılmasına (hafif bir kod incelemesi) ve genel bir genel tartışma başlığına izin verir. Her ikisi de kullanıcılara atanabilir veya kilometre taşlarına düzenlenebilir. Bu bölüm, öncelikle GitLab’ın Git ile ilgili özelliklerine odaklanmış olsa da, olgun bir proje olarak, ekibinizin birlikte çalışmasına yardımcı olacak birçok başka özellik de sunar (örneğin proje vikileri ve sistem bakım araçları). GitLab’ın bir avantajı da sunucu kurulup ve çalıştırıldığı zaman, nadiren bir yapılandırma dosyasını ayarlamak veya sunucuya SSH üzerinden erişmek gibi bir ihtiyacınızın olmamasıdır. Çoğu yönetim ve genel kullanım, tarayıcı tabanlı arayüz üzerinden gerçekleştirilebilir. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/bg/v2/%d0%9d%d0%b0%d1%87%d0%b0%d0%bb%d0%be-%d0%98%d0%bd%d1%81%d1%82%d0%b0%d0%bb%d0%b8%d1%80%d0%b0%d0%bd%d0%b5-%d0%bd%d0%b0-Git
Git - Инсталиране на Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Начало 1.1 За Version Control системите 1.2 Кратка история на Git 1.3 Какво е Git 1.4 Конзолата на Git 1.5 Инсталиране на Git 1.6 Първоначална настройка на Git 1.7 Помощна информация в Git 1.8 Обобщение 2. Основи на Git 2.1 Създаване на Git хранилище 2.2 Запис на промени в хранилището 2.3 Преглед на историята на действията 2.4 Възстановяване на направени действия 2.5 Работа с отдалечени хранилища 2.6 Тагове в Git 2.7 Псевдоними в Git 2.8 Обобщение 3. Клонове в Git 3.1 Накратко за разклоненията 3.2 Основи на клоновете код и сливането 3.3 Управление на клонове 3.4 Стратегии за работа с клонове код 3.5 Отдалечени клонове 3.6 Управление на проект 3.7 Обобщение 4. GitHub 4.1 Създаване и настройка на акаунт 4.2 Как да сътрудничим в проект 4.3 Управление на проект 4.4 Управление на организация 4.5 Автоматизиране с GitHub 4.6 Обобщение 5. Git инструменти 5.1 Избор на къмити 5.2 Интерактивно индексиране 5.3 Stashing и Cleaning 5.4 Подписване на вашата работа 5.5 Търсене 5.6 Манипулация на историята 5.7 Мистерията на командата Reset 5.8 Сливане за напреднали 5.9 Rerere 5.10 Дебъгване с Git 5.11 Подмодули 5.12 Пакети в Git (Bundling) 5.13 Заместване 5.14 Credential Storage система 5.15 Обобщение 6. Настройване на Git 6.1 Git конфигурации 6.2 Git атрибути 6.3 Git Hooks 6.4 Примерна Git-Enforced политика 6.5 Обобщение 7. Git и други системи 7.1 Git като клиент 7.2 Миграция към Git 7.3 Обобщение 8. Git на ниско ниво 8.1 Plumbing и Porcelain команди 8.2 Git обекти 8.3 Git референции 8.4 Packfiles 8.5 Refspec спецификации 8.6 Транспортни протоколи 8.7 Поддръжка и възстановяване на данни 8.8 Environment променливи 8.9 Обобщение 9. Приложение A: Git в други среди 9.1 Графични интерфейси 9.2 Git във Visual Studio 9.3 Git във Visual Studio Code 9.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 9.5 Git в Sublime Text 9.6 Git в Bash 9.7 Git в Zsh 9.8 Git в PowerShell 9.9 Обобщение 10. Приложение B: Вграждане на Git в приложения 10.1 Git от команден ред 10.2 Libgit2 10.3 JGit 10.4 go-git 10.5 Dulwich A1. Приложение C: Git команди A1.1 Настройки и конфигурация A1.2 Издърпване и създаване на проекти A1.3 Snapshotting A1.4 Клонове и сливане A1.5 Споделяне и обновяване на проекти A1.6 Инспекция и сравнение A1.7 Дебъгване A1.8 Patching A1.9 Email команди A1.10 Външни системи A1.11 Административни команди A1.12 Plumbing команди 2nd Edition 1.5 Начало - Инсталиране на Git Инсталиране на Git Преди да започнете да ползвате Git, трябва да го инсталирате на компютъра си. Дори ако вече е инсталиран, добра идея е да обновите до последната версия. Инсталацията става като пакет, чрез друг инсталатор или чрез компилиране на изходния код. Забележка Тази книга е написана за Git версия 2. Понеже Git е достатъчно добър в поддържането на обратната съвместимост, всяка актуална версия би трябвало да работи добре. Въпреки че повечето команди трябва да работят дори в много стари версии на Git, някои от тях може да работят с леки разлики, ако вашата версия е по-стара. Инсталация в Linux Ако искате да инсталирате основните Git инструменти под Linux с binary инсталатор, в общия случай това е лесно с пакетните инструменти на вашата дистибуция. Например под Fedora (или всяка друга подобна, RPM-базирана дистрибуция като RHEL или CentOS), можете да ползвате dnf : $ sudo dnf install git-all Ако предпочитате Debian-базирана дистрибуция като Ubuntu, опитайте с apt : $ sudo apt install git-all За повече подробности и опции касаещи инсталацията в Linux, вижте сайта на Git: https://git-scm.com/download/linux . Инсталация в macOS Има няколко начина за инсталиране на Git в Mac. Може би най-лесният е да инсталирате Xcode Command Line Tools. Под Mavericks (10.9) и по-новите версии, можете да направите това просто като опитате да изпълните команда 'git' в терминала първия път. $ git --version Ако не сте го инсталирали вече, системата ще ви предложи да го направите. Ако желаете по-актуална версия, можете да я инсталирате и през binary инсталатор. OSX Git инсталатор за MacOS се поддържа и може да се изтегли от https://git-scm.com/download/mac . Фигура 7. Git macOS Installer Инсталация в Windows И тук има няколко опции да инсталирате Git. Официалната версия е налична за сваляне от сайта на Git. Отворете https://git-scm.com/download/win и изтеглянето ще започне автоматично. Имайте предвид, че това е проект наречен Git for Windows, който е отделен от самия Git, за повече информация за него, посетете https://gitforwindows.org . Ако искате автоматизирана инсталация, можете да използвате Git Chocolatey package . Chocolatey пакетът се поддържа от общност доброволци. Инсталация от сорс-код Някои хора предпочитат да инсталират Git от изходния код, защото по този начин получават възможно най-актуалната версия. Бинарните инсталатори обикновено са за една идея по-стари версии, макар че това не е толкова важно, защото Git е много съвместим. Ако искате да инсталирате Git от изходен код, ще се нуждатете от библиотеките autotools, curl, zlib, openssl, expat, и libiconv, понеже Git зависи от тях. За Fedora или Debian-базирана дистрибуция, изпълнете долните команди съответно, така че да се сдобиете с минималните изисквания за компилация и инсталиране на Git: $ sudo dnf install dh-autoreconf curl-devel expat-devel gettext-devel \ openssl-devel perl-devel zlib-devel $ sudo apt-get install dh-autoreconf libcurl4-gnutls-dev libexpat1-dev \ gettext libz-dev libssl-dev За да можете да добавите документацията в различни формати (doc, html, info), са необходими допълнителните зависимости отдолу: $ sudo dnf install asciidoc xmlto docbook2X $ sudo apt-get install asciidoc xmlto docbook2x Забележка Потребителите на RHEL и RHEL деривати като CentOS и Scientific Linux трябва да разрешат EPEL хранилището за да изтеглят пакета docbook2X . Ако използвате Debian-базирана дистрибуция (Debian/Ubuntu/Ubuntu-варианти), ще се нуждаете също и от пакета install-info : $ sudo apt-get install install-info Ако използвате RPM дистрибуция (Fedora/RHEL/RHEL-деривати), ще ви трябва пакета getopt (който е наличен по подразбиране в Debian-базираните дистрибуции): $ sudo dnf install getopt Освен това, ако ползвате Fedora/RHEL/RHEL-деривати, трябва да изпълните това: $ sudo ln -s /usr/bin/db2x_docbook2texi /usr/bin/docbook2x-texi поради различия в имената на двоичните файлове. След като се уверите, че имате инсталирани всички зависимости, продължавате напред и изтегляте най-новия архив с изходен код на Git. Това може да стане от няколко места - сайта Kernel.org на адрес https://www.kernel.org/pub/software/scm/git , или от хранилището в GitHub - https://github.com/git/git/releases . Обикновено страницата в GitHub би следвало да е по-актуална, но и Kernel.org също разполага с контролни сигнатури, ако желаете да проверите какво сте изтеглили. Следва компилация и инсталиране: $ tar -zxf git-2.8.0.tar.gz $ cd git-2.8.0 $ make configure $ ./configure --prefix=/usr $ make all doc info $ sudo make install install-doc install-html install-info След като направите това, можете да изтеглите Git от самия Git, за обновявания: $ git clone git://git.kernel.org/pub/scm/git/git.git prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-GitWeb
Git - GitWeb About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.7 Git op de server - GitWeb GitWeb Nu je gewone lees/schrijf en alleen-lezen toegang tot je project hebt, wil je misschien een eenvoudige web-gebaseerde visualisatie instellen. Git levert een CGI script genaamd GitWeb mee, dat soms hiervoor gebruikt wordt. Figuur 49. De GitWeb web-based gebruikers interface. Als je wilt zien hoe GitWeb eruit ziet voor jouw project, kan je een commando wat met Git geleverd wordt gebruiken om een tijdelijke instantie op te starten als je een lichtgewicht server op je systeem hebt als lighttpd of webrick . Op Linux machines is lighttpd vaak geïnstalleerd, dus je zou in staat moeten zijn om het te laten lopen door git instaweb in te typen in je project directory. Als je een Mac gebruikt, Leopard heeft Ruby voor-geïnstalleerd, zou webrick de beste gok kunnen zijn. Om instaweb te starten met een niet-lighttpd handler, kan je het aanroepen met de --httpd optie. $ git instaweb --httpd=webrick [2009-02-21 10:02:21] INFO WEBrick 1.3.1 [2009-02-21 10:02:21] INFO ruby 1.8.6 (2008-03-03) [universal-darwin9.0] Daarmee wordt een HTTPD server op poort 1234 opgestart en daarna een webbrowser die opent op die pagina. Voor jou stelt dit niet veel voor. Als je klaar bent en je de server weer wilt afsluiten, kan je hetzelfde commando met de --stop optie aanroepen: $ git instaweb --httpd=webrick --stop Als je de web interface permanent op een server wilt hebben draaien voor je team of voor een open source project die je host, moet je je reguliere web server inrichten om het CGI script te serveren. Sommige Linux distributies hebben een gitweb package dat je wellicht met apt of yum kunt installeren, wellicht kan je dat eerst proberen. We zullen spoedig het handmatig installeren van GitWeb bespreken. Eerst zal je de Git broncode, waar GitWeb mee geleverd wordt, moeten verkrijgen, en het volgende maatwerk CGI script genereren: $ git clone git://git.kernel.org/pub/scm/git/git.git $ cd git/ $ make GITWEB_PROJECTROOT="/srv/git" prefix=/usr gitweb SUBDIR gitweb SUBDIR ../ make[2]: `GIT-VERSION-FILE' is up to date. GEN gitweb.cgi GEN static/gitweb.js $ sudo cp -Rf gitweb /var/www/ Merk op dat je het commando moet vertellen waar je Git repositories gevonden kunnen worden met de GITWEB_PROJECTROOT variabele. Vervolgens moet je ervoor zorgen dat Apache CGI gebruikt voor dat script, daarvoor kan je een VirtualHost toevoegen: <VirtualHost *:80> ServerName gitserver DocumentRoot /var/www/gitweb <Directory /var/www/gitweb> Options +ExecCGI +FollowSymLinks +SymLinksIfOwnerMatch AllowOverride All order allow,deny Allow from all AddHandler cgi-script cgi DirectoryIndex gitweb.cgi </Directory> </VirtualHost> Nogmaals: GitWeb kan worden geserveerd met elke web server die CGI of Perl ondersteunt, als je toch iets anders wilt gebruiken zou het niet al te moeilijk moeten zijn dit in te richten. Nu zou je in staat moeten zijn om http://gitserver/ op te zoeken en je repositories online te zien. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/tl/v2/Mga-Pangunahing-Kaalaman-sa-Git-Paggawa-gamit-ang-mga-Remote
Git - Paggawa gamit ang mga Remote About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pagsisimula 1.1 Tungkol sa Bersyon Kontrol 1.2 Isang Maikling Kasaysayan ng Git 1.3 Pangunahing Kaalaman sa Git 1.4 Ang Command Line 1.5 Pag-install ng Git 1.6 Unang Beses na Pag-Setup ng Git 1.7 Pagkuha ng Tulong 1.8 Buod 2. Mga Pangunahing Kaalaman sa Git 2.1 Pagkuha ng Repositoryo ng Git 2.2 Pagtatala ng mga Pagbabago sa Repositoryo 2.3 Pagtitingin sa Kasaysayan ng Commit 2.4 Pag-Undo ng mga Bagay 2.5 Paggawa gamit ang mga Remote 2.6 Pag-tag 2.7 Mga Alyas sa Git 2.8 Buod 3. Pag-branch ng Git 3.1 Mga Branch sa Maikling Salita 3.2 Batayan ng Pag-branch at Pag-merge 3.3 Pamamahala ng Branch 3.4 Mga Daloy ng Trabaho sa Pag-branch 3.5 Remote na mga Branch 3.6 Pag-rebase 3.7 Buod 4. Git sa Server 4.1 Ang Mga Protokol 4.2 Pagkuha ng Git sa isang Server 4.3 Ang paglikha ng iyong Pampublikong Susi ng SSH 4.4 Pag-Setup ng Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Mga Opsyon ng Naka-host sa Third Party 4.10 Buod 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Pag-setup at pagsasaayos ng Account 6.2 Pag-aambag sa isang Proyekto 6.3 Pagpapanatili ng isang Proyekto 6.4 Pamamahala ng isang organisasyon 6.5 Pag-iiskrip sa GitHub 6.6 Buod 7. Mga Git na Kasangkapan 7.1 Pagpipili ng Rebisyon 7.2 Staging na Interactive 7.3 Pag-stash at Paglilinis 7.4 Pag-sign sa Iyong Trabaho 7.5 Paghahanap 7.6 Pagsulat muli ng Kasaysayan 7.7 Ang Reset Demystified 7.8 Advanced na Pag-merge 7.9 Ang Rerere 7.10 Pagdebug gamit ang Git 7.11 Mga Submodule 7.12 Pagbibigkis 7.13 Pagpapalit 7.14 Kredensyal na ImbakanCredential Storage 7.15 Buod 8. Pag-aangkop sa Sariling Pangangailagan ng Git 8.1 Kompigurasyon ng Git 8.2 Mga Katangian ng Git 8.3 Mga Hook ng Git 8.4 An Example Git-Enforced Policy 8.5 Buod 9. Ang Git at iba pang mga Sistema 9.1 Git bilang isang Kliyente 9.2 Paglilipat sa Git 9.3 Buod 10. Mga Panloob ng GIT 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 Ang Refspec 10.6 Transfer Protocols 10.7 Pagpapanatili At Pagbalik ng Datos 10.8 Mga Variable sa Kapaligiran 10.9 Buod A1. Appendix A: Git in Other Environments A1.1 Grapikal Interfaces A1.2 Git in Visual Studio A1.3 Git sa Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git sa Powershell A1.7 Summary A2. Appendix B: Pag-embed ng Git sa iyong Mga Aplikasyon A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Mga Kautusan ng Git A3.1 Setup at Config A3.2 Pagkuha at Paglikha ng Mga Proyekto A3.3 Pangunahing Snapshotting A3.4 Branching at Merging A3.5 Pagbabahagi at Pagbabago ng mga Proyekto A3.6 Pagsisiyasat at Paghahambing A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Pagtutuberong mga Utos 2nd Edition 2.5 Mga Pangunahing Kaalaman sa Git - Paggawa gamit ang mga Remote Paggawa gamit ang mga Remote Upang magawa ang pakikipagtutulungan sa anumang proyekto sa Git, kailangan mong malaman kung paano papangasiwaan ang iyong mga remote na repositoryo. Ang mga naka-remote na repositoryo ay mga bersyon ng iyong proyekto na naka-host sa Internet o sa saan mang lugar sa network. Maaari kang magkaroon iilan sa kanila, sa pangkalahatan bawat isa ay maaaring read-only o read/write para sayo. Ang pakikipagtulungan sa iba ay naglalakip sa pamamahala sa naka-remote na mga repositoryo na ito at pag-push at pag-pull ng datos sa patutunguhan at sa pinanggalingan nila kapag kailangan mong magbahagi ng trabaho. Ang pagpapangasiwa ng naka-remote na mga repositoryo ay naglakip nang kaalaman kung paano magdagdag ng naka-remote na mga repositoryo, mag-alis sa mga naka-remote na hindi na magagamit, mangangasiwa ng iba’t ibang mga sanga at tukuyin ang mga ito kung sinusubaybayan o hindi, at marami pang iba. Sa seksyon na ito, tatalakayin natin ang iilan sa mga kakayahan sa pagpapangasiwa ng naka-remote na repositoryo. Example 2. Ang naka-remote na mga repositoryo ay maaaring nasa iyong lokal na makina. Ito ay posible na ikaw ay nagtatrabaho sa isang “naka-remote” na repositoryo na iyan, at sa katunayan, nasa parehong host kung saan nandoon ka. Ang salitang “remote” ay hindi kinakailangan nagpapahiwatig na ang repositoryo ay nasa ibang lugar sa network o Internet, tanging ito ay nasa ibang lugar lang. Ang pagtrabaho sa isang naka-remote na repositoryo gaya nito ay magkakaroon pa rin sa lahat ng karaniwang pag-push, pag-pull at pag-fetch ng mga operasyon tulad ng sa ibang anumang nakaremote. Pagpapakita ng iyong mga Remote Upang makita kung aling remote na mga server ang iyong na-configure, maaari mong patakbuhin ang git remote na utos. Inilista nito ang mga alyas sa bawat remote na handle na iyong itinukoy. Kung na-clone mo ang iyong repositoryo, dapat mong makita ang origin  — iyon ang naka-default na pangalan na ibinigay ng Git sa server na pinagmulan ng iyong naka-clone na repositoryo: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Maaari mo ring ilagay ang -v , na nagpapakita sa iyo sa mga URL na iniimbak ng Git na alyas para magagamit kapag nagbabasa at nagsusulat sa remote na iyon: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Kung mayroon kang higit pa sa isang remote, ang utos ay naglilista sa kanilang lahat. Bilang halimbawa, ang isang repositoryo na may maramihang mga remote para sa pakikipagtulungan sa mga iilang mga taga-ambag ay maaaring maging tulad nito. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Ang ibig sabihin nito ay maaari naming madaling i-pull ang mga kontribusyon mula sa sinumang mga gumagamit Maaari nating idagdag ang mga pahintulot na mag-push sa isa o higit pa sa mga ito, bagaman hindi natin ito masasabi dito. Kung mapansin mo ang mga remote na ito ay gumagamit ng napakaraming mga protokol; tatalakayin pa natin ito nang malaliman sa Pagkuha ng Git sa isang Server . Pagdagdag ng Remote na mga Repositoryo Nabanggit at ibinigay na namin ang ilang mga pagpapakita kung paano ang git clone na utos ay sadyang nagdagdag ng origin na remote para sa iyo. Narito ang paraan kung papaano hayagang magdagdag ng isang bagong remote. Para makadagdag ng bagong remote na repositoryo ng Git bilang isang alyas para madali mong magamit, patakbuhin ang git remote add <shortname> <url> : $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Ngayon maaari mo nang gamitin ang string pb sa command line sa halip ng buong URL. Halimbawa, kung gusto mong kunin ang lahat ng impormasyon na mayroon si Pablo ngunit wala pa sa iyong repositoryo, maaari mong patakbuhin ang git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Ang master branch ni Pablo ay magagamit na ngayon sa lokal bilang pb/master  — maaari mo itong i-merge sa isa sa iyong mga branch, o maaari mong suriin ang lokal na branch sa puntong iyon kung gusto mong siyasatin ito. (Ating tatalakayin kung ano ang mga branch at kung paano gamitin ang mga ito na mas detalyado sa Pag-branch ng Git .) Pag-fetch at Pag-Pull mula sa iyong mga Remote Tulad ng iyong nakikita, upang makakuha ng datos mula sa iyong remote na mga proyekto, maaari kang magtakbo: $ git fetch <remote> Ang utos ay napupunta sa remote na proyekto na iyon at nagkuha sa lahat ng datos mula sa remote na proyekto na iyon na wala ka pa. Pagkatapos mong gawin ito, mayroon ka ng mga reperensiya sa lahat ng mga branch mula sa remote na iyon, kung saan maaari mo itong i-merge o siyasatin sa anumang oras. Kung mag-clone ka ng isang repositoryo, ang utos ay awtomatikong nagdagdag sa remote na repositoryo na iyon sa ilalim ng pangalan “origin”. Kaya, ang git fetch origin ay kumukuha ng anumang bagong trabaho na nai-push sa server na iyon mula nang na-clone (o huling kang nag-fetch) nito. Mahalaga itong tandaan na ang git fetch na utos ay nagda-download lang ng mga datos sa iyong lokal na repositoryo — ito ay hindi awtomatikong nag-merge sa anumang ginawa o binago mo na kasalukuyan mong ginawa. Kailangan mong i-merge ito ng mano-mano sa iyong trabaho kapag ikaw ay handa na. Kung ang iyong kasalukuyang branch ay naka-set up upang subaybayan ang remote na branch (tingnan ang susunod na seksyon at Pag-branch ng Git para sa karagdagang impormasyon), maaari kang gumagamit ng git pull na utos upang awtomatikong mag-fetch at pagkatapos ay mag-merge sa remote na branch sa iyong kasalukuyang branch. Ito ay maaaring maging mas madali o maginhawang proseso para sa iyo; at bilang default, ang git clone na utos ay awtomakong nag-setup sa iyong lokal na branch upang subaybayan ang remote na master branch (o anumang default na branch na tinatawag) sa server na naka-clone ka. Ang pagpapatakbo ng git pull ay karaniwang kumukuha ng datos mula sa server na pinanggalingan ng iyong kopya ng repositoryo at awtomatikong nagsusubok na i-merge ito sa kasalukuyang tinatrabaho mo. Pagtulak sa iyong mga Remote Kapag mayroon kang proyekto na gusto mong ibahagi sa isang punto, kailangan mong i-push ito pataas. Ang utos para dito ay simple lang: git push <remote> <branch> . Kung gusto mo i-push ang iyong master branch sa iyong origin sa server (sa uulitin, ang pag-clone ay awtomatikong nag-setup mga pangalan na iyon para sa iyo), pagkatapos ay maaari mo nang patakbuhin ito para i-push ang anumang mga commit na nagawa mo papunta sa server: $ git push origin master Ang utos na ito ay gagana lang kung ikaw ay naka-clone mula sa isang server na kung saan meron kang pahintulot na magsulat at kung walang pang pansamantalang nag-push. Kung ikaw at sinumang nag-clone sa parehong oras at sila ay nag-push pataas at pagkatapos ikaw ay nag-push pataas, ang iyong pag-push ay hindi tatanggapin. Kailangan mo pang kunin ang kanilang trabaho at isama ito sa iyong nagawa bago ka pahintulotang mag-push. Tingnan ang Pag-branch ng Git para sa mga karagdagang detalye kung papaano mag-push sa remote na mga server. Pagsusuri sa Remote Kung gusto mong makakita pa ng maraming impormasyon tungkol sa partikular na remote, maaari mong gamitin ang git remote show <remote> na utos. Kung papatakbuhin mo ang utos na ito na may isang partikular na alyas, tulad ng origin , makakuha ka ng ganito: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Naglilista ito ng URL para sa remote na repositoryo pati na rin ang impormasyon sa pagsubabay ng branch. Nakakatulong ang utos na magsabi sa iyo na kung ikaw ay nasa master na branch at ikaw ay nagpapatakbo ng git pull , awtomatiko itong i-merge sa master na branch na nasa iyong remote pagkatapos nitong makuha ang lahat ng remote na mga reperensiya. Naglilista din ito sa lahat ng remote na mga reperensiya na nakuha nito pababa. Iyan ay isang simpleng halimbawa na malamang ay matatagpuan mo. Kapag mas madalas ang paggamit mo sa Git, gayunman, maaari kang makakita ng mas maraming impormasyon mula sa git remote show : $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Ang utos na ito ay nagpapakita kung aling branch ang awtomatikong na-push kapag pinatakbo mo ang git push habang nasa ibang branch ka. Ito rin ay nagpapakita sa iyo kung aling remote na mga branch ang nasa server na wala ka pa, kung aling mga remote na branch na meron ka pa na inalis mo na sa server, at maraming lokal na mga branch na awtomatikong mai-merge sa kanilang remote-tracking na branch kapag nagpatakbo ka ng git pull . Pagpapalit ng pangalan at Pagtatangal ng mga Remote Maaari kang magpatakbo ng git remote rename upang baguhin ang alyas ng isang remote. Halimbawa, kung ikaw ay gustong magbago ng pangalan ng pb sa paul , maaari mo itong gawin gamit ang git remote rename : $ git remote rename pb paul $ git remote origin paul Ito ay mahalagang banggitin na ito ay nagbabago rin sa lahat ng iyong remote-tracking na mga pangalan ng branch. Ang dating ginamit na pb/master ay ngayon ay nasa paul/master na. Kung gusto mong tanggalin ang remote para sa kahit anong rason — inilipat mo ang server o hindi na ginagamit ang isang partikular na mirror, o marahil ang isang umaambag ay hindi na mag-aambag muli — maaari mong gamitin ang git remote remove o git remote rm : $ git remote remove paul $ git remote origin Kapag natanggal mo na ang reperensiya sa isang remote sa ganitong paraan, lahat ng remote-tracking na mga branch at mga configuration setting na may kaugnayan sa remote ay tinanggal din. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/categories/web-hosting?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_subtitle-click
Best Web Hosting | Products | LinkedIn Skip to main content LinkedIn Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Clear text Used by Used by Web Developer (51) Web Designer (41) Owner (26) Information Technology Specialist (19) Information Technology Manager (13) See all products Find top products in Web Hosting category Software used to make websites accessible on the Internet through the rental of server space. - Use remote storage for shared, reseller, dedicated, virtual private server, colocation, or cloud hosting - Make sites available per storage limits, bandwidth allowance, and with minimum downtime for server maintenance - Integrate with file and content management systems, server-side code, databases, and e-commerce platforms - Keep site secure with backups, firewalls, and malware detection Can include domain name registration, email accounts, and website builders. 901 results IPv4 Address Leasing Service Web Hosting by LARUS Limited LARUS Limited offers leasing of IPv4 addresses, which are unique numerical identifiers assigned to devices connected to the internet. Leasing IPv4 addresses allows businesses to obtain the necessary IP addresses they need for their online operations without having to purchase them outright. This can be a cost-effective solution for companies that need a temporary or flexible supply of IP addresses. LARUS Limited offers a range of leasing options to suit different needs, with competitive pricing and reliable technical support. Lease IPv4 addresses directly from LARUS's pools and use them like your own without limitation regarding geography and usage. View product Pantheon Platform Web Hosting by Pantheon Pantheon is the only WebOps platform built to unify web infrastructure, workflows, and governance for Drupal, WordPress and Next.js. By eliminating outdated tech and tedious sysadmin tasks, Pantheon frees your team to focus on delivering exceptional results and creating extraordinary digital experiences. Trusted by innovative brands like Tableau, Uber, Okta, Home Depot, and Doctors without Borders, Pantheon powers over 700,000 websites and drives 17 billion monthly pageviews. With a lightning-fast content delivery network, always-on security, and effortless scalability, we empower web teams to build, iterate, and launch with confidence. View product Elastic Metal Web Hosting by Scaleway Elastic Metal combines powerful dedicated servers, with flexible pricing, and native integration into Scaleway complete cloud ecosystem. With this range of Bare Metal, you get full control of the resources and applications installed on the server, for maximum performance and security. Elastic Metal servers are available on two different billing plans: hourly or monthly. View product Web Hosting Web Hosting by SiteGround At SiteGround, we offer fast and secure Website Hosting for personal and business websites with agenda to succeed. - Easy Website start or transfer - Daily Backups, Out-of-the-box Caching, Free SSL & CDN - Optimized for popular CMS applications - Advanced solutions for Ecommerce - Expert Hosting Support available 24/7 - Trusted by the owners of over 2,000,000 domains View product WooCommerce Hosting Web Hosting by SiteGround SiteGround's Managed WooCommerce Hosting is designed to help clients sell more online. It is suitable for both new WooCommerce stores, as well as established Woo sites. With SiteGround you get: - WooCommerce pre-installed - Managed WP & Woo Updates - Fast performance thanks to free CDN and smart caching - Top security with our free SSL, AI anti-bot system, custom WAF - High availability with our Daily Backups and restore functionality - 24/7 Expert Hosting Support and much more. View product Find products trusted by professionals in your network See which products are used by connections in your network and those that share similar job titles Sign in to view full insights Cloudways Autonomous Web Hosting by Cloudways Experience fully-managed WordPress hosting that auto-scales with your growing business. Launch your app in under a minute on a self scaling powerful infrastructure that is pre-configured to scale up & down acccording to the traffic demand. For the tech-savvy, this infrastructure is built on Kubernetes with integrated Cloudflare Enterprise & Object Cache Pro's to help ensure to-notch performance & security Enjoy free 24/7/365 technical support, so you're never left alone when you need assitance. Cloudways Autonomous makes high-performance WordPress hosting simple and accessible, combining advanced infrastructure with user-friendly management. Join the 100,000+ SMBs who scale with Cloudways! View product Sites Web Hosting by Appwrite Sites: The open-source Vercel alternative. ✅ Deploy your site: Monitor your deployment logs to ensure a stable connection with the Appwrite Network. ✅ Integrate with your backend: Easily integrate Appwrite's backend products with your site. Zero configuration needed. ✅ Scale safely and deliver widely: Make your product available anywhere in the world through our Global CDN. View product Front-end hosting Web Hosting by Contentstack Accelerate digital delivery with automated, MACH-compliant hosting. Publish updates in minutes, auto-redeploy across regions and ensure consistently fast experiences through global cache refresh. View product Leaseweb Colocation Web Hosting by Leaseweb Our colocation solutions allow you to interconnect with cloud providers and within the Leaseweb Network to create hybrid architectures. View product one.com Web Hosting Web Hosting by one.com Web hosting with one.com is affordable, fast and secure to keep your website up and running. Free SSL certificate and 24/7 support are included in all plans. View product See more How it works Explore Discover the best product for your need from a growing catalog of 25,000 products and categories trusted by LinkedIn professionals Learn Evaluate new tools, explore trending products in your industry and see who in your network is skilled in the product Grow Join communities of product users to learn best practices, celebrate your progress and accelerate your career LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English Language
2026-01-13T09:29:19
https://www.linkedin.com/products/categories/web-hosting?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_subtitle-click
Best Web Hosting | Products | LinkedIn Skip to main content LinkedIn Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Clear text Used by Used by Web Developer (51) Web Designer (41) Owner (26) Information Technology Specialist (19) Information Technology Manager (13) See all products Find top products in Web Hosting category Software used to make websites accessible on the Internet through the rental of server space. - Use remote storage for shared, reseller, dedicated, virtual private server, colocation, or cloud hosting - Make sites available per storage limits, bandwidth allowance, and with minimum downtime for server maintenance - Integrate with file and content management systems, server-side code, databases, and e-commerce platforms - Keep site secure with backups, firewalls, and malware detection Can include domain name registration, email accounts, and website builders. 901 results IPv4 Address Leasing Service Web Hosting by LARUS Limited LARUS Limited offers leasing of IPv4 addresses, which are unique numerical identifiers assigned to devices connected to the internet. Leasing IPv4 addresses allows businesses to obtain the necessary IP addresses they need for their online operations without having to purchase them outright. This can be a cost-effective solution for companies that need a temporary or flexible supply of IP addresses. LARUS Limited offers a range of leasing options to suit different needs, with competitive pricing and reliable technical support. Lease IPv4 addresses directly from LARUS's pools and use them like your own without limitation regarding geography and usage. View product Pantheon Platform Web Hosting by Pantheon Pantheon is the only WebOps platform built to unify web infrastructure, workflows, and governance for Drupal, WordPress and Next.js. By eliminating outdated tech and tedious sysadmin tasks, Pantheon frees your team to focus on delivering exceptional results and creating extraordinary digital experiences. Trusted by innovative brands like Tableau, Uber, Okta, Home Depot, and Doctors without Borders, Pantheon powers over 700,000 websites and drives 17 billion monthly pageviews. With a lightning-fast content delivery network, always-on security, and effortless scalability, we empower web teams to build, iterate, and launch with confidence. View product Elastic Metal Web Hosting by Scaleway Elastic Metal combines powerful dedicated servers, with flexible pricing, and native integration into Scaleway complete cloud ecosystem. With this range of Bare Metal, you get full control of the resources and applications installed on the server, for maximum performance and security. Elastic Metal servers are available on two different billing plans: hourly or monthly. View product Web Hosting Web Hosting by SiteGround At SiteGround, we offer fast and secure Website Hosting for personal and business websites with agenda to succeed. - Easy Website start or transfer - Daily Backups, Out-of-the-box Caching, Free SSL & CDN - Optimized for popular CMS applications - Advanced solutions for Ecommerce - Expert Hosting Support available 24/7 - Trusted by the owners of over 2,000,000 domains View product WooCommerce Hosting Web Hosting by SiteGround SiteGround's Managed WooCommerce Hosting is designed to help clients sell more online. It is suitable for both new WooCommerce stores, as well as established Woo sites. With SiteGround you get: - WooCommerce pre-installed - Managed WP & Woo Updates - Fast performance thanks to free CDN and smart caching - Top security with our free SSL, AI anti-bot system, custom WAF - High availability with our Daily Backups and restore functionality - 24/7 Expert Hosting Support and much more. View product Find products trusted by professionals in your network See which products are used by connections in your network and those that share similar job titles Sign in to view full insights Cloudways Autonomous Web Hosting by Cloudways Experience fully-managed WordPress hosting that auto-scales with your growing business. Launch your app in under a minute on a self scaling powerful infrastructure that is pre-configured to scale up & down acccording to the traffic demand. For the tech-savvy, this infrastructure is built on Kubernetes with integrated Cloudflare Enterprise & Object Cache Pro's to help ensure to-notch performance & security Enjoy free 24/7/365 technical support, so you're never left alone when you need assitance. Cloudways Autonomous makes high-performance WordPress hosting simple and accessible, combining advanced infrastructure with user-friendly management. Join the 100,000+ SMBs who scale with Cloudways! View product Sites Web Hosting by Appwrite Sites: The open-source Vercel alternative. ✅ Deploy your site: Monitor your deployment logs to ensure a stable connection with the Appwrite Network. ✅ Integrate with your backend: Easily integrate Appwrite's backend products with your site. Zero configuration needed. ✅ Scale safely and deliver widely: Make your product available anywhere in the world through our Global CDN. View product Front-end hosting Web Hosting by Contentstack Accelerate digital delivery with automated, MACH-compliant hosting. Publish updates in minutes, auto-redeploy across regions and ensure consistently fast experiences through global cache refresh. View product Leaseweb Colocation Web Hosting by Leaseweb Our colocation solutions allow you to interconnect with cloud providers and within the Leaseweb Network to create hybrid architectures. View product one.com Web Hosting Web Hosting by one.com Web hosting with one.com is affordable, fast and secure to keep your website up and running. Free SSL certificate and 24/7 support are included in all plans. View product See more How it works Explore Discover the best product for your need from a growing catalog of 25,000 products and categories trusted by LinkedIn professionals Learn Evaluate new tools, explore trending products in your industry and see who in your network is skilled in the product Grow Join communities of product users to learn best practices, celebrate your progress and accelerate your career LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English Language
2026-01-13T09:29:19
https://www.linkedin.com/products/rd-station-rd-station-crm/
RD Station CRM | LinkedIn Skip to main content LinkedIn RD Station in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in RD Station CRM Customer Relationship Management (CRM) Software by RD Station See who's skilled in this Add as skill Request demo Report this product About Lançado em 2018, o RD Station CRM é a ferramenta para controlar e organizar o processo comercial. O software ajuda gestores de vendas a ter uma visão completa do funil, fazer follow-up de tarefas e ter mais visibilidade dos resultados a partir de relatórios automáticos. Foi o primeiro CRM gratuito em português com um número ilimitado de usuários e contatos. Hoje, já registra a marca de 3 mil clientes, além de ser o mais bem avaliado na plataforma B2B Stack. Media Products media viewer No more previous content Conheça o RD Station CRM Como a Foco Turismo passou a vender até 100% mais com o RD Station CRM O que é um CRM? No more next content Featured customers of RD Station CRM Total Express Transportation, Logistics, Supply Chain and Storage 209,470 followers Consórcio Magalu Financial Services 16,861 followers Happy IT Services and IT Consulting 24,765 followers PROCAVE Investimentos e Incorporações Construction 22,365 followers Similar products Sales Cloud Sales Cloud Customer Relationship Management (CRM) Software Zoho CRM Zoho CRM Customer Relationship Management (CRM) Software Bigin by Zoho CRM Bigin by Zoho CRM Customer Relationship Management (CRM) Software Experian DataShare Experian DataShare Customer Relationship Management (CRM) Software Odoo CRM Odoo CRM Customer Relationship Management (CRM) Software Freshsales Freshsales Customer Relationship Management (CRM) Software Sign in to see more Show more Show less RD Station products RD Station Conversas RD Station Conversas Chatbot Software RD Station Marketing RD Station Marketing Marketing Automation Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/az/v2/Git-Al%c9%99tl%c9%99ri-Reviziya-Se%c3%a7imi
Git - Reviziya Seçimi About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlanğıc 1.1 Versiyaya Nəzarət Haqqında 1.2 Git’in Qısa Hekayəsi 1.3 Git Nədir? 1.4 Əmr Sətiri 1.5 Git’i Quraşdırmaq 1.6 İlk Dəfə Git Quraşdırması 1.7 Kömək Almaq 1.8 Qısa Məzmun 2. Git’in Əsasları 2.1 Git Deposunun Əldə Edilməsi 2.2 Depoda Dəyişikliklərin Qeyd Edilməsi 2.3 Commit Tarixçəsinə Baxış 2.4 Ləğv Edilən İşlər (Geri qaytarılan) 2.5 Uzaqdan İşləmək 2.6 Etiketləmə 2.7 Git Alias’lar 2.8 Qısa Məzmun 3. Git’də Branch 3.1 Nutshell’də Branch’lar 3.2 Sadə Branching və Birləşdirmə 3.3 Branch İdarəedilməsi 3.4 Branching İş Axınları 3.5 Uzaq Branch’lar 3.6 Rebasing 3.7 Qısa Məzmun 4. Server’də Git 4.1 Protokollar 4.2 Serverdə Git Əldə Etmək 4.3 Sizin öz SSH Public Key’nizi yaratmaq 4.4 Server qurmaq 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Tərəf Seçimləri 4.10 Qısa Məzmun 5. Paylanmış Git 5.1 Distribyutorluq İş Axınları 5.2 Layihəyə Töhfə vermək 5.3 Layihənin Saxlanılması 5.4 Qısa Məzmun 6. GitHub 6.1 Hesab Qurma və Konfiqurasiya 6.2 Bir Layihəyə Töhfə Vermək 6.3 Bir Layihənin Saxlanılması 6.4 Bir Təşkilatı Idarə Etmək 6.5 GitHub Skriptləmə 6.6 Qısa Məzmun 7. Git Alətləri 7.1 Reviziya Seçimi 7.2 Interaktiv Səhnələşdirmə 7.3 Stashing və Təmizləmə 7.4 İşinizin İmzalanması 7.5 Axtarış 7.6 Tarixi Yenidən Yazmaq 7.7 Reset Demystified 7.8 İnkişaf etmiş Birləşmə 7.9 Rerere 7.10 Git ilə Debugging 7.11 Alt Modullar 7.12 Bundling 7.13 Dəyişdirmək 7.14 Etibarlı Yaddaş 7.15 Qısa Məzmun 8. Git’i Fərdiləşdirmək 8.1 Git Konfiqurasiyası 8.2 Git Atributları 8.3 Git Hook’ları 8.4 Git-Enforced Siyasət Nümunəsi 8.5 Qısa Məzmun 9. Git və Digər Sistemlər 9.1 Git Müştəri kimi 9.2 Git’ə Miqrasiya 9.3 Qısa Məzmun 10. Git’in Daxili İşləri 10.1 Plumbing və Porcelain 10.2 Git Obyektləri 10.3 Git Referansları 10.4 Packfile’lar 10.5 Refspec 10.6 Transfer Protokolları 10.7 Maintenance və Məlumatların Bərpası 10.8 Mühit Dəyişənləri 10.9 Qısa Məzmun A1. Appendix A: Digər Mühitlərdə Git A1.1 Qrafik interfeyslər A1.2 Visual Studio’da Git A1.3 Visual Studio Code’da Git A1.4 Eclipse’də Git A1.5 Sublime Text’də Git A1.6 Bash’da Git A1.7 Zsh’də Git A1.8 PowerShell’də Git A1.9 Qısa Məzmun A2. Appendix B: Proqramlara Git Daxil Etmək A2.1 Əmr-sətri Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Əmrləri A3.1 Quraşdırma və Konfiqurasiya A3.2 Layihələrin Alınması və Yaradılması A3.3 Sadə Snapshotting A3.4 Branching və Birləşmə A3.5 Layihələrin Paylaşılması və Yenilənməsi A3.6 Yoxlama və Müqayisə A3.7 Debugging A3.8 Patching A3.9 E-poçt A3.10 Xarici Sistemlər A3.11 İdarəetmə A3.12 Plumbing Əmrləri 2nd Edition 7.1 Git Alətləri - Reviziya Seçimi İndiyə qədər mənbə kodu nəzarəti üçün Git deposunu idarə etməli və ya saxlamağınız lazım olan gündəlik əmrlərin və iş axınlarının çoxunu öyrəndiniz. Faylları tracking və committing əsas tapşırıqlarını yerinə yetirdiniz və quruluş sahəsinin gücünü və yüngül mövzunun branching’ni və birləşməsini istifadə etdiniz. İndi Git’in gündəlik olaraq istifadə edə bilməyəcəyiniz, ancaq bir anda ehtiyacınız ola biləcəyi çox güclü şeyləri araşdıracaqsınız. Reviziya Seçimi Git, bir sıra commit-lərə, commit-lər dəstinə və ya commit-lərə istinad etməyə imkan verir. Bunlar mütləq açıq deyil, bilmək faydalıdır. Tək Reviziyalar Tamamilə 40 xarakterli SHA-1 hash ilə hər hansı bir commit-ə istinad edə bilərsiniz, lakin commit-ləri ifadə etmənin daha çox insan dostu yolları var. Bu bölüm, hər hansı bir commit-ə istinad edə biləcəyiniz müxtəlif yolları əks etdirir. Qısa SHA-1 Git, SHA-1 hash-nın ilk bir neçə simvolunu verdiyiniz təqdirdə nəyi nəzərdə tutduğunuzu başa düşmək üçün kifayət qədər ağıllıdır, qismən qarışıq ən azı dörd simvol uzun və birmənalıdır; Başqa sözlə, obyekt verilənlər bazasındakı heç bir obyektdə eyni prefikslə başlayan bir hash ola bilməz. Məsələn, müəyyən bir funksionallıq əlavə etdiyinizi bildiyiniz xüsusi bir commit-i araşdırmaq üçün əvvəlcə commit-i tapmaq üçün git log əmrini işə sala bilərsiniz: $ git log commit 734713bc047d87bf7eac9674765ae793478c50d3 Author: Scott Chacon <schacon@gmail.com> Date: Fri Jan 2 18:32:33 2009 -0800 Fix refs handling, add gc auto, update tests commit d921970aadf03b3cf0e71becdaab3147ba71cdef Merge: 1c002dd... 35cfb2b... Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' commit 1c002dd4b536e7479fe34593e72e6c6c1819e53b Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 14:58:32 2008 -0800 Add some blame and merge stuff Bu vəziyyətdə, hash 1c002dd... ilə başlayan commit-lə maraqlandığınızı varsayaq. Aşağıdakı git show varyasyonlarından hər hansı biri ilə əlaqəli olanı yoxlaya bilərsiniz (daha qısa versiyaların birmənalı olduğunu düşünərək): $ git show 1c002dd4b536e7479fe34593e72e6c6c1819e53b $ git show 1c002dd4b536e7479f $ git show 1c002d Git, SHA-1 dəyərləriniz üçün qısa, bənzərsiz bir qısaltmanı müəyyən edə bilər. --abbrev-commit git log əmrinə keçsəniz, çıxış daha qısa dəyərlərdən istifadə edəcək, lakin onları unikal saxlayır; yeddi simvol istifadə etmək üçün standartdır, lakin SHA-1-in birmənalı olması üçün onları daha uzun edir: $ git log --abbrev-commit --pretty=oneline ca82a6d Change the version number 085bb3b Remove unnecessary test code a11bef0 Initial commit Ümumiyyətlə, səkkiz-on simvol bir proyektdə bənzərsiz olmaq üçün kifayətdir. Məsələn, 2019-cu ilin fevral ayından etibarən, Linux kernelinin (olduqca əhəmiyyətli bir layihədir) 875.000-dən çox commit-i və obyekt bazasında təxminən yeddi milyon obyekti var, ilk 12 simvolda SHA-1’ləri eyni olan iki obyekt yoxdur. Note SHA-1 HAQQINDA QISA QEYD Bir çox insan təsadüfi bir şəkildə, eyni SHA-1 dəyərinə qarışan depolarında iki fərqli obyektə sahib olacaqlarından bir anda narahat olurlar. Bəs onda nə etmək lazımdır? Əgər deponuzdakı əvvəlki fəərqli obyekti ilə eyni SHA-1 dəyərinə bərabər olan bir obyekt törətmisinizsə, Git əvvəlki obyekti Git verilənlər bazanızda görəcək, artıq yazıldığını düşünün və sadəcə yenidən istifadə edin. Bir nöqtədə yenidən həmin obyekti yoxlamağa çalışsanız, həmişə ilk obyektin məlumatlarını əldə edəcəksiniz. Bununla birlikdə, bu ssenarinin nə qədər gülünc bir şəkildə ehtimal olunmadığının fərqində olmalısınız. SHA-1 həcmi 20 bayt və ya 160 bitdir. Tək bir toqquşma ehtimalının 50% olmasını təmin etmək üçün lazım olan təsadüfi yığılmış obyektlərin sayı təxminən 2 80 -dir (toqquşma ehtimalını müəyyənləşdirmək üçün düstur p = (n(n-1)/2) * (1/2^160)) . 2 80 1.2 x 10 24 təşkil edir və ya 1 milyon milyard. Bu, yer üzündə qum dənələrinin sayından 1,200 dəfə çoxdur. Burada SHA-1 toqquşması üçün nə lazım olduğunu düşünmək üçün bir nümunə var. Yer üzündəki 6,5 milyard insanın hamısı proqramlaşdırma aparsaydı və hər saniyədə hər biri bütün Linux nüvə tarixinə (6.5 milyon Git obyekt) bərabər olan bir kod istehsal etsəydi və onu böyük bir Git deposuna salsaydı, təxminən 2 il çəkərdi. Bu depoda bir SHA-1 obyektinin toqquşma ehtimalı 50% -ə çatacaq qədər obyekt var qədər. Beləliklə, SHA-1 toqquşması, proqramlaşma komandanızın hər bir üzvünün eyni gecədə əlaqəsi olmayan hadisələrdə canavarların hücumuna məruz qalması və öldürülməsi ehtimalı daha azdır. Branch Referansları Müəyyən bir commit-ə istinad etməyin bir sadə yolu branch-ın ucundakı commit-in olmasıdır; bu halda, sadəcə bir commit-ə istinad gözləyən hər hansı bir Git əmrində branch adını istifadə edə bilərsiniz. Məsələn, bir branch-dakı son commit obyektini araşdırmaq istəyirsinizsə, aşağıdakı mövzuda əmrlər ekvivalentdir ki, topic1 branch-ının ca82a6d... işarə etdiyini göstərir: $ git show ca82a6dff817ec66f44342007202690a93763949 $ git show topic1 Bir branch-ın hansı konkret SHA-1-ə işarə etdiyini görmək və ya bu nümunələrdən hər hansı birinin SHA-1-lər baxımından nəyə bənzədiyini görmək istəyirsinizsə, rev-parse adlı Git santexnika alətindən istifadə edə bilərsiniz. Santexnika alətləri haqqında daha çox məlumat üçün Git’in Daxili İşləri -ə baxa bilərsiniz; əsasən, rev-parse aşağı səviyyəli əməliyyatlar üçün mövcuddur və gündəlik əməliyyatlarda istifadə üçün nəzərdə tutulmayıb. Ancaq bəzən həqiqətən nələrin baş verdiyini görmək lazım olduqda faydalı ola bilər. Burada branch-ınızda rev-parse işlədə bilərsiniz. $ git rev-parse topic1 ca82a6dff817ec66f44342007202690a93763949 RefLog Qısa Adları Gitin arxada işləyərkən arxa planda gördüyü işlərdən biri də reflog saxlamaqdır - HEAD və branch istinadlarınızın son bir neçə ayda olduğu bir qeyd. Reflogunuzu git reflog istifadə edərək görə bilərsiniz: $ git reflog 734713b HEAD@{0}: commit: Fix refs handling, add gc auto, update tests d921970 HEAD@{1}: merge phedders/rdocs: Merge made by the 'recursive' strategy. 1c002dd HEAD@{2}: commit: Add some blame and merge stuff 1c36188 HEAD@{3}: rebase -i (squash): updating HEAD 95df984 HEAD@{4}: commit: # This is a combination of two commits. 1c36188 HEAD@{5}: rebase -i (squash): updating HEAD 7e05da5 HEAD@{6}: rebase -i (pick): updating HEAD Branch ucunuz hər hansı bir səbəbdən yeniləndikdə Git bu məlumatları sizin üçün bu müvəqqəti tarixdə saxlayır. Köhnə commit-lərə də istinad etmək üçün reflog məlumatlarınızı istifadə edə bilərsiniz. Məsələn, deposunuzun HEAD-in əvvəlki beşinci dəyərini görmək istəyirsinizsə, reflog çıxışında gördüyünüz @{5} istinadından istifadə edə bilərsiniz: $ git show HEAD@{5} Bu sintaksisdən branch-ın müəyyən bir müddət əvvəl harada olduğunu görmək üçün də istifadə edə bilərsiniz. Məsələn, dünən master branch-ınızın harada olduğunu görmək üçün yaza bilərsiniz: $ git show master@{yesterday} Bu, dünən " master branch-ınızın ucunun harada olduğunu göstərəcəkdir. Bu texnika yalnız hələ də qeydlərinizdə olan məlumatlar üçün işləyir, buna görə də bir neçə aydan yuxarı commit-lər axtarmaq üçün istifadə edə bilməzsiniz. Reflog məlumatlarını git log çıxışı kimi formatlanmış şəkildə görmək üçün git log -g işlədə bilərsiniz: $ git log -g master commit 734713bc047d87bf7eac9674765ae793478c50d3 Reflog: master@{0} (Scott Chacon <schacon@gmail.com>) Reflog message: commit: Fix refs handling, add gc auto, update tests Author: Scott Chacon <schacon@gmail.com> Date: Fri Jan 2 18:32:33 2009 -0800 Fix refs handling, add gc auto, update tests commit d921970aadf03b3cf0e71becdaab3147ba71cdef Reflog: master@{1} (Scott Chacon <schacon@gmail.com>) Reflog message: merge phedders/rdocs: Merge made by recursive. Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' Qeyd etmək vacibdir ki, reflog məlumatları ciddi şəkildə localdır - bu, yalnız deponuzda etdiyiniz işlərin bir qeydidir. Referanslar başqasının deposunun kopyasında eyni olmayacaq; ayrıca, əvvəlcə bir deponu klonladıqdan dərhal sonra, depoda hələ heç bir fəaliyyət baş vermədiyi üçün boş bir refloqa sahib olacaqsınız. git show HEAD@{2.months.ago} -ı işə salmaq, sizə yalnız ən azı iki ay əvvəl layihəni klonlaşdırdığınız təqdirdə uyğunlaşma commit-ini göstərəcəkdir - daha yaxınlarda klonlaşdırsanız, yalnız ilk local commit-i görəcəksiniz. Tip Reflogu Git-in shell tarixinin versiyası kimi düşünün UNIX və ya Linux arxa planınız varsa, reflog-u Git-in shell tarixinin versiyası olaraq düşünə bilərsiniz, burada olanların yalnız sizin və sizin “sessiyanız” üçün açıq şəkildə əlaqəli olduğunu vurğulayan və eyni maşında işləyə başqa heç kimlə ilə əlaqəsi yoxdur. Ancestry Referansları Bir commit-i müəyyənləşdirməyin digər əsas yolu əcdadı ilə bağlıdır. Bir referansın sonunda bir ^ (caret) qoysanız, Git, bu commit-in valideynini ifadə etmək üçün onu təhlil edir. Tutaq ki, layihənizin tarixinə nəzər yetirdiniz: $ git log --pretty=format:'%h %s' --graph * 734713b Fix refs handling, add gc auto, update tests * d921970 Merge commit 'phedders/rdocs' |\ | * 35cfb2b Some rdoc changes * | 1c002dd Add some blame and merge stuff |/ * 1c36188 Ignore *.gem * 9b29157 Add open3_detach to gemspec file list Daha sonra, “the parent of HEAD” mənasını verən HEAD^ göstərərək əvvəlki commit-i görə bilərsiniz: $ git show HEAD^ commit d921970aadf03b3cf0e71becdaab3147ba71cdef Merge: 1c002dd... 35cfb2b... Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 15:08:43 2008 -0800 Merge commit 'phedders/rdocs' Note Craet-i Windows-dan xilas etmək Windows-da cmd.exe , ^ xüsusi bir xarakter daşıyır və fərqli davranılmalıdır. Ya ikiqat edə bilərsiniz, ya da commit arayışını quote-lara daxil edə bilərsiniz: $ git show HEAD^ # will NOT work on Windows $ git show HEAD^^ # OK $ git show "HEAD^" # OK İstədiyiniz valideynin hansı olduğunu müəyyən etmək üçün ^ -dən sonra bir rəqəm də göstərə bilərsiniz; məsələn, d921970^2 “d921970-in ikinci valideynidir.” deməkdir. Bu sintaksis yalnız birdən çox valideynə sahib olan birləşmə commit-ləri üçün faydalıdır - birləşdirmə commit-inin birinci valideyni birləşdikdə olduğunuz branch-dan (tez-tez master ), birləşmə commit-nin ikinci valideyn hissəsi isə birləşdirilmiş branch-dan ( topic deyək): $ git show d921970^ commit 1c002dd4b536e7479fe34593e72e6c6c1819e53b Author: Scott Chacon <schacon@gmail.com> Date: Thu Dec 11 14:58:32 2008 -0800 Add some blame and merge stuff $ git show d921970^2 commit 35cfb2b795a55793d7cc56a6cc2060b4bb732548 Author: Paul Hedderly <paul+git@mjr.org> Date: Wed Dec 10 22:22:03 2008 +0000 Some rdoc changes Digər əsas əcdad spesifikasiyası ~ (tilde)-dir. Bu da birinci valideynə aiddir, buna görə HEAD~ və HEAD^ bərabərdir. Fərq bir rəqəm göstərdiyiniz zaman aydın olur. HEAD~2 , “ilk valideynin birinci valideyni” və ya “nənə və baba” deməkdir - ilk valideynlərə göstərdiyiniz vaxt keçir. Məsələn, əvvəllər sadalanan tarixdə HEAD~3 : $ git show HEAD~3 commit 1c3618887afb5fbcbea25b7c013f4e2114448b8d Author: Tom Preston-Werner <tom@mojombo.com> Date: Fri Nov 7 13:47:59 2008 -0500 Ignore *.gem Yenidən ilk valideynin ilk valideyninin ilk valideyni olan HEAD~~~ yazıla bilər: $ git show HEAD~~~ commit 1c3618887afb5fbcbea25b7c013f4e2114448b8d Author: Tom Preston-Werner <tom@mojombo.com> Date: Fri Nov 7 13:47:59 2008 -0500 Ignore *.gem Bu sintaksisləri də birləşdirə bilərsiniz - əvvəlcədən istinadın ikinci əsas hissəsini (birləşdirmə əmri olduğunu düşünərək) HEAD~3^2 və s. istifadə edərək əldə edə bilərsiniz. Commit Aralıqları İndi fərdi commit-lər təyin edə bildiyinizə görə, commit-lərin hüdudlarını necə təyin edəcəyimizə baxaq. Bu, branch-larınızı idarə etmək üçün xüsusilə faydalıdır - çox sayda branch-ınız varsa, “Bu branch-da hələ əsas branch-a birləşdirmədiyim hansı iş var?” kimi suallara cavab vermək üçün spesifikasiyalardan istifadə edə bilərsiniz. Cüt nöqtə Ən geniş yayılmış spesifikasiya cüt nöqtəli sintaksisdir. Bu, əsasən Git-dən bir commit-dən əldə edilə bilən, digərinə çatmayan bir sıra commit-ləri həll etməsini xahiş edir. Məsələn, Aralıq seçimi üçün nümunə tarixçəsi kimi görünən commit tarixçəniz olduğunu söyləyin. Figure 137. Aralıq seçimi üçün nümunə tarixçəsi experiment branch-ınızda hələ master branch-ınıza birləşdirilməyənləri görmək istədiyinizi söyləyin. Git-dən sizə yalnız master..experiment ilə işləyənlərin bir jurnalını göstərməsini xahiş edə bilərsiniz - bu, ‘` master -dən əldə edilə bilməyən experiment -dən əldə edilə bilən bütün commit-lər’' deməkdir. Bu nümunələrdə qısalıq və aydınlıq üçün, diaqramdakı commit obyektlərinin hərfləri göstərəcəkləri qaydada həqiqi log çıxışı yerinə istifadə olunur: $ git log master..experiment D C Digər tərəfdən bunun əksini görmək istəyirsinizsə - bütün commit-lər experiment -də olmayan master -də işləyirsə - branch adlarını tərsinə çevirə bilərsiniz. experiment..master sizə experiment -dən əlçatmaz olan hər şeyi master -də göstərir: $ git log experiment..master F E Bu, experiment branch-ını yeniləmək və birləşdirmək istədiklərinizi önizləmək istəsəniz faydalıdır. Bu sintaksisin başqa bir tez-tez istifadəsi uzaq məsafəyə nəyi push edəcəyinizi görməkdir: $ git log origin/master..HEAD Bu əmr sizə cari branch-ınızdakı origin remote-dakı master branch-ında olmayan hər hansı bir commit-i göstərir. Bir git push işə salırsınızsa və mövcud branch-ınız origin/master izləyirsə, git log origin/master..HEAD tərəfindən sadalanan commit-lər serverə ötürülən commit-lərdir. Git-in HEAD olduğunu qəbul etməsi üçün sintaksisin bir tərəfini də tərk edə bilərsiniz. Məsələn, git log origin/master.. yazaraq əvvəlki nümunədəki ilə eyni nəticələr əldə edə bilərsiniz - bir tərəfi yoxdursa, HEAD əvəzlə. Birdən Çox Pal İkili nöqtəli sintaksis shorthand kimi faydalıdır, lakin bəlkə də hazırda olduğunuz branch-da olmayan bir neçə branch-dan birinin nə olduğunu görmək kimi düzəlişlərinizi göstərmək üçün ikidən çox branch göstərmək istəyirsiniz. Git, əlçatan commit-lər görmək istəmədiyiniz hər hansı bir istinaddan əvvəl ^ simvolunu və ya --not istifadə edərək bunu etməyə imkan verir. Beləliklə, aşağıdakı üç əmr bərabərdir: $ git log refA..refB $ git log ^refA refB $ git log refB --not refA Bu çox yaxşıdır, çünki bu sintaksislə sorğunuzda ikiqat nöqtəli sintaksis ilə edə bilməyəcəyiniz ikidən çox istinad daxil edə bilərsiniz. Məsələn, refA ya da refB -dən əldə edilə bilən, ancaq refC -dən edilməyən, bütün commit-ləri görmək istəyirsinizsə, aşağıdakılardan birini istifadə edə bilərsiniz: $ git log refA refB ^refC $ git log refA refB --not refC Bu, branch-larınızda nə olduğunu anlamanıza kömək edəcək çox güclü bir revizyon sorğu sistemi yaradır. Üçqat Nöqtə Son böyük aralıq seçmə sintaksisi, hər ikisindən deyil, iki istinadın hər ikisi tərəfindən də əldə edilə bilən bütün commit-ləri təyin edən üç nöqtəli sintaksisdir. Aralıq seçimi üçün nümunə tarixçəsi -dakı commit tarixçəsinə baxın. master və ya experiment -də olanları görmək istəsəniz, lakin ümumi istinadları yox görmək istəməsəniz, işlədə bilərsiniz: $ git log master...experiment F E D C Yenə də, bu sizə normal bir log çıxışı verir, ancaq ənənəvi commit tarixi sifarişində görünən yalnız bu dörd commit üçün commit məlumatlarını göstərir. Bu halda log əmri ilə istifadə olunan ümumi bir keçid, hər bir commit aralığın hansı tərəfində olduğunu göstərən --left-right -dır. Bu, nəticənin daha faydalı olmasına kömək edir: $ git log --left-right master...experiment < F < E > D > C Bu vasitələrlə Git-ə nəyi yoxlamaq istədiyinizi və ya commit-lərinizi daha asanlıqla bildirə bilərsiniz. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/radware-attack-mitigation-system/?trk=products_seo_search
Attack Mitigation System | LinkedIn Skip to main content LinkedIn Radware in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Attack Mitigation System DDoS Protection Software by Radware See who's skilled in this Add as skill Learn more Report this product About Continuously adaptive real-time DDoS services for the most sophisticated web security threats through best-in-class cloud WAF and DDoS protection technologies. Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less Radware products APSolute Vision APSolute Vision Network Management Software Radware Alteon Radware Alteon Load Balancing Software Radware Cloud DDoS Protection Service Radware Cloud DDoS Protection Service DDoS Protection Software Radware Cloud WAF Radware Cloud WAF Web Application Firewalls (WAF) Radware Cloud Workload Protection Radware Cloud Workload Protection Cloud Workload Protection Platforms Radware DefensePro Radware DefensePro DDoS Protection Software Radware Security for Azure Radware Security for Azure DDoS Protection Software Radware Threat Intelligence Radware Threat Intelligence Threat Intelligence Platforms Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/de/v2/Git-Grundlagen-Mit-Remotes-arbeiten
Git - Mit Remotes arbeiten About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Erste Schritte 1.1 Was ist Versionsverwaltung? 1.2 Kurzer Überblick über die Historie von Git 1.3 Was ist Git? 1.4 Die Kommandozeile 1.5 Git installieren 1.6 Git Basis-Konfiguration 1.7 Hilfe finden 1.8 Zusammenfassung 2. Git Grundlagen 2.1 Ein Git-Repository anlegen 2.2 Änderungen nachverfolgen und im Repository speichern 2.3 Anzeigen der Commit-Historie 2.4 Ungewollte Änderungen rückgängig machen 2.5 Mit Remotes arbeiten 2.6 Taggen 2.7 Git Aliases 2.8 Zusammenfassung 3. Git Branching 3.1 Branches auf einen Blick 3.2 Einfaches Branching und Merging 3.3 Branch-Management 3.4 Branching-Workflows 3.5 Remote-Branches 3.6 Rebasing 3.7 Zusammenfassung 4. Git auf dem Server 4.1 Die Protokolle 4.2 Git auf einem Server einrichten 4.3 Erstellung eines SSH-Public-Keys 4.4 Einrichten des Servers 4.5 Git-Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Von Drittanbietern gehostete Optionen 4.10 Zusammenfassung 5. Verteiltes Git 5.1 Verteilter Arbeitsablauf 5.2 An einem Projekt mitwirken 5.3 Ein Projekt verwalten 5.4 Zusammenfassung 6. GitHub 6.1 Einrichten und Konfigurieren eines Kontos 6.2 Mitwirken an einem Projekt 6.3 Ein Projekt betreuen 6.4 Verwalten einer Organisation 6.5 Skripte mit GitHub 6.6 Zusammenfassung 7. Git Tools 7.1 Revisions-Auswahl 7.2 Interaktives Stagen 7.3 Stashen und Bereinigen 7.4 Deine Arbeit signieren 7.5 Suchen 7.6 Den Verlauf umschreiben 7.7 Reset entzaubert 7.8 Fortgeschrittenes Merging 7.9 Rerere 7.10 Debuggen mit Git 7.11 Submodule 7.12 Bundling 7.13 Replace (Ersetzen) 7.14 Anmeldeinformationen speichern 7.15 Zusammenfassung 8. Git einrichten 8.1 Git Konfiguration 8.2 Git-Attribute 8.3 Git Hooks 8.4 Beispiel für Git-forcierte Regeln 8.5 Zusammenfassung 9. Git und andere VCS-Systeme 9.1 Git als Client 9.2 Migration zu Git 9.3 Zusammenfassung 10. Git Interna 10.1 Basisbefehle und Standardbefehle (Plumbing and Porcelain) 10.2 Git Objekte 10.3 Git Referenzen 10.4 Packdateien (engl. Packfiles) 10.5 Die Referenzspezifikation (engl. Refspec) 10.6 Transfer Protokolle 10.7 Wartung und Datenwiederherstellung 10.8 Umgebungsvariablen 10.9 Zusammenfassung A1. Anhang A: Git in anderen Umgebungen A1.1 Grafische Schnittstellen A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Zusammenfassung A2. Anhang B: Git in deine Anwendungen einbetten A2.1 Die Git-Kommandozeile A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Anhang C: Git Kommandos A3.1 Setup und Konfiguration A3.2 Projekte importieren und erstellen A3.3 Einfache Snapshot-Funktionen A3.4 Branching und Merging A3.5 Projekte gemeinsam nutzen und aktualisieren A3.6 Kontrollieren und Vergleichen A3.7 Debugging A3.8 Patchen bzw. Fehlerkorrektur A3.9 E-mails A3.10 Externe Systeme A3.11 Administration A3.12 Basisbefehle 2nd Edition 2.5 Git Grundlagen - Mit Remotes arbeiten Mit Remotes arbeiten Um an Git-Projekte mitarbeiten zu können, musst du wissen, wie du deine Remote-Repositorys verwalten kannst. Remote-Repositorys sind Versionen deines Projekts, die im Internet oder im Netzwerk irgendwo gehostet werden. Du kannst Mehrere davon haben, von denen jedes in der Regel entweder schreibgeschützt oder beschreibbar ist. Die Zusammenarbeit mit Anderen erfordert die Verwaltung dieser Remote-Repositorys sowie das Pushing und Pulling von Daten zu und von den Repositorys, wenn du deine Arbeit teilen möchtest. Die Verwaltung von Remote-Repositorys umfasst das Wissen, wie man entfernte Repositorys hinzufügt, nicht mehr gültige Remotes entfernt, verschiedene Remote-Branches verwaltet, sie als versioniert oder nicht versioniert definiert und vieles mehr. In diesem Abschnitt werden wir einige dieser Remote-Management-Fertigkeiten erörtern. Anmerkung Remote-Repositorys können auch auf deinem lokalen Rechner liegen. Es ist durchaus möglich, dass du mit einem „entfernten“ Repository arbeiten kannst, das sich eigentlich auf demselben Host befindet auf dem du gerade arbeitest. Das Wort „remote“ bedeutet nicht unbedingt, dass sich das Repository an einem anderen Ort im Netzwerk oder Internet befindet, sondern nur, dass es an einem anderen Ort liegt. Die Arbeit mit einem solchen entfernten Repository würde immer noch alle üblichen Push-, Pull- und Fetch-Operationen einschließen, wie bei jedem anderen Remote-Repository. Auflisten der Remotes Um zu sehen, welche Remote-Server bei dir konfiguriert sind, kannst du den Befehl git remote aufrufen. Er listet die Kurznamen der einzelnen von dir festgelegten Remote-Handles auf. Wenn du dein Repository geklont hast, solltest du zumindest origin sehen – das ist der Standardname, den Git dem Server gibt, von dem du geklont hast: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Du kannst zusätzlich auch -v angeben, das dir die URLs anzeigt, die Git für den Kurznamen gespeichert hat, um beim Lesen und Schreiben auf diesen Remote zuzugreifen: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Wenn du mehr als einen Remote hast, listet der Befehl sie alle auf. Ein Repository mit mehreren Remotes für die Arbeit mit mehreren Beteiligten könnte beispielsweise so aussehen. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Das bedeutet, dass wir Beiträge von jedem dieser Benutzer ziemlich einfach abrufen können. Möglicherweise haben wir zusätzlich die Erlaubnis, auf einen oder mehrere von diesen zu pushen, obwohl wir das hier nicht erkennen können. Beachte: Diese Remotes verwenden unterschiedliche Protokollen; wir werden mehr darüber hier erfahren: Git auf einem Server installieren . Hinzufügen von Remote-Repositorys Wir haben bereits erwähnt und einige Beispiele gezeigt, wie der Befehl git clone stillschweigend den Remote origin hinzufügt. Im Folgendem beschreiben wir, wie du einen neuen Remote hinzufügen kannst. Um ein neues Remote-Git-Repository als Kurzname hinzuzufügen, auf das du verweisen kannst, führe git remote add <shortname> <url> aus: $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Nun kannst du die Zeichenfolge pb auf der Kommandozeile anstelle der gesamten URL verwenden. Wenn du beispielsweise alle Informationen fetchen möchtest, die Paul hat, die aber noch nicht in deinem Repository enthalten sind, kannst du git fetch pb ausführen: $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Pauls master Branch ist nun lokal als pb/master erreichbar – Du kannst ihn in eine deiner Branches mergen oder ihn in einen lokalen Branch auschecken, wenn du ihn inspizieren möchten. Wir werden in Git Branching detailliert beschreiben, was Branches sind und wie man sie nutzt. Fetchen und Pullen deiner Remotes Wie du gerade gesehen hast, kannst du Daten aus deinem Remote-Projekt abrufen: $ git fetch <remote> Der Befehl geht an das Remote-Projekt und zieht (engl. pull) alle Daten von diesem Remote-Projekt, die du noch nicht hast. Danach solltest du Referenzen auf alle Branches dieses Remote haben, die du jederzeit mergen oder inspizieren kannst. Wenn du ein Repository klonst, fügt der Befehl dieses entfernte Repository automatisch unter dem Namen „origin“ hinzu. So holt git fetch origin alle Änderungen, die seit dem Klonen (oder dem letzten Abholen) auf diesen Server abgelegt (engl. pushed) wurden. Es ist jedoch wichtig zu beachten, dass der Befehl git fetch nur die Daten in dein lokales Repository herunterlädt – er merged sie nicht automatisch mit deiner Arbeit zusammen oder ändert das, woran du gerade arbeitest. Du musst das Ganze manuell mit deiner Arbeit zusammenführen, wenn du soweit bist. Wenn dein aktueller Branch so eingerichtet ist, dass er einen entfernten Branch verfolgt (engl. tracking), kannst du den Befehl git pull verwenden, um diesen entfernten Branch automatisch zu fetchen und dann mit deinem aktuellen Branch zu mergen (siehe den nächsten Abschnitt und Git Branching für weitere Informationen). Das könnte ein einfacherer oder komfortablerer Workflow sein. Standardmäßig richtet der Befehl git clone deinen lokalen master Branch automatisch so ein, dass er den entfernten master Branch (oder wie auch immer der Standard-Branch genannt wird) auf dem Server versioniert von dem du geklont hast. Wenn du git pull ausführst, werden normalerweise Daten von dem Server abgerufen, von dem du ursprünglich geklont hast. Anschliessend wird automatisch versucht diese Daten in deinem Code zu mergen. Anmerkung Ab der Version 2.27 von Git wird git pull eine Warnung ausgeben, wenn die Variable pull.rebase nicht gesetzt ist. Git wird dich so lange warnen, bis du die Variable setzt. Falls du das Standardverhalten (möglichst ein fast-forward, ansonsten einen Merge-Commit erstellen) von Git beibehalten willst: git config --global pull.rebase "false" Wenn du jedoch mit dem Pullen einen Rebase machen willst: git config --global pull.rebase "true" Zu deinen Remotes Pushen Wenn du dein Projekt an einem Punkt hast, an dem du es teilen möchtest, können wir es zum Upstream verschieben (engl. pushen). Der Befehl dafür lautet: git push <remote> <branch> . Wenn du deinen master Branch auf den origin Server pushen möchtest (Zur Erinnerung: das Klonen richtet im Regelfall diese beiden Branches automatisch ein), dann kannst du diesen Befehl auch nutzen, um alle Commits, die du durchgeführt hast, auf den Server zu pushen: $ git push origin master Dieser Befehl funktioniert allerdings nur, wenn du von einem Server geklont hast, auf den du Schreibzugriff hast und wenn in der Zwischenzeit noch niemand anderes gepusht hat. Wenn du und ein anderer Benutzer gleichzeitig klonen und beide Upstream pushen wollen, du aber etwas später nach Upstream pushst, dann wird dein Push zu Recht abgelehnt. Du musst zuerst dessen Bearbeitung abholen und in deine einbinden, bevor du pushen kannst. Siehe Kapitel 3 Git Branching mit ausführlicheren Details zum Pushen auf Remote-Server. Inspizieren eines Remotes Wenn du mehr Informationen über einen bestimmten Remote sehen möchten, kannst du den Befehl git remote show <remote> verwenden. Wenn du diesen Befehl mit einem remote Kurznamen ausführen, wie z.B. origin , erhältst du bspw. folgende Meldung: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Er listet die URL für das Remote-Repository sowie die Informationen zu den Tracking-Branches auf. Der Befehl teilt dir mit: Wenn du dich im Master-Branch befindest und git pull ausführst, der master Branch des remotes nach dem abrufen (engl. fetched) automatisch mit dem lokalen Branch gemerged wird. Er listet auch alle Remote-Referenzen auf, die er abgerufen hat. Das ist nur ein einfaches Beispiel, auf das du möglicherweise treffen wirst. Wenn du Git hingegen intensiver verwendest, könnte die git remote show Ausgabe wesentlich umfangreicher sein: $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Dieser Befehl zeigt an, zu welchem Branch automatisch gepusht wird, wenn du git push ausführst, während du dich in bestimmten Branches befindest. Er zeigt dir auch, welche entfernten Branches auf dem Server vorhanden sind, die du noch nicht hast, welche entfernten Branches du hast, die aber vom Server gelöscht wurden und die lokalen Branches, die automatisch mit deinen Remote-Tracking-Branch zusammengeführt (gemerged) werden können, wenn du git pull ausführst. Umbenennen und Entfernen von Remotes Du kannst git remote rename ausführen, um den Kurznamen eines Remotes zu ändern. Wenn du beispielsweise pb in paul umbenennen möchtest, kannst du dies mit dem Befehl git remote rename erreichen: $ git remote rename pb paul $ git remote origin paul Es ist zu beachten, dass dadurch auch alle deine Remote-Tracking-Branchnamen geändert werden. Was vorher unter pb/master referenziert wurde, ist jetzt unter paul/master zu finden. Wenn du einen Remote aus irgendwelchen Gründen entfernen möchten – Du hast den Server verschoben oder verwendest einen bestimmten Mirror nicht länger oder ein Beitragender ist nicht mehr dabei – dann kannst du entweder git remote remove oder git remote rm verwenden: $ git remote remove paul $ git remote origin Sobald du die Referenz auf einen Remote auf diese Weise gelöscht hast, werden auch alle mit diesem Remote verbundenen Remote-Tracking-Branches und Konfigurationseinstellungen gelöscht. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-infinistreamng-isng/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_image-click
InfiniStreamNG (ISNG) | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in InfiniStreamNG (ISNG) Business Continuity Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Organizations embracing digital transformation face network and application visibility challenges, regardless of the next-generation platforms involved in such transitions. NETSCOUT provides organizations with the flexibility to deploy the InfiniStreamNG (ISNG) appliance’s technology in a manner best-suited to specific infrastructure, visibility requirements, and budgetary goals. Media Products media viewer No more previous content What is Visibility Without Borders? Visibility Across Any Environment, Any Workload, Any Time You're faced with transforming your technology infrastructure at extraordinary rates in an increasingly complex and fragmented digital universe. At NETSCOUT, we have made an unprecedented investment of time, money and resources to solve this challenge and bring to market a practical solution that produces a common view of your digital services across technology and organizational boundaries, across all networks, all locations and all users. We call this Visibility Without Borders. InfiniStreamNG Hardware InfiniStreamNG Infrastructure No more next content Similar products Business Operations Business Operations Business Continuity Software Continuity Continuity Business Continuity Software BC in the Cloud BC in the Cloud Business Continuity Software SafetyCulture (formerly iAuditor) SafetyCulture (formerly iAuditor) Business Continuity Software أجود أجود Business Continuity Software Arcserve Replication and High Availability (RHA) Arcserve Replication and High Availability (RHA) Business Continuity Software Sign in to see more Show more Show less NETSCOUT products Arbor Edge Defense Arbor Edge Defense DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/redwood-software-cerberus-by-redwood/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_full-click
Cerberus by Redwood | LinkedIn Skip to main content LinkedIn Redwood Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Cerberus by Redwood Managed File Transfer (MFT) Software by Redwood Software See who's skilled in this Add as skill Download now Report this product About Secure file transfer just got easier with Cerberus FTP by Redwood. Since 2001, Cerberus has been simplifying SFTP for small, midsize and enterprise organizations. Known for its ease of use, robust security features and award-winning support, Cerberus is the preferred Windows SFTP solution for IT professionals. Security and compliance features help you meet requirements like HIPAA, GPDR and more, plus regular penetration testing and third-party auditing identifies any potential product weaknesses for remediation. Start sharing your files securely and reliably with Cerberus FTP today. Media Products media viewer No more previous content Cerberus FTP Server v13 IP Manager Cerberus FTP Server v13 Active Directory Integration Cerberus FTP Server v13 Summary Page No more next content Featured customers of Cerberus by Redwood Microsoft Software Development 27,296,747 followers Duke Health Technology Solutions IT Services and IT Consulting 3,253 followers Apple Computers and Electronics Manufacturing 18,041,896 followers SpaceX Aviation and Aerospace Component Manufacturing 3,485,414 followers Intel Corporation Semiconductor Manufacturing 4,100,344 followers COMPanion Corporation Software Development 476 followers NHS Food and Beverage Services 712,120 followers Show more Show less Similar products Axway Managed File Transfer Axway Managed File Transfer Managed File Transfer (MFT) Software Progress MOVEit Progress MOVEit Managed File Transfer (MFT) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software JSCAPE by Redwood JSCAPE by Redwood Managed File Transfer (MFT) Software MLADU MLADU Managed File Transfer (MFT) Software dDataBox dDataBox Managed File Transfer (MFT) Software Sign in to see more Show more Show less Redwood Software products ActiveBatch by Redwood ActiveBatch by Redwood Workload Automation Software Finance Automation by Redwood Finance Automation by Redwood Financial Close Software JSCAPE by Redwood JSCAPE by Redwood Managed File Transfer (MFT) Software RunMyJobs by Redwood RunMyJobs by Redwood Workload Automation Software Tidal by Redwood Tidal by Redwood Workload Automation Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/pt-br/v2/Come%c3%a7ando-O-B%c3%a1sico-do-Git
Git - O Básico do Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Começando 1.1 Sobre Controle de Versão 1.2 Uma Breve História do Git 1.3 O Básico do Git 1.4 A Linha de Comando 1.5 Instalando o Git 1.6 Configuração Inicial do Git 1.7 Pedindo Ajuda 1.8 Sumário 2. Fundamentos de Git 2.1 Obtendo um Repositório Git 2.2 Gravando Alterações em Seu Repositório 2.3 Vendo o histórico de Commits 2.4 Desfazendo coisas 2.5 Trabalhando de Forma Remota 2.6 Criando Tags 2.7 Apelidos Git 2.8 Sumário 3. Branches no Git 3.1 Branches em poucas palavras 3.2 O básico de Ramificação (Branch) e Mesclagem (Merge) 3.3 Gestão de Branches 3.4 Fluxo de Branches 3.5 Branches remotos 3.6 Rebase 3.7 Sumário 4. Git no servidor 4.1 Os Protocolos 4.2 Getting Git on a Server 4.3 Gerando Sua Chave Pública SSH 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Sumário 5. Distributed Git 5.1 Fluxos de Trabalho Distribuídos 5.2 Contribuindo com um Projeto 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Configurando uma conta 6.2 Contribuindo em um projeto 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Funcionamento Interno do Git 10.1 Encanamento e Porcelana 10.2 Objetos do Git 10.3 Referências do Git 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Variáveis de ambiente 10.9 Sumário A1. Appendix A: Git em Outros Ambientes A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Resumo A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 1.3 Começando - O Básico do Git O Básico do Git Então, em poucas palavras, o que é o Git ? Esta é uma parte que é importante aprender, porque se você entender o que o Git é e os fundamentos de como ele funciona, em seguida, provavelmente usar efetivamente o Git será muito mais fácil para você. Enquanto você estiver aprendendo sobre o Git, tente esquecer das coisas que você pode saber sobre outros VCSs, como Subversion e Perforce; isso vai ajudá-lo a evitar a confusão sutil ao usar a ferramenta. O Git armazena e vê informações de forma muito diferente do que esses outros sistemas, mesmo que a interface do usuário seja bem semelhante, e entender essas diferenças o ajudará a não ficar confuso. (Perforce ) Imagens, Não Diferenças A principal diferença entre o Git e qualquer outro VCS (Subversion e similares) é a maneira como o Git trata seus dados. Conceitualmente, a maioria dos outros sistemas armazenam informação como uma lista de mudanças nos arquivos. Estes sistemas (CVS, Subversion, Perforce, Bazaar, e assim por diante) tratam a informação como um conjunto de arquivos e as mudanças feitas em cada arquivo ao longo do tempo. Figure 4. Armazenando dados como alterações em uma versão básica de cada arquivo. O Git não trata nem armazena seus dados desta forma. Em vez disso, o Git trata seus dados mais como um conjunto de imagens de um sistema de arquivos em miniatura. Toda vez que você fizer um commit , ou salvar o estado de seu projeto no Git, ele basicamente tira uma foto de todos os seus arquivos e armazena uma referência para esse conjunto de arquivos. Para ser eficiente, se os arquivos não foram alterados, o Git não armazena o arquivo novamente, apenas um link para o arquivo idêntico anterior já armazenado. O Git trata seus dados mais como um fluxo do estado dos arquivos . Figure 5. Armazenando dados como um estado do conjunto de arquivos do projeto ao longo do tempo. Esta é uma diferença importante entre o Git e quase todos os outros VCSs. Isto faz o Git reconsiderar quase todos os aspectos de controle de versão que a maioria dos outros sistemas copiaram da geração anterior. Isso faz com que o Git seja mais como um mini sistema de arquivos com algumas ferramentas incrivelmente poderosas, ao invés de simplesmente um VCS. Vamos explorar alguns dos benefícios que você ganha ao tratar seus dados desta forma quando cobrirmos ramificações no Git [ch03-git-branching] . Quase Todas as Operações são Locais A maioria das operações no Git só precisa de arquivos e recursos locais para operar - geralmente nenhuma informação é necessária de outro computador da rede. Se você estiver acostumado com um CVCS onde a maioria das operações têm aquela demora causada pela latência da rede, este aspecto do Git vai fazer você pensar que os deuses da velocidade abençoaram o Git com poderes extraterrestres. Como você tem toda a história do projeto ali mesmo em seu disco local, a maioria das operações parecem quase instantâneas. Por exemplo, para pesquisar o histórico do projeto, o Git não precisa sair para o servidor para obter a história e exibi-lo para você - ele simplesmente lê diretamente do seu banco de dados local. Isto significa que você vê o histórico do projeto quase que instantaneamente. Se você quiser ver as alterações introduzidas entre a versão atual de um arquivo e o arquivo de um mês atrás, o Git pode procurar o arquivo de um mês atrás e fazer um cálculo de diferença local, em vez de ter que quer pedir a um servidor remoto para fazê-lo ou puxar uma versão mais antiga do arquivo do servidor remoto para fazê-lo localmente. Isto também significa que há muito pouco que você não pode fazer se você estiver desconectado ou sem VPN. Se você estiver em um avião ou um trem e quiser trabalhar um pouco, você pode fazer commits alegremente até conseguir conexão de rede e enviar os arquivos. Se você chegar em casa e não conseguir conectar ao VPN, você ainda poderá trabalhar. Em muitos outros sistemas, isso é impossível ou doloroso. No Perforce, por exemplo, você não pode fazer quase nada se você não estiver conectado ao servidor; e no Subversion e CVS, você pode editar os arquivos, mas não poderá enviar commits das alterações ao seu banco de dados (porque você não está conectado ao seu banco de dados). Isso pode não parecer muito, mas você poderá se surpreender com a grande diferença que isso pode fazer. Git Tem Integridade Tudo no Git passa por uma soma de verificações ( checksum ) antes de ser armazenado e é referenciado por esse checksum . Isto significa que é impossível mudar o conteúdo de qualquer arquivo ou pasta sem que Git saiba. Esta funcionalidade está incorporada no Git nos níveis mais baixos e é parte integrante de sua filosofia. Você não perderá informação durante a transferência e não receberá um arquivo corrompido sem que o Git seja capaz de detectar. O mecanismo que o Git utiliza para esta soma de verificação é chamado um hash SHA-1. Esta é uma sequência de 40 caracteres composta de caracteres hexadecimais (0-9 e-f) e é calculada com base no conteúdo de uma estrutura de arquivo ou diretório no Git. Um hash SHA-1 é algo como o seguinte: 24b9da6552252987aa493b52f8696cd6d3b00373 Você vai ver esses valores de hash em todo o lugar no Git porque ele os usa com frequência. Na verdade, o Git armazena tudo em seu banco de dados não pelo nome do arquivo, mas pelo valor de hash do seu conteúdo. O Git Geralmente Somente Adiciona Dados Quando você faz algo no Git, quase sempre dados são adicionados no banco de dados do Git - e não removidos. É difícil fazer algo no sistema que não seja reversível ou fazê-lo apagar dados de forma alguma. Como em qualquer VCS, você pode perder alterações que ainda não tenham sido adicionadas em um commit ; mas depois de fazer o commit no Git do estado atual das alterações, é muito difícil que haja alguma perda, especialmente se você enviar regularmente o seu banco de dados para outro repositório. Isso faz com que o uso do Git seja somente alegria, porque sabemos que podemos experimentar sem o perigo de estragar algo. Para um olhar mais aprofundado de como o Git armazena seus dados e como você pode recuperar dados que parecem perdidos, consulte Desfazendo coisas . Os Três Estados Agora, preste atenção. Esta é a principal coisa a lembrar sobre Git se você quiser que o resto do seu processo de aprendizagem ocorra sem problemas. O Git tem três estados principais que seus arquivos podem estar: committed , modificado ( modified ) e preparado ( staged ). Committed significa que os dados estão armazenados de forma segura em seu banco de dados local. Modificado significa que você alterou o arquivo, mas ainda não fez o commit no seu banco de dados. Preparado significa que você marcou a versão atual de um arquivo modificado para fazer parte de seu próximo commit . Isso nos leva a três seções principais de um projeto Git: o diretório Git, o diretório de trabalho e área de preparo. Figure 6. Diretório de trabalho, área de preparo, e o diretório Git. O diretório Git é onde o Git armazena os metadados e o banco de dados de objetos de seu projeto. Esta é a parte mais importante do Git, e é o que é copiado quando você clona um repositório de outro computador. O diretório de trabalho é uma simples cópia de uma versão do projeto. Esses arquivos são pegos do banco de dados compactado no diretório Git e colocados no disco para você usar ou modificar. A área de preparo é um arquivo, geralmente contido em seu diretório Git, que armazena informações sobre o que vai entrar em seu próximo commit . É por vezes referido como o “índice”, mas também é comum referir-se a ele como área de preparo ( staging area ). O fluxo de trabalho básico Git é algo assim: Você modifica arquivos no seu diretório de trabalho. Você prepara os arquivos, adicionando imagens deles à sua área de preparo. Você faz commit , o que leva os arquivos como eles estão na área de preparo e armazena essa imagens de forma permanente para o diretório do Git. Se uma versão específica de um arquivo está no diretório Git, é considerado commited . Se for modificado, mas foi adicionado à área de preparo, é considerado preparado. E se ele for alterado depois de ter sido carregado, mas não foi preparado, ele é considerado modificado. Em [ch02-git-basics] , você vai aprender mais sobre esses estados e como você pode tirar proveito deles ou pular a parte de preparação inteiramente. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/be/v2/Git-Basics-Tagging
Git - Tagging About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Першыя крокі 1.1 About Version Control 1.2 A Short History of Git 1.3 What is Git? 1.4 The Command Line 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Getting Help 1.8 Падсумаваньне 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Дадатак A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Summary A2. Дадатак B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Дадатак C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.6 Git Basics - Tagging Tagging Like most VCSs, Git has the ability to tag specific points in a repository’s history as being important. Typically, people use this functionality to mark release points ( v1.0 , v2.0 and so on). In this section, you’ll learn how to list existing tags, how to create and delete tags, and what the different types of tags are. Listing Your Tags Listing the existing tags in Git is straightforward. Just type git tag (with optional -l or --list ): $ git tag v1.0 v2.0 This command lists the tags in alphabetical order; the order in which they are displayed has no real importance. You can also search for tags that match a particular pattern. The Git source repo, for instance, contains more than 500 tags. If you’re interested only in looking at the 1.8.5 series, you can run this: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Заўвага Listing tag wildcards requires -l or --list option If you want just the entire list of tags, running the command git tag implicitly assumes you want a listing and provides one; the use of -l or --list in this case is optional. If, however, you’re supplying a wildcard pattern to match tag names, the use of -l or --list is mandatory. Creating Tags Git supports two types of tags: lightweight and annotated . A lightweight tag is very much like a branch that doesn’t change — it’s just a pointer to a specific commit. Annotated tags, however, are stored as full objects in the Git database. They’re checksummed; contain the tagger name, email, and date; have a tagging message; and can be signed and verified with GNU Privacy Guard (GPG). It’s generally recommended that you create annotated tags so you can have all this information; but if you want a temporary tag or for some reason don’t want to keep the other information, lightweight tags are available too. Annotated Tags Creating an annotated tag in Git is simple. The easiest way is to specify -a when you run the tag command: $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 The -m specifies a tagging message, which is stored with the tag. If you don’t specify a message for an annotated tag, Git launches your editor so you can type it in. You can see the tag data along with the commit that was tagged by using the git show command: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number That shows the tagger information, the date the commit was tagged, and the annotation message before showing the commit information. Lightweight Tags Another way to tag commits is with a lightweight tag. This is basically the commit checksum stored in a file — no other information is kept. To create a lightweight tag, don’t supply any of the -a , -s , or -m options, just provide a tag name: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 This time, if you run git show on the tag, you don’t see the extra tag information. The command just shows the commit: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Tagging Later You can also tag commits after you’ve moved past them. Suppose your commit history looks like this: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme Now, suppose you forgot to tag the project at v1.2, which was at the “Update rakefile” commit. You can add it after the fact. To tag that commit, you specify the commit checksum (or part of it) at the end of the command: $ git tag -a v1.2 9fceb02 You can see that you’ve tagged the commit: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... Sharing Tags By default, the git push command doesn’t transfer tags to remote servers. You will have to explicitly push tags to a shared server after you have created them. This process is just like sharing remote branches — you can run git push origin <tagname> . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 If you have a lot of tags that you want to push up at once, you can also use the --tags option to the git push command. This will transfer all of your tags to the remote server that are not already there. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Now, when someone else clones or pulls from your repository, they will get all your tags as well. Заўвага git push pushes both types of tags git push <remote> --tags will push both lightweight and annotated tags. There is currently no option to push only lightweight tags, but if you use git push <remote> --follow-tags only annotated tags will be pushed to the remote. Deleting Tags To delete a tag on your local repository, you can use git tag -d <tagname> . For example, we could remove our lightweight tag above as follows: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Note that this does not remove the tag from any remote servers. There are two common variations for deleting a tag from a remote server. The first variation is git push <remote> :refs/tags/<tagname> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw The way to interpret the above is to read it as the null value before the colon is being pushed to the remote tag name, effectively deleting it. The second (and more intuitive) way to delete a remote tag is with: $ git push origin --delete <tagname> Checking out Tags If you want to view the versions of files a tag is pointing to, you can do a git checkout of that tag, although this puts your repository in “detached HEAD” state, which has some ill side effects: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image In “detached HEAD” state, if you make changes and then create a commit, the tag will stay the same, but your new commit won’t belong to any branch and will be unreachable, except by the exact commit hash. Thus, if you need to make changes — say you’re fixing a bug on an older version, for instance — you will generally want to create a branch: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' If you do this and make a commit, your version2 branch will be slightly different than your v2.0.0 tag since it will move forward with your new changes, so do be careful. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/dataport-a%C3%B6r-dataport-consulting/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_full-click
Dataport Consulting | LinkedIn Skip to main content LinkedIn Dataport AöR in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Dataport Consulting Strategic Planning Software by Dataport AöR See who's skilled in this Add as skill Contact us Report this product About Der öffentlichen Verwaltung eröffnen sich z. B. durch das OZG, KI sowie eine notwendige Ende-zu-Ende-Digitalisierung neue Möglichkeiten. Dataport Consulting, Ihr Digitalisierungs- und Innovationspartner, unterstützt Sie dabei, passende Lösungen für Sie zu finden und Veränderungen nachhaltig in Ihrer Organisation zu verankern. Lassen Sie sich beraten und kontaktieren Sie uns. ✅ Strategieberatung ✅ Innovationsberatung und Service Design ✅ Prozessmanagement und -automation ✅ Changemanagement und Organisationsentwicklung This product is intended for Head of Section Resort Manager Department Manager Senior Administrative Officer Civil Servant State Secretary Chief Information Security Officer Chief Digital Officer Chief Information Officer Process Manager Media Products media viewer No more previous content Dataport Consulting. Verwaltungswissen. Beratungskompetenz. Trendaffinität. No more next content Similar products Focus Focus Strategic Planning Software Aha! Roadmaps Aha! Roadmaps Strategic Planning Software DIGITAL PRODUCTS DIGITAL PRODUCTS Strategic Planning Software Cascade Cascade Strategic Planning Software WorkBoard WorkBoard Strategic Planning Software Stratsys platform for compliance and efficiency Stratsys platform for compliance and efficiency Strategic Planning Software Sign in to see more Show more Show less Dataport AöR products data[port]ai data[port]ai Data Science & Machine Learning Platforms dDataBox dDataBox Managed File Transfer (MFT) Software dMessenger dMessenger Enterprise Messaging Software dProjectTracking dProjectTracking Project Management Software dWebService dWebService Web Hosting dWorkflow dWorkflow Workflow Management Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://pt-br.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fmessenger.com%252F%26amp%253Bh%3DAT0NYKzIfH0K_3n8zEg2FDOZcaDpnHRvH0R8dh-lPmKwTCHeq_8KnftluFblL_25g--6zmzfczgAJeQwSsawTrtkNWyUyAgtdjIzSYKuD5W1hnR8LO8GCp5d3SZTsqtKSfCJcbVVJU2fV45q
Facebook Facebook Email ou telefone Senha Esqueceu a conta? Criar nova conta Você está bloqueado temporariamente Você está bloqueado temporariamente Parece que você estava usando este recurso de forma indevida. Bloqueamos temporariamente sua capacidade de usar o recurso. Back Português (Brasil) 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Français (France) Deutsch Cadastre-se Entrar Messenger Facebook Lite Vídeo Meta Pay Meta Store Meta Quest Ray-Ban Meta Meta AI Mais conteúdo da Meta AI Instagram Threads Central de Informações de Votação Política de Privacidade Central de Privacidade Sobre Criar anúncio Criar Página Desenvolvedores Carreiras Cookies Escolhas para anúncios Termos Ajuda Upload de contatos e não usuários Configurações Registro de atividades Meta © 2026
2026-01-13T09:29:19
http://git-scm.com/book/de/v2/Erste-Schritte-Was-ist-Versionsverwaltung%3F
Git - Was ist Versionsverwaltung? About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Erste Schritte 1.1 Was ist Versionsverwaltung? 1.2 Kurzer Überblick über die Historie von Git 1.3 Was ist Git? 1.4 Die Kommandozeile 1.5 Git installieren 1.6 Git Basis-Konfiguration 1.7 Hilfe finden 1.8 Zusammenfassung 2. Git Grundlagen 2.1 Ein Git-Repository anlegen 2.2 Änderungen nachverfolgen und im Repository speichern 2.3 Anzeigen der Commit-Historie 2.4 Ungewollte Änderungen rückgängig machen 2.5 Mit Remotes arbeiten 2.6 Taggen 2.7 Git Aliases 2.8 Zusammenfassung 3. Git Branching 3.1 Branches auf einen Blick 3.2 Einfaches Branching und Merging 3.3 Branch-Management 3.4 Branching-Workflows 3.5 Remote-Branches 3.6 Rebasing 3.7 Zusammenfassung 4. Git auf dem Server 4.1 Die Protokolle 4.2 Git auf einem Server einrichten 4.3 Erstellung eines SSH-Public-Keys 4.4 Einrichten des Servers 4.5 Git-Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Von Drittanbietern gehostete Optionen 4.10 Zusammenfassung 5. Verteiltes Git 5.1 Verteilter Arbeitsablauf 5.2 An einem Projekt mitwirken 5.3 Ein Projekt verwalten 5.4 Zusammenfassung 6. GitHub 6.1 Einrichten und Konfigurieren eines Kontos 6.2 Mitwirken an einem Projekt 6.3 Ein Projekt betreuen 6.4 Verwalten einer Organisation 6.5 Skripte mit GitHub 6.6 Zusammenfassung 7. Git Tools 7.1 Revisions-Auswahl 7.2 Interaktives Stagen 7.3 Stashen und Bereinigen 7.4 Deine Arbeit signieren 7.5 Suchen 7.6 Den Verlauf umschreiben 7.7 Reset entzaubert 7.8 Fortgeschrittenes Merging 7.9 Rerere 7.10 Debuggen mit Git 7.11 Submodule 7.12 Bundling 7.13 Replace (Ersetzen) 7.14 Anmeldeinformationen speichern 7.15 Zusammenfassung 8. Git einrichten 8.1 Git Konfiguration 8.2 Git-Attribute 8.3 Git Hooks 8.4 Beispiel für Git-forcierte Regeln 8.5 Zusammenfassung 9. Git und andere VCS-Systeme 9.1 Git als Client 9.2 Migration zu Git 9.3 Zusammenfassung 10. Git Interna 10.1 Basisbefehle und Standardbefehle (Plumbing and Porcelain) 10.2 Git Objekte 10.3 Git Referenzen 10.4 Packdateien (engl. Packfiles) 10.5 Die Referenzspezifikation (engl. Refspec) 10.6 Transfer Protokolle 10.7 Wartung und Datenwiederherstellung 10.8 Umgebungsvariablen 10.9 Zusammenfassung A1. Anhang A: Git in anderen Umgebungen A1.1 Grafische Schnittstellen A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Zusammenfassung A2. Anhang B: Git in deine Anwendungen einbetten A2.1 Die Git-Kommandozeile A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Anhang C: Git Kommandos A3.1 Setup und Konfiguration A3.2 Projekte importieren und erstellen A3.3 Einfache Snapshot-Funktionen A3.4 Branching und Merging A3.5 Projekte gemeinsam nutzen und aktualisieren A3.6 Kontrollieren und Vergleichen A3.7 Debugging A3.8 Patchen bzw. Fehlerkorrektur A3.9 E-mails A3.10 Externe Systeme A3.11 Administration A3.12 Basisbefehle 2nd Edition 1.1 Erste Schritte - Was ist Versionsverwaltung? In diesem Kapitel geht es um den Einstieg in Git. Wir erklären zunächst einige Hintergrundinformationen zu Versionskontrolltools, gehen dann dazu über, wie du Git auf deinem System zum Laufen bringst und wie du es schließlich so einrichtest, dass du damit arbeiten kannst. Am Ende dieses Kapitels solltest du verstehen, wozu Git gut ist und weshalb man es verwenden sollte. Du solltest dann in der Lage sein, mit deiner Arbeit in Git loslegen zu können. Was ist Versionsverwaltung? Was ist „Versionsverwaltung“, und warum solltest du dich dafür interessieren? Versionsverwaltung ist ein System, welches die Änderungen an einer oder einer Reihe von Dateien über die Zeit hinweg protokolliert, sodass man später auf eine bestimmte Version zurückgreifen kann. Die Dateien, die in den Beispielen in diesem Buch unter Versionsverwaltung gestellt werden, enthalten Quelltext von Software. Tatsächlich kann in der Praxis nahezu jede Art von Datei per Versionsverwaltung nachverfolgt werden. Wenn du Grafik- oder Webdesigner bist und jede Version eines Bildes oder Layouts behalten möchtest (was Du mit Sicherheit möchtest), ist die Verwendung eines Versionskontrollsystems (VCS) eine sehr kluge Sache. Damit kannst du ausgewählte Dateien auf einen früheren Zustand zurücksetzen, das gesamte Projekt auf einen früheren Zustand zurücksetzen, zurückliegende Änderungen vergleichen, sehen wer zuletzt etwas geändert hat, das möglicherweise ein Problem verursacht, wer wann ein Problem verursacht hat und vieles mehr. Die Verwendung eines VCS bedeutet im Allgemeinen auch, dass du problemlos eine Wiederherstellung durchführen könntest, wenn etwas kaputt geht oder Dateien verloren gehen. All diese Vorteile erhält man mit einen nur sehr geringen zusätzlichen Aufwand. Lokale Versionsverwaltung Viele Menschen betreiben Versionsverwaltung, indem sie einfach all ihre Dateien in ein separates Verzeichnis kopieren (die Schlaueren darunter verwenden ein Verzeichnis mit Zeitstempel im Namen). Dieser Ansatz ist sehr verbreitet, weil er so einfach ist, aber auch unglaublich fehleranfällig. Man arbeitet sehr leicht im falschen Verzeichnis, bearbeitet damit die falschen Dateien oder überschreibt Dateien, die man eigentlich nicht überschreiben wollte. Aus diesem Grund, haben Programmierer bereits vor langer Zeit lokale Versionsverwaltungssysteme entwickelt, die alle Änderungen an allen relevanten Dateien in einer Datenbank verwalten. Abbildung 1. Lokale Versionsverwaltung Eines der populäreren Versionsverwaltungssysteme war RCS, welches auch heute noch mit vielen Computern ausgeliefert wird. RCS arbeitet nach dem Prinzip, dass für jede Änderung ein Patch (ein Patch umfasst alle Änderungen an einer oder mehreren Dateien) in einem speziellen Format auf der Festplatte gespeichert wird. Um eine bestimmte Version einer Datei wiederherzustellen, wendet es alle Patches bis zur gewünschten Version an und rekonstruiert damit die Datei in der gewünschten Version. Zentrale Versionsverwaltung Ein weiteres großes Problem, mit dem sich viele Leute konfrontiert sahen, bestand in der Zusammenarbeit mit anderen Entwicklern auf anderen Systemen. Um dieses Problem zu lösen, wurden zentralisierte Versionsverwaltungssysteme entwickelt (engl. Centralized Version Control System, CVCS). Diese Systeme, wozu beispielsweise CVS, Subversion und Perforce gehören, basieren auf einem zentralen Server, der alle versionierte Dateien verwaltet. Die Clients können die Dateien von diesem zentralen Ort abholen und auf ihren PC übertragen. Den Vorgang des Abholens nennt man Auschecken (engl. to check out). Diese Art von System war über viele Jahre hinweg der Standard für Versionsverwaltungssysteme. Abbildung 2. Zentrale Versionsverwaltung Dieser Ansatz hat viele Vorteile, besonders gegenüber lokalen Versionsverwaltungssystemen. Zum Beispiel weiß jeder mehr oder weniger genau darüber Bescheid, was andere an einem Projekt Beteiligte gerade tun. Administratoren haben die Möglichkeit, detailliert festzulegen, wer was tun kann. Und es ist sehr viel einfacher, ein CVCS zu administrieren als lokale Datenbanken auf jedem einzelnen Anwenderrechner zu verwalten. Allerdings hat dieser Aufbau auch einige erhebliche Nachteile. Der offensichtlichste Nachteil ist das Risiko eines Systemausfalls bei Ausfall einer einzelnen Komponente, d.h. wenn der zentrale Server ausfällt. Wenn dieser Server nur für eine Stunde nicht verfügbar ist, dann kann in dieser einen Stunde niemand in irgendeiner Form mit anderen zusammenarbeiten oder Dateien, an denen gerade gearbeitet wird, versioniert abspeichern. Wenn die auf dem zentralen Server eingesetzte Festplatte beschädigt wird und keine Sicherheitskopien erstellt wurden, dann sind all diese Daten unwiederbringlich verloren – die komplette Historie des Projektes, abgesehen natürlich von dem jeweiligen Zustand, den Mitstreiter gerade zufällig auf ihrem Rechner noch vorliegen haben. Lokale Versionskontrollsysteme haben natürlich dasselbe Problem: Wenn man die Historie eines Projektes an einer einzigen, zentralen Stelle verwaltet, riskiert man, sie vollständig zu verlieren, wenn irgendetwas an dieser zentralen Stelle schief läuft. Verteilte Versionsverwaltung Und an dieser Stelle kommen verteilte Versionsverwaltungssysteme (engl. Distributed Version Control System, DVCS) ins Spiel. In einem DVCS (wie z.B. Git, Mercurial, Bazaar oder Darcs) erhalten Anwender nicht einfach nur den jeweils letzten Zustand des Projektes von einem Server: Sie erhalten stattdessen eine vollständige Kopie des Repositories. Auf diese Weise kann, wenn ein Server beschädigt wird, jedes beliebige Repository von jedem beliebigen Anwenderrechner zurückkopiert werden und der Server so wiederhergestellt werden. Jede Kopie, ein sogenannter Klon (engl. clone), ist ein vollständiges Backup der gesamten Projektdaten. Abbildung 3. Verteilte Versionsverwaltung Darüber hinaus können derartige Systeme hervorragend mit verschiedenen externen Repositories, sogenannten Remote-Repositories, umgehen, sodass man mit verschiedenen Gruppen von Leuten simultan auf verschiedene Art und Weise an einem Projekt zusammenarbeiten kann. Damit ist es möglich, verschiedene Arten von Arbeitsabläufen zu erstellen und anzuwenden, welche mit zentralisierten Systemen nicht möglich wären. Dazu gehören zum Beispiel hierarchische Arbeitsabläufe. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://scholar.google.com/citations?view_op=search_authors&hl=ko&oe=ASCII&mauthors=label:ai_security_safety
:root{--gm3-sys-color-on-surface-rgb:31,31,31}@media screen and (prefers-color-scheme:dark){:root{--gm3-sys-color-on-surface-rgb:227,227,227}}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.q4Wquf,.nfoC7c{display:block;height:25vh;position:relative}@media (min-width:600px){.q4Wquf,.nfoC7c{height:150px}}.q4Wquf.Irjbwb{height:auto}@media screen and (prefers-color-scheme:dark){.q4Wquf:not(.GtvzYd){display:none}}.nfoC7c{margin:0;overflow:hidden}.PwpMUe,.lVUmD{display:block;height:100%;margin:0 auto;width:100%}.St9mde{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.St9mde{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.PwpMUe,.lVUmD,.St9mde{-o-object-fit:contain;object-fit:contain}.wsArZ[data-ss-mode="1"] .q4Wquf,.wsArZ[data-ss-mode="1"] .St9mde{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .St9mde{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .q4Wquf,.NQ5OL .St9mde{height:auto;width:100%}.NQ5OL .St9mde{max-width:400px}}.q4Wquf.NWba7e,.q4Wquf.NWba7e .St9mde{height:auto}.q4Wquf.NWba7e .St9mde{height:auto;max-width:312px;width:100%}.q4Wquf.NWba7e.zpCp3 .St9mde{max-width:unset}.q4Wquf.IiQozc .St9mde{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.q4Wquf.Irjbwb .St9mde{height:auto;width:100%}.q4Wquf.EEeaqf .St9mde{max-height:144px;max-width:144px}.SnAaEd{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.SnAaEd{display:none}}.SnAaEd::after,.SnAaEd::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.SnAaEd::after,.SnAaEd::before{width:150px}}.SnAaEd::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.SnAaEd::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .SnAaEd~.St9mde{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .SnAaEd~.St9mde{width:auto}}.RHNWk .St9mde{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .RHNWk .St9mde{width:115px}}.cf660d .St9mde{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.tUhwwc .St9mde{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.Jkvqxd .St9mde{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.onc8Ic .St9mde{-webkit-transform:translate(var( --c-ps-s,24px ),0);-ms-transform:translate(var( --c-ps-s,24px ),0);transform:translate(var( --c-ps-s,24px ),0)}.WA89Yb .St9mde{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .XEN8Yb .St9mde{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .XEN8Yb .St9mde{max-width:115px}}@media (min-width:1240px) and (orientation:landscape),all and (min-width:1600px){.KyPKed .NQ5OL .XEN8Yb .St9mde{max-width:115px}}.IsSr6b .St9mde{max-width:300px}.mmskdd .St9mde{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@-webkit-keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@-webkit-keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@-webkit-keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}:root{--wf-tfs:calc(var(--c-tfs,32)/16*1rem);--wf-tfs-bp2:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp3:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp5:calc(var(--c-tfs,44)/16*1rem);--wf-stfs:calc(var(--c-stfs,16)/16*1rem);--wf-stfs-bp5:calc(var(--c-stfs,16)/16*1rem)}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.Dzz9Db,.GpMPBe{display:block;height:25vh;position:relative}@media (min-width:600px){.Dzz9Db,.GpMPBe{height:150px}}@media screen and (prefers-color-scheme:dark){.Dzz9Db:not(.GtvzYd){display:none}}.Dzz9Db.Irjbwb{height:auto}.GpMPBe{margin:0;overflow:hidden}.UFQPDd,.JNOvdd{display:block;height:100%;margin:0 auto;-o-object-fit:contain;object-fit:contain;width:100%}.f4ZpM{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.f4ZpM{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.wsArZ[data-ss-mode="1"] .Dzz9Db,.wsArZ[data-ss-mode="1"] .f4ZpM{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .f4ZpM{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .Dzz9Db,.NQ5OL .f4ZpM{height:auto;width:100%}.NQ5OL .f4ZpM{max-width:400px}}.Dzz9Db.utFBGf,.Dzz9Db.utFBGf .f4ZpM{height:auto}.Dzz9Db.utFBGf .f4ZpM{height:auto;max-width:312px;width:100%}.Dzz9Db.utFBGf.zpCp3 .f4ZpM{max-width:unset}.Dzz9Db.IiQozc .f4ZpM{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.Dzz9Db.Irjbwb .f4ZpM{height:auto;width:100%}.Dzz9Db.EEeaqf .f4ZpM{max-height:144px;max-width:144px}.nPt1pc{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.nPt1pc{display:none}}.nPt1pc::after,.nPt1pc::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.nPt1pc::after,.nPt1pc::before{width:150px}}.nPt1pc::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.nPt1pc::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .nPt1pc~.f4ZpM{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .nPt1pc~.f4ZpM{width:auto}}.ZS7CGc .f4ZpM{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .ZS7CGc .f4ZpM{width:115px}}.qiRZ5e .f4ZpM{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.vIv7Gf .f4ZpM{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.nvYXVd .f4ZpM{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.uOhnzd .f4ZpM{-webkit-transform:translate(24px,0);-ms-transform:translate(24px,0);transform:translate(24px,0)}.MsYMaf .f4ZpM{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .YIi9qf .f4ZpM{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .YIi9qf .f4ZpM{max-width:115px}}.QG3Xbe .f4ZpM{max-width:300px}.F6gtje .f4ZpM{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@-webkit-keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@-webkit-keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}@keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}.VfPpkd-ksKsZd-XxIAqe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity;position:relative;outline:none;overflow:hidden}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-ksKsZd-XxIAqe::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-ksKsZd-XxIAqe::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded],.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd{overflow:visible}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:0;left:0;width:100%;height:100%}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-ksKsZd-XxIAqe:hover::before,.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe::before,.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):focus::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d)::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):active::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe{font-size:24px;width:48px;height:48px;padding:12px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-Jh9lGc{width:40px;height:40px;margin-top:4px;margin-bottom:4px;margin-right:4px;margin-left:4px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-J1Ukfc-LhBDec{max-height:40px;max-width:40px}.VfPpkd-Bz112c-LgbsSe:disabled{color:rgba(0,0,0,.38);color:var(--mdc-theme-text-disabled-on-light,rgba(0,0,0,.38))}.VfPpkd-Bz112c-LgbsSe svg,.VfPpkd-Bz112c-LgbsSe img{width:24px;height:24px}.VfPpkd-Bz112c-LgbsSe{display:inline-block;position:relative;-webkit-box-sizing:border-box;box-sizing:border-box;border:none;outline:none;background-color:transparent;fill:currentColor;color:inherit;text-decoration:none;cursor:pointer;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;z-index:0;overflow:visible}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-RLmnJb{position:absolute;top:50%;height:48px;left:50%;width:48px;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-J1Ukfc-LhBDec,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-J1Ukfc-LhBDec{display:block}}.VfPpkd-Bz112c-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-Bz112c-LgbsSe[hidden]{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-KVuj8d-Q3DXx{-webkit-box-align:center;-webkit-align-items:center;align-items:center;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center}.VfPpkd-Bz112c-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:100%;width:100%;display:none}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{border-color:CanvasText}}.VfPpkd-Bz112c-kBDsod{display:inline-block}.VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd,.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd{display:inline-block}.VfPpkd-Bz112c-mRLv6{height:100%;left:0;outline:none;position:absolute;top:0;width:100%}.VfPpkd-Bz112c-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;width:100%;height:100%}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-Bz112c-LgbsSe:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe:disabled:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc{height:100%;left:0;pointer-events:none;position:absolute;top:0;width:100%;z-index:-1}.VfPpkd-dgl2Hf-ppHlrf-sM5MNb{display:inline}.VfPpkd-LgbsSe{position:relative;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-align:center;-webkit-align-items:center;align-items:center;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center;-webkit-box-sizing:border-box;box-sizing:border-box;min-width:64px;border:none;outline:none;line-height:inherit;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;-webkit-appearance:none;overflow:visible;vertical-align:middle;background:transparent}.VfPpkd-LgbsSe .VfPpkd-BFbNVe-bF1uUb{width:100%;height:100%;top:0;left:0}.VfPpkd-LgbsSe::-moz-focus-inner{padding:0;border:0}.VfPpkd-LgbsSe:active{outline:none}.VfPpkd-LgbsSe:hover{cursor:pointer}.VfPpkd-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-LgbsSe[hidden]{display:none}.VfPpkd-LgbsSe .VfPpkd-kBDsod{margin-left:0;margin-right:8px;display:inline-block;position:relative;vertical-align:top}[dir=rtl] .VfPpkd-LgbsSe .VfPpkd-kBDsod,.VfPpkd-LgbsSe .VfPpkd-kBDsod[dir=rtl]{margin-left:8px;margin-right:0}.VfPpkd-LgbsSe .VfPpkd-UdE5de-uDEFge{font-size:0;position:absolute;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);top:50%;left:50%;line-height:normal}.VfPpkd-LgbsSe .VfPpkd-vQzf8d{position:relative}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px);display:none}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-color:CanvasText}}@media screen and (forced-colors:active){.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-J1Ukfc-LhBDec,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-J1Ukfc-LhBDec{display:block}}.VfPpkd-LgbsSe .VfPpkd-RLmnJb{position:absolute;top:50%;height:48px;left:0;right:0;-webkit-transform:translateY(-50%);-ms-transform:translateY(-50%);transform:translateY(-50%)}.VfPpkd-vQzf8d+.VfPpkd-kBDsod{margin-left:8px;margin-right:0}[dir=rtl] .VfPpkd-vQzf8d+.VfPpkd-kBDsod,.VfPpkd-vQzf8d+.VfPpkd-kBDsod[dir=rtl]{margin-left:0;margin-right:8px}svg.VfPpkd-kBDsod{fill:currentColor}.VfPpkd-LgbsSe-OWXEXe-dgl2Hf{margin-top:6px;margin-bottom:6px}.VfPpkd-LgbsSe{-moz-osx-font-smoothing:grayscale;-webkit-font-smoothing:antialiased;text-decoration:none}.VfPpkd-LgbsSe{padding:0 8px 0 8px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-INsAgc{border-style:solid;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-style:solid;border-color:transparent}.VfPpkd-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{z-index:0}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Jh9lGc{position:absolute;-webkit-box-sizing:content-box;box-sizing:content-box;overflow:hidden;z-index:0;top:0;left:0;bottom:0;right:0}.VfPpkd-LgbsSe{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-text-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-filled-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;-webkit-box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12);box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-protected-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled):active{-webkit-box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12);box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{-webkit-box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12);box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;padding:0 15px 0 15px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-outlined-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 11px 0 15px}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 15px 0 11px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{top:-1px;left:-1px;bottom:-1px;right:-1px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-RLmnJb{left:-1px;width:calc(100% + 2px)}.nCP5yc{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.nCP5yc .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.nCP5yc:not(:disabled){background-color:rgb(26,115,232);background-color:var(--gm-fillbutton-container-color,rgb(26,115,232))}.nCP5yc:not(:disabled){color:#fff;color:var(--gm-fillbutton-ink-color,#fff)}.nCP5yc:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-fillbutton-disabled-container-color,rgba(60,64,67,.12))}.nCP5yc:disabled{color:rgba(60,64,67,.38);color:var(--gm-fillbutton-disabled-ink-color,rgba(60,64,67,.38))}.nCP5yc .VfPpkd-Jh9lGc::before,.nCP5yc .VfPpkd-Jh9lGc::after{background-color:rgb(32,33,36);background-color:var(--gm-fillbutton-state-color,rgb(32,33,36))}.nCP5yc:hover .VfPpkd-Jh9lGc::before,.nCP5yc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.16;opacity:var(--mdc-ripple-hover-opacity,.16)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24;opacity:var(--mdc-ripple-focus-opacity,.24)}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.2;opacity:var(--mdc-ripple-press-opacity,.2)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.2)}.nCP5yc .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:#fff}@media (-ms-high-contrast:active),screen and (forced-colors:active){.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.nCP5yc:hover{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:hover .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:active{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:active .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:disabled{-webkit-box-shadow:none;box-shadow:none}.nCP5yc:disabled:hover .VfPpkd-Jh9lGc::before,.nCP5yc:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.Rj2Mlf .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.Rj2Mlf:not(:disabled){color:rgb(26,115,232);color:var(--gm-hairlinebutton-ink-color,rgb(26,115,232))}.Rj2Mlf:not(:disabled){border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled):hover{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled).VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.Rj2Mlf:not(:disabled):not(.VfPpkd-ksKsZd-mWPk3d):focus{border-color:rgb(23,78,166);border-color:var(--gm-hairlinebutton-outline-color--stateful,rgb(23,78,166))}.Rj2Mlf:not(:disabled):active,.Rj2Mlf:not(:disabled):focus:active{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:disabled{color:rgba(60,64,67,.38);color:var(--gm-hairlinebutton-disabled-ink-color,rgba(60,64,67,.38))}.Rj2Mlf:disabled{border-color:rgba(60,64,67,.12);border-color:var(--gm-hairlinebutton-disabled-outline-color,rgba(60,64,67,.12))}.Rj2Mlf:hover:not(:disabled),.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.Rj2Mlf:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-hairlinebutton-ink-color--stateful,rgb(23,78,166))}.Rj2Mlf .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.Rj2Mlf .VfPpkd-Jh9lGc::before,.Rj2Mlf .VfPpkd-Jh9lGc::after{background-color:rgb(26,115,232);background-color:var(--gm-hairlinebutton-state-color,rgb(26,115,232))}.Rj2Mlf:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.Rj2Mlf:disabled:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.b9hyVd{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);border-width:0;-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15))}.b9hyVd .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.b9hyVd:not(:disabled){background-color:#fff;background-color:var(--gm-protectedbutton-container-color,#fff)}.b9hyVd:not(:disabled){color:rgb(26,115,232);color:var(--gm-protectedbutton-ink-color,rgb(26,115,232))}.b9hyVd:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-protectedbutton-disabled-container-color,rgba(60,64,67,.12))}.b9hyVd:disabled{color:rgba(60,64,67,.38);color:var(--gm-protectedbutton-disabled-ink-color,rgba(60,64,67,.38))}.b9hyVd:hover:not(:disabled),.b9hyVd.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.b9hyVd:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.b9hyVd:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-protectedbutton-ink-color--stateful,rgb(23,78,166))}.b9hyVd .VfPpkd-BFbNVe-bF1uUb{opacity:0}.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-B
2026-01-13T09:29:19
https://www.linkedin.com/products/cloudways-autonomous/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
Cloudways Autonomous | LinkedIn Skip to main content LinkedIn Cloudways in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Cloudways Autonomous Web Hosting by Cloudways See who's skilled in this Add as skill Learn more Report this product About Experience fully-managed WordPress hosting that auto-scales with your growing business. Launch your app in under a minute on a self scaling powerful infrastructure that is pre-configured to scale up & down acccording to the traffic demand. For the tech-savvy, this infrastructure is built on Kubernetes with integrated Cloudflare Enterprise & Object Cache Pro's to help ensure to-notch performance & security Enjoy free 24/7/365 technical support, so you're never left alone when you need assitance. Cloudways Autonomous makes high-performance WordPress hosting simple and accessible, combining advanced infrastructure with user-friendly management. Join the 100,000+ SMBs who scale with Cloudways! Media Products media viewer No more previous content Stop Worrying About Traffic Spikes, Servers, or Scaling. Cloudways Autonomous: Fully managed WordPress hosting that auto scales to meet the needs of your growing business. Autonomous adapts to your business in real-time Thanks to Autonomous, your applications remain online and accessible, regardless of traffic spikes. No more next content Featured customers of Cloudways Autonomous ClickMinded Marketing Services 1,681 followers Cootehill Chamber - representing businesses for over 60 years in the North-East Cavan. Non-profit Organizations 202 followers Odysséa Organisation Spectator Sports 4,612 followers Magnitude Surveys Environmental Services 1,024 followers INKAS® Group Manufacturing 10,731 followers Minny Grown Manufacturing 1,618 followers Similar products IPv4 Address Leasing Service IPv4 Address Leasing Service Web Hosting Pantheon Platform Pantheon Platform Web Hosting Elastic Metal Elastic Metal Web Hosting Web Hosting Web Hosting Web Hosting WooCommerce Hosting WooCommerce Hosting Web Hosting Sites Sites Web Hosting Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/id/v2/Git-Basics-Tagging
Git - Tagging About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Memulai 1.1 Tentang Version Control 1.2 Sejarah Singkat Git 1.3 Dasar-dasar Git 1.4 Command Line 1.5 Memasang Git 1.6 Pengaturan Awal Git 1.7 Mendapatkan Bantuan 1.8 Kesimpulan 2. Git Basics 2.1 Mendapatkan Repository Git 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Alias Git 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git di Server 4.1 Protokol 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Ringkasan 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Pengaturan dan Konfigurasi Akun 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Mengelola Organization 6.5 Scripting GitHub 6.6 Ringkasan 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Kostumisasi Git 8.1 Konfigurasi Git 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Ringkasan 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.6 Git Basics - Tagging Tagging Like most VCSs, Git has the ability to tag specific points in history as being important. Typically people use this functionality to mark release points (v1.0, and so on). In this section, you’ll learn how to list the available tags, how to create new tags, and what the different types of tags are. Listing Your Tags Listing the available tags in Git is straightforward. Just type git tag : $ git tag v0.1 v1.3 This command lists the tags in alphabetical order; the order in which they appear has no real importance. You can also search for tags with a particular pattern. The Git source repo, for instance, contains more than 500 tags. If you’re only interested in looking at the 1.8.5 series, you can run this: $ git tag -l 'v1.8.5*' v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Creating Tags Git uses two main types of tags: lightweight and annotated. A lightweight tag is very much like a branch that doesn’t change – it’s just a pointer to a specific commit. Annotated tags, however, are stored as full objects in the Git database. They’re checksummed; contain the tagger name, e-mail, and date; have a tagging message; and can be signed and verified with GNU Privacy Guard (GPG). It’s generally recommended that you create annotated tags so you can have all this information; but if you want a temporary tag or for some reason don’t want to keep the other information, lightweight tags are available too. Annotated Tags Creating an annotated tag in Git is simple. The easiest way is to specify -a when you run the tag command: $ git tag -a v1.4 -m 'my version 1.4' $ git tag v0.1 v1.3 v1.4 The -m specifies a tagging message, which is stored with the tag. If you don’t specify a message for an annotated tag, Git launches your editor so you can type it in. You can see the tag data along with the commit that was tagged by using the git show command: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number That shows the tagger information, the date the commit was tagged, and the annotation message before showing the commit information. Lightweight Tags Another way to tag commits is with a lightweight tag. This is basically the commit checksum stored in a file – no other information is kept. To create a lightweight tag, don’t supply the -a , -s , or -m option: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 This time, if you run git show on the tag, you don’t see the extra tag information. The command just shows the commit: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Tagging Later You can also tag commits after you’ve moved past them. Suppose your commit history looks like this: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 beginning write support 0d52aaab4479697da7686c15f77a3d64d9165190 one more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc added a commit function 4682c3261057305bdd616e23b64b0857d832627b added a todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a started write support 9fceb02d0ae598e95dc970b74767f19372d61af8 updated rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a updated readme Now, suppose you forgot to tag the project at v1.2, which was at the “updated rakefile” commit. You can add it after the fact. To tag that commit, you specify the commit checksum (or part of it) at the end of the command: $ git tag -a v1.2 9fceb02 You can see that you’ve tagged the commit: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... Sharing Tags By default, the git push command doesn’t transfer tags to remote servers. You will have to explicitly push tags to a shared server after you have created them. This process is just like sharing remote branches – you can run git push origin [tagname] . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 If you have a lot of tags that you want to push up at once, you can also use the --tags option to the git push command. This will transfer all of your tags to the remote server that are not already there. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Now, when someone else clones or pulls from your repository, they will get all your tags as well. Checking out Tags You can’t really check out a tag in Git, since they can’t be moved around. If you want to put a version of your repository in your working directory that looks like a specific tag, you can create a new branch at a specific tag: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Of course if you do this and do a commit, your version2 branch will be slightly different than your v2.0.0 tag since it will move forward with your new changes, so do be careful. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/ru/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d1%8b-Git-%d0%a1%d0%be%d0%b7%d0%b4%d0%b0%d0%bd%d0%b8%d0%b5-Git-%d1%80%d0%b5%d0%bf%d0%be%d0%b7%d0%b8%d1%82%d0%be%d1%80%d0%b8%d1%8f
Git - Создание Git-репозитория About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Введение 1.1 О системе контроля версий 1.2 Краткая история Git 1.3 Что такое Git? 1.4 Командная строка 1.5 Установка Git 1.6 Первоначальная настройка Git 1.7 Как получить помощь? 1.8 Заключение 2. Основы Git 2.1 Создание Git-репозитория 2.2 Запись изменений в репозиторий 2.3 Просмотр истории коммитов 2.4 Операции отмены 2.5 Работа с удалёнными репозиториями 2.6 Работа с тегами 2.7 Псевдонимы в Git 2.8 Заключение 3. Ветвление в Git 3.1 О ветвлении в двух словах 3.2 Основы ветвления и слияния 3.3 Управление ветками 3.4 Работа с ветками 3.5 Удалённые ветки 3.6 Перебазирование 3.7 Заключение 4. Git на сервере 4.1 Протоколы 4.2 Установка Git на сервер 4.3 Генерация открытого SSH ключа 4.4 Настраиваем сервер 4.5 Git-демон 4.6 Умный HTTP 4.7 GitWeb 4.8 GitLab 4.9 Git-хостинг 4.10 Заключение 5. Распределённый Git 5.1 Распределённый рабочий процесс 5.2 Участие в проекте 5.3 Сопровождение проекта 5.4 Заключение 6. GitHub 6.1 Настройка и конфигурация учётной записи 6.2 Внесение собственного вклада в проекты 6.3 Сопровождение проекта 6.4 Управление организацией 6.5 Создание сценариев GitHub 6.6 Заключение 7. Инструменты Git 7.1 Выбор ревизии 7.2 Интерактивное индексирование 7.3 Припрятывание и очистка 7.4 Подпись 7.5 Поиск 7.6 Перезапись истории 7.7 Раскрытие тайн reset 7.8 Продвинутое слияние 7.9 Rerere 7.10 Обнаружение ошибок с помощью Git 7.11 Подмодули 7.12 Создание пакетов 7.13 Замена 7.14 Хранилище учётных данных 7.15 Заключение 8. Настройка Git 8.1 Конфигурация Git 8.2 Атрибуты Git 8.3 Хуки в Git 8.4 Пример принудительной политики Git 8.5 Заключение 9. Git и другие системы контроля версий 9.1 Git как клиент 9.2 Переход на Git 9.3 Заключение 10. Git изнутри 10.1 Сантехника и Фарфор 10.2 Объекты Git 10.3 Ссылки в Git 10.4 Pack-файлы 10.5 Спецификации ссылок 10.6 Протоколы передачи данных 10.7 Обслуживание репозитория и восстановление данных 10.8 Переменные окружения 10.9 Заключение A1. Приложение A: Git в других окружениях A1.1 Графические интерфейсы A1.2 Git в Visual Studio A1.3 Git в Visual Studio Code A1.4 Git в Eclipse A1.5 Git в IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.6 Git в Sublime Text A1.7 Git в Bash A1.8 Git в Zsh A1.9 Git в PowerShell A1.10 Заключение A2. Приложение B: Встраивание Git в ваши приложения A2.1 Git из командной строки A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Приложение C: Команды Git A3.1 Настройка и конфигурация A3.2 Клонирование и создание репозиториев A3.3 Основные команды A3.4 Ветвление и слияния A3.5 Совместная работа и обновление проектов A3.6 Осмотр и сравнение A3.7 Отладка A3.8 Внесение исправлений A3.9 Работа с помощью электронной почты A3.10 Внешние системы A3.11 Администрирование A3.12 Низкоуровневые команды 2nd Edition 2.1 Основы Git - Создание Git-репозитория Если вы хотите начать работать с Git, прочитав всего одну главу, то эта глава — то, что вам нужно. Здесь рассмотрены все базовые команды, необходимые вам для решения подавляющего большинства задач, возникающих при работе с Git. После прочтения этой главы вы научитесь настраивать и инициализировать репозиторий, начинать и прекращать контроль версий файлов, а также подготавливать и фиксировать изменения. Мы также продемонстрируем вам, как настроить в Git игнорирование отдельных файлов или их групп, как быстро и просто отменить ошибочные изменения, как просмотреть историю вашего проекта и изменения между отдельными коммитами (commit), а также как отправлять (push) и получать (pull) изменения в/из удалённого (remote) репозитория. Создание Git-репозитория Обычно вы получаете репозиторий Git одним из двух способов: Вы можете взять локальный каталог, который в настоящее время не находится под версионным контролем, и превратить его в репозиторий Git, либо Вы можете клонировать существующий репозиторий Git из любого места. В обоих случаях вы получите готовый к работе Git репозиторий на вашем компьютере. Создание репозитория в существующем каталоге Если у вас уже есть проект в каталоге, который не находится под версионным контролем Git, то для начала нужно перейти в него. Если вы не делали этого раньше, то для разных операционных систем это выглядит по-разному: для Linux: $ cd /home/user/my_project для macOS: $ cd /Users/user/my_project для Windows: $ cd C:/Users/user/my_project а затем выполните команду: $ git init Эта команда создаёт в текущем каталоге новый подкаталог с именем .git , содержащий все необходимые файлы репозитория — структуру Git репозитория. На этом этапе ваш проект ещё не находится под версионным контролем. Подробное описание файлов, содержащихся в только что созданном вами каталоге .git , приведено в главе Git изнутри Если вы хотите добавить под версионный контроль существующие файлы (в отличие от пустого каталога), вам стоит добавить их в индекс и осуществить первый коммит изменений. Добиться этого вы сможете запустив команду git add несколько раз, указав индексируемые файлы, а затем выполнив git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' Мы разберём, что делают эти команды чуть позже. Теперь у вас есть Git-репозиторий с отслеживаемыми файлами и начальным коммитом. Клонирование существующего репозитория Для получения копии существующего Git-репозитория, например, проекта, в который вы хотите внести свой вклад, необходимо использовать команду git clone . Если вы знакомы с другими системами контроля версий, такими как Subversion, то заметите, что команда называется «clone», а не «checkout». Это важное различие — вместо того, чтобы просто получить рабочую копию, Git получает копию практически всех данных, которые есть на сервере. При выполнении git clone с сервера забирается (pulled) каждая версия каждого файла из истории проекта. Фактически, если серверный диск выйдет из строя, вы можете использовать любой из клонов на любом из клиентов, для того, чтобы вернуть сервер в то состояние, в котором он находился в момент клонирования (вы можете потерять часть серверных хуков (server-side hooks) и т. п., но все данные, помещённые под версионный контроль, будут сохранены, подробнее об этом смотрите в разделе Установка Git на сервер главы 4). Клонирование репозитория осуществляется командой git clone <url> . Например, если вы хотите клонировать библиотеку libgit2 , вы можете сделать это следующим образом: $ git clone https://github.com/libgit2/libgit2 Эта команда создаёт каталог libgit2 , инициализирует в нём подкаталог .git , скачивает все данные для этого репозитория и извлекает рабочую копию последней версии. Если вы перейдёте в только что созданный каталог libgit2 , то увидите в нём файлы проекта, готовые для работы или использования. Для того, чтобы клонировать репозиторий в каталог с именем, отличающимся от libgit2 , необходимо указать желаемое имя, как параметр командной строки: $ git clone https://github.com/libgit2/libgit2 mylibgit Эта команда делает всё то же самое, что и предыдущая, только результирующий каталог будет назван mylibgit . В Git реализовано несколько транспортных протоколов, которые вы можете использовать. В предыдущем примере использовался протокол https:// , вы также можете встретить git:// или user@server:path/to/repo.git , использующий протокол передачи SSH. В разделе Установка Git на сервер главы 4 мы познакомимся со всеми доступными вариантами конфигурации сервера для обеспечения доступа к вашему Git репозиторию, а также рассмотрим их достоинства и недостатки. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/tr/v2/Git-Temelleri-Bir-Git-Reposu-Olu%c5%9fturma/Kopyalama
Git - Bir Git Reposu Oluşturma/Kopyalama About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlangıç 1.1 Sürüm Denetimi 1.2 Git’in Kısa Tarihçesi 1.3 Git Nedir? 1.4 Komut Satırı 1.5 Git’i Yüklemek 1.6 Git’i İlk Defa Kurmak 1.7 Yardım Almak 1.8 Özet 2. Git Temelleri 2.1 Bir Git Reposu Oluşturma/Kopyalama 2.2 Değişikliklerin Repoya Kaydedilmesi 2.3 Katkı Geçmişini Görüntüleme 2.4 Değişiklikleri Geri Alma 2.5 Uzak Repo ile Çalışmak 2.6 Etiketleme 2.7 Komut Kısayolu (Alias) Ayarlama 2.8 Özet 3. Git Dalları 3.1 Dallar 3.2 Kısaca Dallandırma ve Birleştirme Temelleri 3.3 Dal Yönetimi 3.4 İş Akışı Dallandırması 3.5 Uzak Dallar 3.6 Yeniden Temelleme (rebase) 3.7 Özet 4. Bir Sunucuda Git Kurma 4.1 İletişim Kuralları (Protocols) 4.2 Bir Sunucuda Git Kurma 4.3 SSH Ortak Anahtarınızı Oluşturma 4.4 Sunucu Kurma 4.5 Git Cini (Daemon) 4.6 Akıllı HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Taraf Barındırma (Hosting) Seçenekleri 4.10 Özet 5. Dağıtık Git 5.1 Dağıtık İş Akışları 5.2 Projenin Gelişiminde Rol Almak 5.3 Bir Projeyi Yürütme 5.4 Özet 6. GitHub 6.1 Bir Projeye Katkıda Bulunmak 6.2 Proje Bakımı 6.3 Kurumsal Yönetim 6.4 GitHub’ı otomatikleştirme 6.5 Özet 7. Git Araçları 7.1 Düzeltme Seçimi 7.2 Etkileşimli İzlemleme (Staging) 7.3 Saklama ve Silme 7.4 Çalışmanızı İmzalama 7.5 Arama 7.6 Geçmişi Yeniden Yazma 7.7 Reset Komutunun Gizemleri 7.8 İleri Seviye Birleştirme 7.9 Rerere 7.10 Git’le Hata Ayıklama 7.11 Alt Modüller 7.12 Demetleme (Bundling) 7.13 Git Nesnesini Değiştirme 7.14 Kimlik Bilgisi Depolama 7.15 Özet 8. Git’i Özelleştirmek 8.1 Git Yapılandırması 8.2 Git Nitelikleri 8.3 Git Kancaları (Hooks) 8.4 Bir Örnek: Mecburi Git Politikası 8.5 Özet 9. Git ve Diğer Sistemler 9.1 İstemci Olarak Git 9.2 Git’e Geçiş 9.3 Özet 10. Dahili Git Ögeleri 10.1 Tesisat ve Döşeme (Plumbing ve Porcelain) 10.2 Git Nesneleri 10.3 Git Referansları 10.4 Packfiles 10.5 Refspec 10.6 Transfer Protokolleri 10.7 Bakım ve Veri Kurtarma 10.8 Ortam Değişkenleri 10.9 Özet A1. Ek bölüm A: Diğer Ortamlarda Git A1.1 Görsel Arayüzler A1.2 Visual Studio ile Git A1.3 Visual Studio Code ile Git A1.4 Eclipse ile Git A1.5 Sublime Text ile Git A1.6 Bash ile Git A1.7 Zsh ile Git A1.8 PowerShell ile Git A1.9 Özet A2. Ek bölüm B: Git’i Uygulamalarınıza Gömmek A2.1 Git Komut Satırı A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Ek bölüm C: Git Komutları A3.1 Kurulum ve Yapılandırma Komutları A3.2 Proje Oluşturma Komutları A3.3 Kısaca Poz (Snapshot) Alma A3.4 Dallandırma ve Birleştirme Komutları A3.5 Projeleri Paylaşma ve Güncelleme Komutları A3.6 İnceleme ve Karşılaştırma Komutları A3.7 Hata Ayıklama (Debugging) Komutları A3.8 Yamalama (Patching) A3.9 E-Posta Komutları A3.10 Harici Sistemler A3.11 Yönetim A3.12 Tesisat (Plumbing) Komutları 2nd Edition 2.1 Git Temelleri - Bir Git Reposu Oluşturma/Kopyalama Git’e başlamak için yalnızca bir bölüm okuyabilecek vaktiniz varsa, işte bu aradığınız bölümdür. Bu bölüm, Git’te zamanınızı harcayacağınız şeylerin büyük çoğunluğunu yapmak için ihtiyacınız olan tüm temel komutları kapsar. Bölümün sonunda, bir Git reposunu yapılandırıp başlatabilmeniz, dosyaları izlemeyi başlatıp durdurabilmeniz ve değişikliklerinizi izleme alıp (stage) uzak repoya kaydedebilecek seviyeye geleceksiniz. Ayrıca Git’i belirli dosyaları ve dosya kalıplarını yok sayacak şekilde nasıl ayarlayacağınızı, hataları hızlı ve kolay bir şekilde nasıl geri alacağınızı, projenizin geçmişine nasıl göz atacağınızı, katkılar (commit) arasındaki değişiklikleri nasıl görüntüleyeceğinizi ve uzak repolarla nasıl kod alışverişi yapacağınızı göstereceğiz. Git’e başlamak için yalnızca bir bölüm okuyabilecek vaktiniz varsa, işte bu aradığınız bölümdür. Bu bölüm, Git’te zamanınızı harcayacağınız şeylerin büyük çoğunluğunu yapmak için ihtiyacınız olan tüm temel komutları kapsar. Bölümün sonunda, bir Git reposunu yapılandırıp başlatabilmeniz, dosyaları izlemeyi başlatıp durdurabilmeniz ve değişikliklerinizi izleme alıp (stage) uzak repoya kaydedebilecek seviyeye geleceksiniz. Ayrıca Git’i belirli dosyaları ve dosya kalıplarını yok sayacak şekilde nasıl ayarlayacağınızı, hataları hızlı ve kolay bir şekilde nasıl geri alacağınızı, projenizin geçmişine nasıl göz atacağınızı, katkılar (commit) arasındaki değişiklikleri nasıl görüntüleyeceğinizi ve uzak repolarla nasıl kod alışverişi yapacağınızı göstereceğiz. Bir Git Reposu Oluşturma/Kopyalama Tipik olarak bir Git reposu oluşturmanın 2 yolu vardır: Şu anda sürüm denetimi (version control) altında olmayan bir yerel dizini (dosya yolu) alabilir ve onu bir Git reposuna dönüştürebilirsiniz. Veya Başka bir yerden, var olan bir Git reposunu kopyalayabilirsiniz. Her iki durumda da, yerel makinenizde çalışmaya hazır bir Git reposuna sahip olursunuz. Varolan Bir Dizinde Repo Başlatma Eğer şu anda sürüm denetimi altında olmayan bir proje diziniz varsa ve onu Git ile takip etmeye başlamak istiyorsanız, önce o projenin dizinine gitmeniz gerekmektedir. Eğer bunu daha önce yapmadıysanız, hangi işletim sisteminde çalıştığınıza bağlı olarak proje dizininiz farklı görünebilir. Linux için: $ cd /home/user/my_project macOS için: $ cd /Users/user/my_project Windows için: $ cd /c/user/my_project Proje dizinine girdikten sonra şunu yazın: $ git init Bu, tüm gerekli repo dosyalarını içeren .git adında yeni bir alt dizin oluşturur (Yani bir Git repo temeli). Bu aşamada, projenizde henüz takip edilen bir şey yoktur. (Yeni oluşturduğunuz .git dizininde hangi dosyaların bulunduğuna dair daha fazla bilgi edinmek için Dahili Git Ögeleri bölümüne bakabilirsiniz.) Eğer halihazırda mevcut olan dosyaları sürüm denetimine almak istiyorsanız, boş bir dizin yerine, bu hazır dosyaları takip etmeye başlamalı ve bir initial commit (ilk katkı) yapmalısınız. Bunu, takip etmek istediğiniz dosyaları belirten birkaç git add komutu ve ardından bir git commit komutu ile başarabilirsiniz: $ git add *.c $ git add LICENSE $ git commit -m 'initial project version' Bu komutların ne yaptığını birazdan inceleyeceğiz. Bu aşamada, takip edilen dosyaları içeren bir Git reponuz ve başlangıç 'initial commit’iniz vardır. Mevcut bir Git Reposunu Kopyalama (Klonlama) Mevcut bir Git deposunun (örneğin katkıda bulunmak istediğiniz bir projenin) bir kopyasını almak istiyorsanız git clone komutunu kullanmalısınız. Eğer Subversion veya benzeri diğer VCS sistemlerine aşina iseniz, "checkout" yerine "clone" komutu kullanıldığını fark edeceksiniz. Bu önemli bir ayrımdır! Zira Git, yalnızca çalışan bir kopya almak yerine, sunucunun sahip olduğu neredeyse tüm verilerin tam bir kopyasını alır. git clone komutunu çalıştırdığınızda, projenin geçmişindeki her dosyanın her sürümü varsayılan olarak indirilir. Aslında, sunucu diskiniz bozulursa, sunucuyu kopyalandığı zamanki durumuna geri döndürmek için herhangi bir istemcideki kopyaların neredeyse tümünü kullanabilirsiniz (bazı sunucu tarafı kancalarını vb. kaybedebilirsiniz, ancak sürümlendirilmiş tüm verilere sahip olacaksınız. Daha fazla ayrıntı için: Bir Sunucuda Git Kurma ). Bir repoyu git clone <url> komutu ile kopyalayabilirsiniz. Örneğin, "libgit2" adlı Git’e bağlanabilir kitaplığı kopyalamak istiyorsanız bunu şu şekilde yapabilirsiniz: $ git clone https://github.com/libgit2/libgit2 Bu, içinde bir .git dizini olan libgit2 adında bir dizin oluşturur, bu repoya ait tüm verileri çeker ve en son sürümün çalışan bir kopyasını alır. Yeni oluşturulan libgit2 dizinine giderseniz, orada üzerinde çalışılmaya veya kullanılmaya hazır proje dosyalarını görebilirsiniz. Repoyu libgit2 dışında bir isimle adlandırılan bir dizine kopyalamak istiyorsanız, yeni dizin adını ek bir argüman olarak belirtebilirsiniz: $ git clone https://github.com/libgit2/libgit2 mylibgit Bu komut öncekiyle aynı şeyi yapar, ancak hedef dizine mylibgit adını verir. Git’te kullanabileceğiniz bir dizi farklı aktarım protokolü vardır. Önceki örnekte https:// protokolü kullanılmıştır ancak git:// veya SSH aktarım protokolünü kullanan user@server:path/to/repo.git ifadelerini de görebilirsiniz. Bir Sunucuda Git Kurma bölümünde sunucunun Git reponuza erişmek için ayarlayabileceği mevcut tüm seçenekleri ve her birinin artılarını ve eksilerini göreceksiniz. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/livechat/?trk=products_seo_search
Text | LinkedIn Skip to main content LinkedIn Text in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Text Live Chat Software by Text See who's skilled in this Add as skill Contact us Report this product About Text App is an all-in-one AI support platform that connects live chat, smart ticketing, and intelligent automation. It empowers teams to resolve issues more quickly, operate more efficiently, and deliver customer experiences that truly stand out. Key features include: ✦ Live chat: Connect with customers instantly and resolve issues in real time. ✦ Smart ticketing: Organize, prioritize, and automate support requests efficiently. ✦ AI agent: Handle up to 80% of routine inquiries autonomously. ✦ Automation workflows: Streamline repetitive tasks and boost team productivity. ✦ Analytics & insights: Monitor response times, CSAT, and team performance to optimize results. ✦ Omnichannel support: Manage conversations across chat, email, Messenger, WhatsApp, and more — all in one place. Media Products media viewer No more previous content Text App Context-aware AI engine handles up to 80% of routine questions, suggests real-time replies, and speeds up resolution. Chat Widget Text Intelligence AI Agent Metrics and reports No more next content Featured customers of Text IKEA Retail 3,494,359 followers McDonald's Restaurants 2,378,938 followers Valley Driving School Truck Transportation 383 followers Mercedes-Benz AG Motor Vehicle Manufacturing 2,306,610 followers Ingram Micro IT Services and IT Consulting 584,277 followers Huawei Telecommunications 5,372,667 followers Nikon Appliances, Electrical, and Electronics Manufacturing 102,279 followers Plus500™ Financial Services 32,513 followers PayPal Software Development 1,615,662 followers Henkel Manufacturing 1,697,595 followers Virgin Venture Capital and Private Equity Principals 320,544 followers Comcast Telecommunications 746,229 followers Toyota Motor Corporation Motor Vehicle Manufacturing 2,293,377 followers Unilever Manufacturing 20,625,355 followers LG Electronics Computers and Electronics Manufacturing 1,392,216 followers ING Banking 617,583 followers Atlassian Software Development 2,340,590 followers CBS Entertainment Providers 151,818 followers Adobe Software Development 5,252,484 followers Stryker Medical Equipment Manufacturing 1,707,589 followers Ryanair - Europe's Favourite Airline Airlines and Aviation 866,770 followers Show more Show less Similar products Freshchat Freshchat Live Chat Software ChatR ChatR Live Chat Software ICE Chat ICE Chat Live Chat Software Webchat Webchat Live Chat Software Microsoft Teams Microsoft Teams Live Chat Software LiveChat LiveChat Live Chat Software Sign in to see more Show more Show less Text products ChatBot ChatBot Chatbot Platforms Software HelpDesk HelpDesk Help Desk Software LiveChat LiveChat Live Chat Software OpenWidget OpenWidget LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/aitana-microsoft-teams/?trk=products_seo_search
Microsoft Teams | LinkedIn Skip to main content LinkedIn AITANA in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Microsoft Teams Live Chat Software by AITANA See who's skilled in this Add as skill Learn more Report this product About Microsoft Teams es una plataforma de comunicación diseñada para el trabajo en equipo. Con esta aplicación, incluida en el paquete de Office 365, tus empleados podrán hacer videollamadas, chatear, realizar y grabar reuniones, compartir archivos e incluso crear equipos de trabajo. Esta herramienta cuenta ya con más de 85 millones de usuarios en todo el mundo. Además, en su área de trabajo también están integradas las herramientas ofimáticas de Office 365 (ahora Microsoft 365): Word, Excel, PowerPoint, SharePoint, Power BI, etc. Gran parte del éxito en las empresas se basa en el trabajo en equipo. Por ello, Microsoft ha diseñado esta plataforma donde las compañías pueden apoyarse para fomentar la colaboración y crear conversaciones fluidas entre los empleados. Media Products media viewer No more previous content Microsoft Teams Te presentamos Microsoft Teams, la solución de Office 365 (ahora Microsoft 365) para realizar llamadas, videollamadas, chatear en grupo, crear equipos de trabajo... una forma óptima de trabajar en equipo. El teletrabajo ya es posible con esta herramienta de comunicación. Demostración de Microsoft Teams. Teletrabaja de forma eficiente En este vídeo te explicamos una de las herramientas más conocidas y usadas hoy en día en el mercado para poner en marcha el teletrabajo. Se trata de una sesión práctica que comienza con una breve introducción sobre Microsoft Teams y todo lo que puede ofrecer a las empresas y sus trabajadores, y en la que se realiza una demostración en directo. Trabaja con Dynamics 365 desde Microsoft Teams Conocemos cómo se puede trabajar con Dynamics 365 Sales (CRM) desde Microsoft Teams. Básicamente veremos cómo Teams, además de ser ese canal de comunicación interno o ese hub de comunicación, es una plataforma con la que se pueden gestionar las relaciones comerciales con los clientes. La 9ª sinfonía de Microsoft Business Central integrado en Teams Realiza todas las tareas de Business Central sin salir de Microsoft Teams y aumenta la productividad de tu equipo. En este webinar te enseñaremos cómo la integración de Business Central con Teams te permite ser más productivo y evitar pérdidas de tiempo e información. Microsoft Teams a fondo Microsoft Teams es el área de trabajo basada en chat que se integra con Office 365. Ofrece a los trabajadores acceso instantáneo a todo lo que necesitan en un lugar centralizado donde el chat, el contenido, las personas y las herramientas conviven juntos. En este vídeo te explicamos cómo moverte por Microsoft Teams y cuáles son las ventajas que te puede ofrecer. No more next content Similar products Freshchat Freshchat Live Chat Software ChatR ChatR Live Chat Software ICE Chat ICE Chat Live Chat Software Webchat Webchat Live Chat Software LiveChat LiveChat Live Chat Software Chat Chat Live Chat Software Sign in to see more Show more Show less AITANA products Dynamics 365 Customer Service Dynamics 365 Customer Service Customer Relationship Management (CRM) Software Dynamics 365 Sales Dynamics 365 Sales Customer Relationship Management (CRM) Software Finance BI Finance BI Business Intelligence (BI) Software Microsoft 365 Microsoft 365 Office Suites Software Portal del empleado - Intranet Portal del empleado - Intranet Intranet Software Power Apps Power Apps Robotic Process Automation (RPA) Software Power Automate Power Automate Robotic Process Automation (RPA) Software Sales BI Sales BI Business Intelligence (BI) Software SharePoint SharePoint Intranet Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://scholar.google.com/citations?hl=ko&oe=ASCII&user=5HoF_9oAAAAJ&view_op=list_works
:root{--gm3-sys-color-on-surface-rgb:31,31,31}@media screen and (prefers-color-scheme:dark){:root{--gm3-sys-color-on-surface-rgb:227,227,227}}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.q4Wquf,.nfoC7c{display:block;height:25vh;position:relative}@media (min-width:600px){.q4Wquf,.nfoC7c{height:150px}}.q4Wquf.Irjbwb{height:auto}@media screen and (prefers-color-scheme:dark){.q4Wquf:not(.GtvzYd){display:none}}.nfoC7c{margin:0;overflow:hidden}.PwpMUe,.lVUmD{display:block;height:100%;margin:0 auto;width:100%}.St9mde{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.St9mde{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.PwpMUe,.lVUmD,.St9mde{-o-object-fit:contain;object-fit:contain}.wsArZ[data-ss-mode="1"] .q4Wquf,.wsArZ[data-ss-mode="1"] .St9mde{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .St9mde{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .q4Wquf,.NQ5OL .St9mde{height:auto;width:100%}.NQ5OL .St9mde{max-width:400px}}.q4Wquf.NWba7e,.q4Wquf.NWba7e .St9mde{height:auto}.q4Wquf.NWba7e .St9mde{height:auto;max-width:312px;width:100%}.q4Wquf.NWba7e.zpCp3 .St9mde{max-width:unset}.q4Wquf.IiQozc .St9mde{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.q4Wquf.Irjbwb .St9mde{height:auto;width:100%}.q4Wquf.EEeaqf .St9mde{max-height:144px;max-width:144px}.SnAaEd{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.SnAaEd{display:none}}.SnAaEd::after,.SnAaEd::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.SnAaEd::after,.SnAaEd::before{width:150px}}.SnAaEd::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.SnAaEd::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .SnAaEd~.St9mde{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .SnAaEd~.St9mde{width:auto}}.RHNWk .St9mde{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .RHNWk .St9mde{width:115px}}.cf660d .St9mde{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.tUhwwc .St9mde{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.Jkvqxd .St9mde{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.onc8Ic .St9mde{-webkit-transform:translate(var( --c-ps-s,24px ),0);-ms-transform:translate(var( --c-ps-s,24px ),0);transform:translate(var( --c-ps-s,24px ),0)}.WA89Yb .St9mde{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .XEN8Yb .St9mde{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .XEN8Yb .St9mde{max-width:115px}}@media (min-width:1240px) and (orientation:landscape),all and (min-width:1600px){.KyPKed .NQ5OL .XEN8Yb .St9mde{max-width:115px}}.IsSr6b .St9mde{max-width:300px}.mmskdd .St9mde{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@-webkit-keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@-webkit-keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@-webkit-keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}:root{--wf-tfs:calc(var(--c-tfs,32)/16*1rem);--wf-tfs-bp2:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp3:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp5:calc(var(--c-tfs,44)/16*1rem);--wf-stfs:calc(var(--c-stfs,16)/16*1rem);--wf-stfs-bp5:calc(var(--c-stfs,16)/16*1rem)}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.Dzz9Db,.GpMPBe{display:block;height:25vh;position:relative}@media (min-width:600px){.Dzz9Db,.GpMPBe{height:150px}}@media screen and (prefers-color-scheme:dark){.Dzz9Db:not(.GtvzYd){display:none}}.Dzz9Db.Irjbwb{height:auto}.GpMPBe{margin:0;overflow:hidden}.UFQPDd,.JNOvdd{display:block;height:100%;margin:0 auto;-o-object-fit:contain;object-fit:contain;width:100%}.f4ZpM{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.f4ZpM{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.wsArZ[data-ss-mode="1"] .Dzz9Db,.wsArZ[data-ss-mode="1"] .f4ZpM{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .f4ZpM{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .Dzz9Db,.NQ5OL .f4ZpM{height:auto;width:100%}.NQ5OL .f4ZpM{max-width:400px}}.Dzz9Db.utFBGf,.Dzz9Db.utFBGf .f4ZpM{height:auto}.Dzz9Db.utFBGf .f4ZpM{height:auto;max-width:312px;width:100%}.Dzz9Db.utFBGf.zpCp3 .f4ZpM{max-width:unset}.Dzz9Db.IiQozc .f4ZpM{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.Dzz9Db.Irjbwb .f4ZpM{height:auto;width:100%}.Dzz9Db.EEeaqf .f4ZpM{max-height:144px;max-width:144px}.nPt1pc{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.nPt1pc{display:none}}.nPt1pc::after,.nPt1pc::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.nPt1pc::after,.nPt1pc::before{width:150px}}.nPt1pc::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.nPt1pc::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .nPt1pc~.f4ZpM{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .nPt1pc~.f4ZpM{width:auto}}.ZS7CGc .f4ZpM{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .ZS7CGc .f4ZpM{width:115px}}.qiRZ5e .f4ZpM{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.vIv7Gf .f4ZpM{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.nvYXVd .f4ZpM{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.uOhnzd .f4ZpM{-webkit-transform:translate(24px,0);-ms-transform:translate(24px,0);transform:translate(24px,0)}.MsYMaf .f4ZpM{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .YIi9qf .f4ZpM{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .YIi9qf .f4ZpM{max-width:115px}}.QG3Xbe .f4ZpM{max-width:300px}.F6gtje .f4ZpM{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@-webkit-keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@-webkit-keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}@keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}.VfPpkd-ksKsZd-XxIAqe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity;position:relative;outline:none;overflow:hidden}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-ksKsZd-XxIAqe::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-ksKsZd-XxIAqe::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded],.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd{overflow:visible}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:0;left:0;width:100%;height:100%}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-ksKsZd-XxIAqe:hover::before,.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe::before,.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):focus::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d)::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):active::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe{font-size:24px;width:48px;height:48px;padding:12px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-Jh9lGc{width:40px;height:40px;margin-top:4px;margin-bottom:4px;margin-right:4px;margin-left:4px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-J1Ukfc-LhBDec{max-height:40px;max-width:40px}.VfPpkd-Bz112c-LgbsSe:disabled{color:rgba(0,0,0,.38);color:var(--mdc-theme-text-disabled-on-light,rgba(0,0,0,.38))}.VfPpkd-Bz112c-LgbsSe svg,.VfPpkd-Bz112c-LgbsSe img{width:24px;height:24px}.VfPpkd-Bz112c-LgbsSe{display:inline-block;position:relative;-webkit-box-sizing:border-box;box-sizing:border-box;border:none;outline:none;background-color:transparent;fill:currentColor;color:inherit;text-decoration:none;cursor:pointer;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;z-index:0;overflow:visible}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-RLmnJb{position:absolute;top:50%;height:48px;left:50%;width:48px;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-J1Ukfc-LhBDec,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-J1Ukfc-LhBDec{display:block}}.VfPpkd-Bz112c-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-Bz112c-LgbsSe[hidden]{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-KVuj8d-Q3DXx{-webkit-box-align:center;-webkit-align-items:center;align-items:center;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center}.VfPpkd-Bz112c-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:100%;width:100%;display:none}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{border-color:CanvasText}}.VfPpkd-Bz112c-kBDsod{display:inline-block}.VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd,.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd{display:inline-block}.VfPpkd-Bz112c-mRLv6{height:100%;left:0;outline:none;position:absolute;top:0;width:100%}.VfPpkd-Bz112c-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;width:100%;height:100%}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-Bz112c-LgbsSe:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe:disabled:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc{height:100%;left:0;pointer-events:none;position:absolute;top:0;width:100%;z-index:-1}.VfPpkd-dgl2Hf-ppHlrf-sM5MNb{display:inline}.VfPpkd-LgbsSe{position:relative;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-align:center;-webkit-align-items:center;align-items:center;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center;-webkit-box-sizing:border-box;box-sizing:border-box;min-width:64px;border:none;outline:none;line-height:inherit;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;-webkit-appearance:none;overflow:visible;vertical-align:middle;background:transparent}.VfPpkd-LgbsSe .VfPpkd-BFbNVe-bF1uUb{width:100%;height:100%;top:0;left:0}.VfPpkd-LgbsSe::-moz-focus-inner{padding:0;border:0}.VfPpkd-LgbsSe:active{outline:none}.VfPpkd-LgbsSe:hover{cursor:pointer}.VfPpkd-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-LgbsSe[hidden]{display:none}.VfPpkd-LgbsSe .VfPpkd-kBDsod{margin-left:0;margin-right:8px;display:inline-block;position:relative;vertical-align:top}[dir=rtl] .VfPpkd-LgbsSe .VfPpkd-kBDsod,.VfPpkd-LgbsSe .VfPpkd-kBDsod[dir=rtl]{margin-left:8px;margin-right:0}.VfPpkd-LgbsSe .VfPpkd-UdE5de-uDEFge{font-size:0;position:absolute;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);top:50%;left:50%;line-height:normal}.VfPpkd-LgbsSe .VfPpkd-vQzf8d{position:relative}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px);display:none}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-color:CanvasText}}@media screen and (forced-colors:active){.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-J1Ukfc-LhBDec,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-J1Ukfc-LhBDec{display:block}}.VfPpkd-LgbsSe .VfPpkd-RLmnJb{position:absolute;top:50%;height:48px;left:0;right:0;-webkit-transform:translateY(-50%);-ms-transform:translateY(-50%);transform:translateY(-50%)}.VfPpkd-vQzf8d+.VfPpkd-kBDsod{margin-left:8px;margin-right:0}[dir=rtl] .VfPpkd-vQzf8d+.VfPpkd-kBDsod,.VfPpkd-vQzf8d+.VfPpkd-kBDsod[dir=rtl]{margin-left:0;margin-right:8px}svg.VfPpkd-kBDsod{fill:currentColor}.VfPpkd-LgbsSe-OWXEXe-dgl2Hf{margin-top:6px;margin-bottom:6px}.VfPpkd-LgbsSe{-moz-osx-font-smoothing:grayscale;-webkit-font-smoothing:antialiased;text-decoration:none}.VfPpkd-LgbsSe{padding:0 8px 0 8px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-INsAgc{border-style:solid;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-style:solid;border-color:transparent}.VfPpkd-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{z-index:0}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Jh9lGc{position:absolute;-webkit-box-sizing:content-box;box-sizing:content-box;overflow:hidden;z-index:0;top:0;left:0;bottom:0;right:0}.VfPpkd-LgbsSe{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-text-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-filled-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;-webkit-box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12);box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-protected-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled):active{-webkit-box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12);box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{-webkit-box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12);box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;padding:0 15px 0 15px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-outlined-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 11px 0 15px}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 15px 0 11px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{top:-1px;left:-1px;bottom:-1px;right:-1px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-RLmnJb{left:-1px;width:calc(100% + 2px)}.nCP5yc{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.nCP5yc .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.nCP5yc:not(:disabled){background-color:rgb(26,115,232);background-color:var(--gm-fillbutton-container-color,rgb(26,115,232))}.nCP5yc:not(:disabled){color:#fff;color:var(--gm-fillbutton-ink-color,#fff)}.nCP5yc:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-fillbutton-disabled-container-color,rgba(60,64,67,.12))}.nCP5yc:disabled{color:rgba(60,64,67,.38);color:var(--gm-fillbutton-disabled-ink-color,rgba(60,64,67,.38))}.nCP5yc .VfPpkd-Jh9lGc::before,.nCP5yc .VfPpkd-Jh9lGc::after{background-color:rgb(32,33,36);background-color:var(--gm-fillbutton-state-color,rgb(32,33,36))}.nCP5yc:hover .VfPpkd-Jh9lGc::before,.nCP5yc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.16;opacity:var(--mdc-ripple-hover-opacity,.16)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24;opacity:var(--mdc-ripple-focus-opacity,.24)}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.2;opacity:var(--mdc-ripple-press-opacity,.2)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.2)}.nCP5yc .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:#fff}@media (-ms-high-contrast:active),screen and (forced-colors:active){.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.nCP5yc:hover{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:hover .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:active{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:active .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:disabled{-webkit-box-shadow:none;box-shadow:none}.nCP5yc:disabled:hover .VfPpkd-Jh9lGc::before,.nCP5yc:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.Rj2Mlf .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.Rj2Mlf:not(:disabled){color:rgb(26,115,232);color:var(--gm-hairlinebutton-ink-color,rgb(26,115,232))}.Rj2Mlf:not(:disabled){border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled):hover{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled).VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.Rj2Mlf:not(:disabled):not(.VfPpkd-ksKsZd-mWPk3d):focus{border-color:rgb(23,78,166);border-color:var(--gm-hairlinebutton-outline-color--stateful,rgb(23,78,166))}.Rj2Mlf:not(:disabled):active,.Rj2Mlf:not(:disabled):focus:active{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:disabled{color:rgba(60,64,67,.38);color:var(--gm-hairlinebutton-disabled-ink-color,rgba(60,64,67,.38))}.Rj2Mlf:disabled{border-color:rgba(60,64,67,.12);border-color:var(--gm-hairlinebutton-disabled-outline-color,rgba(60,64,67,.12))}.Rj2Mlf:hover:not(:disabled),.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.Rj2Mlf:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-hairlinebutton-ink-color--stateful,rgb(23,78,166))}.Rj2Mlf .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.Rj2Mlf .VfPpkd-Jh9lGc::before,.Rj2Mlf .VfPpkd-Jh9lGc::after{background-color:rgb(26,115,232);background-color:var(--gm-hairlinebutton-state-color,rgb(26,115,232))}.Rj2Mlf:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.Rj2Mlf:disabled:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.b9hyVd{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);border-width:0;-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15))}.b9hyVd .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.b9hyVd:not(:disabled){background-color:#fff;background-color:var(--gm-protectedbutton-container-color,#fff)}.b9hyVd:not(:disabled){color:rgb(26,115,232);color:var(--gm-protectedbutton-ink-color,rgb(26,115,232))}.b9hyVd:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-protectedbutton-disabled-container-color,rgba(60,64,67,.12))}.b9hyVd:disabled{color:rgba(60,64,67,.38);color:var(--gm-protectedbutton-disabled-ink-color,rgba(60,64,67,.38))}.b9hyVd:hover:not(:disabled),.b9hyVd.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.b9hyVd:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.b9hyVd:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-protectedbutton-ink-color--stateful,rgb(23,78,166))}.b9hyVd .VfPpkd-BFbNVe-bF1uUb{opacity:0}.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-B
2026-01-13T09:29:19
http://www.trello.com/inbox
Trello Inbox: Never miss a beat | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Trello Inbox Capture, organize, and conquer every to-do Try Inbox for free! Learn more about Trello’s plans and pricing. Capture anywhere, anytime Say goodbye to lost to-dos and scattered ideas! Trello Inbox saves you from the chaos of emails, messages, and notes. Jot down thoughts as they come or effortlessly snatch snippets from your favorite tools—no need to organize them right away. Instant capture No more missed opportunities! Capture tasks and ideas from emails and your favorite messaging apps—like Slack and Microsoft Teams—so nothing important slips away. Seamless organization Ready to tidy up? Simply drag and drop your captured items into the right boards, making organization a breeze. Intuitive organization Got quick to-dos? Mark them "Done" straight from the Inbox and keep your momentum going! Join the Trello Inbox revolution! We’re thrilled to bring you Trello Inbox, and we want your input to make it even better! Try it out, share your feedback, and help us innovate the future of to-do list management. Get started Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://www.linkedin.com/products/freshworks-inc-freshworks-customer-service-suite/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_full-click#main-content
Freshdesk Omni | LinkedIn Skip to main content LinkedIn Freshworks in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Freshdesk Omni Help Desk Software by Freshworks See who's skilled in this Add as skill Learn more Report this product About Freshdesk Omni provides a seamless experience across channels, from automated self-service to agent-led conversational support, plus advanced ticketing management for issues that require collaboration across teams, all powered by Freddy AI This product is intended for Customer Service Specialist Customer Service Representative Customer Support Specialist Customer Service Manager Media Products media viewer No more previous content All-in-one solution Improve customer experience and agent productivity with generative AI. One seamless omnichannel solution for self-service, conversational support, and ticketing. Supercharge agent productivity Empower agents to personalize service on the fly. View all customer interactions across channels in one place. Freddy Copilot is an always-on assistant that can suggest responses, recommend actions, and summarize events. Enable seamless collaboration across teams Reach out to colleagues to swiftly resolve customer issues. Swarm fast. Rally a dream team. Track accountability. Advanced ticketing keeps all the information in one place, so it’s easy to stay on top of progress and no one drops the ball. Make smart decisions faster Powered by generative AI, Freddy Insights proactively analyzes performance data for opportunities to optimize productivity and head off potential issues. Freddy Insights even recommends remediation actions—and helps execute them. Drive productivity and enable excellent support Give admins the tools to streamline operations and boost efficiency with AI-powered workflow automations. No more next content Featured customers of Freshdesk Omni Klarna Software Development 390,433 followers Travix Software Development 16,312 followers MultiChoice Group Broadcast Media Production and Distribution 483,508 followers BEL USA LLC Printing Services 7,794 followers PhonePe Software Development 1,417,091 followers Blue Nile Retail 33,285 followers Trainline Technology, Information and Internet 55,507 followers Show more Show less Similar products Service Cloud Service Cloud Help Desk Software Zoho Desk Zoho Desk Help Desk Software Freshdesk Freshdesk Help Desk Software Zendesk Suite Zendesk Suite Help Desk Software Zendesk for Employee Experience Zendesk for Employee Experience Help Desk Software Hiver Hiver Help Desk Software Sign in to see more Show more Show less Freshworks products Freshchat Freshchat Live Chat Software Freshdesk Freshdesk Help Desk Software Freshmarketer Freshmarketer Marketing Automation Software Freshsales Freshsales Customer Relationship Management (CRM) Software Freshservice Freshservice Service Desk Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
http://www.trello.com/home
Capture, organize, and tackle your to-dos from anywhere | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Accelerate your teams' work with AI features 🤖 now available for all Premium and Enterprise! Learn more. Capture, organize, and tackle your to-dos from anywhere. Escape the clutter and chaos—unleash your productivity with Trello. Sign up - it’s free! By entering my email, I acknowledge the Atlassian Privacy Policy Watch video Hero Trello 101 Your productivity powerhouse Stay organized and efficient with Inbox, Boards, and Planner. Every to-do, idea, or responsibility—no matter how small—finds its place, keeping you at the top of your game. Inbox When it’s on your mind, it goes in your Inbox. Capture your to-dos from anywhere, anytime. Boards Your to-do list may be long, but it can be manageable! Keep tabs on everything from "to-dos to tackle" to "mission accomplished!” Planner Drag, drop, get it done. Snap your top tasks into your calendar and make time for what truly matters. Inbox When it’s on your mind, it goes in your Inbox. Capture your to-dos from anywhere, anytime. Boards Your to-do list may be long, but it can be manageable! Keep tabs on everything from "to-dos to tackle" to "mission accomplished!” Planner Drag, drop, get it done. Snap your top tasks into your calendar and make time for what truly matters. From message to action Quickly turn communication from your favorite apps into to-dos, keeping all your discussions and tasks organized in one place. EMAIL MAGIC Easily turn your emails into to-dos! Just forward them to your Trello Inbox, and they’ll be transformed by AI into organized to-dos with all the links you need. MESSAGE APP SORCERY Need to follow up on a message from Slack or Microsoft Teams? Send it directly to your Trello board! Your favorite app interface lets you save messages that appear in your Trello Inbox with AI-generated summaries and links. WORK SMARTER Do more with Trello Customize the way you organize with easy integrations, automation, and mirroring of your to-dos across multiple locations. Integrations Connect the apps you are already using into your Trello workflow or add a Power-Up to fine-tune your specific needs. Browse Integrations Automation No-code automation is built into every Trello board. Focus on the work that matters most and let the robots do the rest. Get to know Automation Card mirroring View all your to-dos from multiple boards in one place. Mirror a card to keep track of work wherever you need it! Compare plans [Trello is] great for simplifying complex processes. As a manager, I can chunk [processes] down into bite-sized pieces for my team and then delegate that out, but still keep a bird's-eye view. Joey Rosenberg Global Leadership Director at Women Who Code Read the story 75% of organizations report that Trello delivers value to their business within 30 days. Trello TechValidate Survey Whether someone is in the office, working from home, or working on-site with a client, everyone can share context and information through Trello. Sumeet Moghe Product Manager at ThoughtWorks Read the story 81% of customers chose Trello for its ease of use. Trello TechValidate Survey We used Trello to provide clarity on steps, requirements, and procedures. This was exceptional when communicating with teams that had deep cultural and language differences. Jefferson Scomacao Development Manager at IKEA/PTC Read the story 74% of customers say Trello has improved communication with their co-workers and teams. Trello TechValidate Survey Join a community of millions of users globally who are using Trello to get more done. Join a community of millions of users globally who are using Trello to get more done. Get started with Trello today Sign up - it’s free! By entering my email, I acknowledge the Atlassian Privacy Policy Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://sre.google/workbook/table-of-contents/
Google SRE - SRE workbook table of content Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Table of Contents Table of Contents Foreword I Foreword II Preface Chapter 1 - How SRE Relates to DevOps Part I - Foundations Chapter 2 - Implementing SLOs Chapter 3 - SLO Engineering Case Studies Chapter 4 - Monitoring Chapter 5 - Alerting on SLOs Chapter 6 - Eliminating Toil Chapter 7 - Simplicity Part II - Practices Chapter 8 - On-Call Chapter 9 - Incident Response Chapter 10 - Postmortem Culture: Learning from Failure Chapter 11 - Managing Load Chapter 12 - Introducing Non-Abstract Large System Design Chapter 13 - Data Processing Pipelines Chapter 14 - Configuration Design and Best Practices Chapter 15 - Configuration Specifics Chapter 16 - Canarying Releases Part III - Processes Chapter 17 - Identifying and Recovering from Overload Chapter 18 - SRE Engagement Model Chapter 19 - SRE: Reaching Beyond Your Walls Chapter 20 - SRE Team Lifecycles Chapter 21 - Organizational Change Management in SRE Conclusion Appendix A - Example SLO Document Appendix B - Example Error Budget Policy Appendix C - Results of Postmortem Analysis Index About the Editors Colophon Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://www.dataport.de/dwebservice/?mtm_campaign=dwebservice&mtm_source=linkedin&mtm_medium=pp&mtm_content=awareness&mtm_group=rv13&trk=products_details_guest_secondary_call_to_action
dWebService: Hosting. Webportale. Analysen. | Dataport Ihre Cookie-Einstellungen für diese Webseite Treffen Sie hier Ihre persönlichen Einstellungen Erforderliche Cookies Erforderliche Cookies sind für die Nutzung der Webseite zwingend erforderlich. Die Webseite kann ohne diese Cookies nicht richtig funktionieren. Einzelne Cookies Cookiename: cookiehint; Anbieter: Dataport; Zweck: Dieser Cookie speichert die Einstellung zu Cookies; Laufzeit: 1 Jahr Statistik Diese Cookies helfen uns bei der Analyse des Nutzerverhaltens, um die Qualität unserer Webseite fortlaufend zu verbessern. Einzelne Cookies Cookiename: _pk_id.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, wie die Webseite von Besucher*innen genutzt wird; Laufzeit: 12 Monate Cookiename: _pk_ref.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, woher die Besucher*innen dieser Webseite gekommen sind; Laufzeit: 6 Monate Cookiename: _pk_ses.*, _pk_cvar.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, wie die Webseite von Besucher*innen genutzt wird; Laufzeit: 30 Minuten Weitere Informationen zu Cookies finden Sie in der Datenschutzerklärung Abbrechen Speichern Ihre Daten sind bei uns sicher Als Datenschutzexpert*innen klären wir transparent darüber auf, welche Daten wie und warum verarbeitet werden ( zur Datenschutzerklärung zum Impressum ). Auf dieser Website verwenden wir Cookies. Das sind kleine Dateien, die lokal im Browser gespeichert werden. Einige dieser Cookies sind unbedingt erforderlich, damit Sie diese Website nutzen können und alles richtig funktioniert. Über die Verwendung anderer Cookies entscheiden Sie selbst. Ganz gleich wie Sie sich entscheiden – wir respektieren Ihre Privatsphäre und wir gehen stets verantwortungsvoll mit den Daten um, die Sie uns anvertrauen. Ich entscheide selbst Ich bin einverstanden Ich lehne ab Zur Suche Ansprechpartner*innen finden Störung melden Zu den Jobs Zum Inhalt Menü Unternehmen Dataport Nachhaltigkeit Diversity Digitale Souveränität Geschäftsbericht Organe Partner Services & Produkte Consulting Innovationsmanagement Projektmanagement Rechenzentrum Druckzentrum Software-Entwicklung Weiterbildung Client-Betriebsmodelle IT-Beschaffung SAP-Kompetenzzentrum Fachanwendungen IT-Produkte Unsere Themen Künstliche Intelligenz IT Sicherheit Cloud Onlinezugangsgesetz Open Source Netze Alle Themen Karriere Arbeiten bei Dataport Onboarding Benefits Rund ums Bewerben Qualifizierung Einstieg Aufgabengebiete Veranstaltungen Insights Podcast Jobs Newsroom Übersicht Stories News & Presse Downloads Datareport Support Leichte Sprache Gebärden-sprache English Summary Dialog schließen Ihr direkter Weg zum Ziel Ihr Suchbegriff Suchen Dialog schließen Ansprechpartner*innen Bitte wählen Sie Ihr Anliegen Bitte wählen Pressefragen Karrierefragen Produkte und Vertrieb Internetseite Britta Heinrich Pressesprecherin 040 42846-3047 0171 3342284 E-Mail schreiben Karen Hoffmann-Abraham Stellvertretende Pressesprecherin 040 42846-3736 0176 42860910 E-Mail schreiben Karriereteam E-Mail schreiben Michael Schulz-Heirich Bereichsleiter Vertrieb 040 42846-2881 E-Mail schreiben Online-Redaktion E-Mail schreiben Zentrale: 0431 3295-0 poststelle@dataport.de Mo–Fr: 06:30–18:00 Uhr Dialog schließen Störungsmeldung Region wählen Bitte wählen Hamburg Schleswig-Holstein Bremen Sachsen-Anhalt Kreis Dithmarschen Störungsart wählen Bitte wählen Bei Problemen mit dem BASIS-Standardarbeitsplatz Bei Problemen mit der Schulverwaltungsplattform TUVAS Bei Problemen mit LOGINEO Bei allen anderen IT-Störungen Bei Telefonstörung Bei Mobilfunkproblemen Störungsart wählen Bitte wählen Bei Problemen mit der Plattform ABS Bei Problemen mit dem +1-Arbeitsplatz Bei anderen IT-Störungen Störungsart wählen Bitte wählen Bei Problemen mit dem BASIS-Standardarbeitsplatz Bei Problemen mit dem Fachverfahren MESO Bei allen anderen IT-Störungen Bei Mobilfunkproblemen Störungsart wählen Bitte wählen Bei IT-Problemen mit den StApl-Arbeitsplätzen Bei IT-Problemen aller anderen Kunden Bei IT-Problemen der Polizei ST Störungsart wählen Bitte wählen Bei IT-Störungen Ihre Rufnummer 040 428-333 Ihre Rufnummer 040 42846-3990 Ihre Rufnummer 040 42846-4284 Ihre Rufnummer 040 42846-750 Ihre Rufnummer 040 428-22 Ihre Rufnummer 040 42846-1965 Ihre Rufnummer 0431 3295 7000 Ihre Rufnummer 0431 3295-888 Ihre Rufnummer 0431 3295-444 Ihre Rufnummer 0421 361-55555 Ihre Rufnummer 0421 83558-3559 Ihre Rufnummer 0421 83558-444 Ihre Rufnummer 040 42846-1965 Ihre Rufnummer 0391 6627-1500 Ihre Rufnummer 0391 6627-1545 Ihre Rufnummer 0391-6627-1501 Ihre Rufnummer 0431 3295-3295 Dialog schließen Jobs Zum Jobportal Jetzt Jobalert einrichten Startseite Services & Produkte IT-Produkte dWebService dWebService – Hosting. Webportale. Analysen. Jetzt Beratungstermin vereinbaren Kurz erklärt Ob umfangreiche Webpräsenz oder einfach nur Informationen bereitstellen – ohne Webseite geht es nicht mehr. Dabei unterstützt Sie dWebService. Einfach, schnell, flexibel. Vom simplen Webspace bis hin zu einem vollständigen Content Management System. Inklusive sicherem Hosting, Barrierefreiheits-Check und smarten Nutzungsanalysen. Alles aus einer Hand. Mit einer festen Kontaktperson, die für Sie da ist. Produktvorteile Barrierefreie Webportale Gemeinsam entwickeln Sie mit uns Webportale, die etablierten, anerkannten Standards im Public Sektor entsprechen. Mit TYPO3 als Open-Source-Lösung setzen Sie auf eine moderne, skalierbare und inklusive Plattform, die für alle Nutzenden zugänglich ist. Internet- und Intranet-Lösungen Spezialisiert für die Anforderungen Ihres jeweiligen Geschäftsfelds und der öffentlichen Verwaltung Natürlich barrierefrei mit zertifizierten Standards Digital souverän umgesetzt mit dem Open-Source-CMS TYPO3 Zertifizierte Berater*innen und Entwickler*innen Sicheres Hosting  Vertrauen Sie dank des BSI-zertifizierten Rechenzentrums von Dataport auf höchste Sicherheitsstandards – egal, ob Sie sich für Shared Services oder individuelle Hosting-Optionen entscheiden. So sind Ihre Daten bestens geschützt und Sie behalten die Flexibilität, die Sie für Ihre Projekte benötigen. Hosting einem der besten Rechenzentren Europas BSI-zertifiziert für maximale Sicherheit Herausragend gesichert gegen physische Angriffe und Hacker Kostengünstige, schnell verfügbare Shared Services Individual-Hosting-Services mit Premium Service Level Agreements Smarte Nutzungsanalysen Mit unserer DSGVO- und TDDDG-konformen Analyse-Software Matomo gewinnen Sie wertvolle Einblicke in das Nutzerverhalten – ganz ohne Cookies. So können Sie datenschutzkonform fundierte Entscheidungen treffen und Ihre Online-Strategie optimieren. Nutzungsanalysen für jede Form von Webportalen DSGVO- und TDDDG-konform mit dem Open-Source-Werkzeug Matomo Cookie-less Tracking Hilft bei Verbesserungen und Weiterentwicklungen Schluss mit gefühlten Wahrheiten – datenbasierte Entscheidungen Leistungen im Überblick Individuelle Beratung und maßgeschneiderte Entwicklung von Webportalen und IT-Lösungen Schulungen und Workshops zur Einführung und optimalen Nutzung unserer Lösungen Hosting im BSI-zertifizierten Twin Data Center für höchste Sicherheitsstandards Barrierefreie Webentwicklung nach zertifizierten Standards 24/7-Überwachung und Schutz durch unser Security Operations Center Optimierung und Standardisierung Ihrer Prozesse für mehr Effizienz DSGVO-konforme Nutzungsanalysen mit der Open-Source-Software Matomo Entwicklung und Betreuung von Webportalen mit dem Open-Source-CMS TYPO3 Integration von Active Directories und Single-Sign-On für eine nahtlose Benutzerverwaltung Umsetzung durch zertifizierte Expert*innen mit langjähriger Erfahrung Downloads Produktblatt dWebService (PDF) 279 KB Jetzt starten! Sie möchten, dass dWebService für Sie Hosting, Webportalentwicklung und Nutzeranalysen übernimmt? Dann melden Sie sich jetzt zu einem unverbindlichen Beratungsgespräch an. * Pflichtfelder Vorname * Nachname * E-Mail-Adresse * Telefon Behörde/Institution/Firma * Ihre Nachricht Einwilligung für die elektronische Übermittlung * Ich willige in die elektronische Übermittlung meiner Daten gemäß Art. 6 Abs. 1 lit. a DSGVO ein (Opt-In). Eine Weiterverarbeitung der übermittelten Daten erfolgt ggf. auf Basis einer anderen Rechtsgrundlage. Absenden Weitere Informationen finden Sie in der Datenschutzerklärung . Bitte dieses Feld NICHT ausfüllen! Ihr Kontakt Michael Schulz-Heirich Bereichsleiter Vertrieb 040 42846-2881 E-Mail schreiben Team dWebService E-Mail schreiben Weiterführende Informationen Bildnachweise Bildnachweise © Dataport © ronstik/stock.adobe.com © REDPIXEL/stock.adobe.com © Dataport Schwerpunktthemen Open Source IT Sicherheit Onlinezugangsgesetz Cloud Netze Services & Produkte Consulting Innovationsmanagement Projektmanagement Rechenzentrum Druckzentrum Software-Entwicklung Weiterbildung Client-Betriebsmodelle IT-Beschaffung SAP-Kompetenzzentrum Fachanwendungen IT-Produkte Wir machen digitale Zukunft Starten Sie mit uns durch! Karriere bei Dataport Zu den Stellenangeboten Kontakt Dataport Altenholz Standort (Zentrale) Altenholzer Straße 10-14 24161 Altenholz 0431 3295-0 poststelle@dataport.de Kontakt Zertifizierung Portale Kundenportal Jobportal Weiterbildungsportal Shop Vergabeplattform dataport.kommunal Folgen Sie uns Kontakt Impressum AGB Datenschutz Barrierefreiheit Vergabeverfahren Fördermittelprojekte Sitemap
2026-01-13T09:29:19
https://git-scm.com/book/fr/v2/Les-bases-de-Git-D%c3%a9marrer-un-d%c3%a9p%c3%b4t-Git
Git - Démarrer un dépôt Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Démarrage rapide 1.1 À propos de la gestion de version 1.2 Une rapide histoire de Git 1.3 Rudiments de Git 1.4 La ligne de commande 1.5 Installation de Git 1.6 Paramétrage à la première utilisation de Git 1.7 Obtenir de l’aide 1.8 Résumé 2. Les bases de Git 2.1 Démarrer un dépôt Git 2.2 Enregistrer des modifications dans le dépôt 2.3 Visualiser l’historique des validations 2.4 Annuler des actions 2.5 Travailler avec des dépôts distants 2.6 Étiquetage 2.7 Les alias Git 2.8 Résumé 3. Les branches avec Git 3.1 Les branches en bref 3.2 Branches et fusions : les bases 3.3 Gestion des branches 3.4 Travailler avec les branches 3.5 Branches de suivi à distance 3.6 Rebaser (Rebasing) 3.7 Résumé 4. Git sur le serveur 4.1 Protocoles 4.2 Installation de Git sur un serveur 4.3 Génération des clés publiques SSH 4.4 Mise en place du serveur 4.5 Démon (Daemon) Git 4.6 HTTP intelligent 4.7 GitWeb 4.8 GitLab 4.9 Git hébergé 4.10 Résumé 5. Git distribué 5.1 Développements distribués 5.2 Contribution à un projet 5.3 Maintenance d’un projet 5.4 Résumé 6. GitHub 6.1 Configuration et paramétrage d’un compte 6.2 Contribution à un projet 6.3 Maintenance d’un projet 6.4 Gestion d’un regroupement 6.5 Écriture de scripts pour GitHub 6.6 Résumé 7. Utilitaires Git 7.1 Sélection des versions 7.2 Indexation interactive 7.3 Remisage et nettoyage 7.4 Signer votre travail 7.5 Recherche 7.6 Réécrire l’historique 7.7 Reset démystifié 7.8 Fusion avancée 7.9 Rerere 7.10 Déboguer avec Git 7.11 Sous-modules 7.12 Empaquetage (bundling) 7.13 Replace 7.14 Stockage des identifiants 7.15 Résumé 8. Personnalisation de Git 8.1 Configuration de Git 8.2 Attributs Git 8.3 Crochets Git 8.4 Exemple de politique gérée par Git 8.5 Résumé 9. Git et les autres systèmes 9.1 Git comme client 9.2 Migration vers Git 9.3 Résumé 10. Les tripes de Git 10.1 Plomberie et porcelaine 10.2 Les objets de Git 10.3 Références Git 10.4 Fichiers groupés 10.5 La refspec 10.6 Les protocoles de transfert 10.7 Maintenance et récupération de données 10.8 Les variables d’environnement 10.9 Résumé A1. Annexe A: Git dans d’autres environnements A1.1 Interfaces graphiques A1.2 Git dans Visual Studio A1.3 Git dans Visual Studio Code A1.4 Git dans IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git dans Sublime Text A1.6 Git dans Bash A1.7 Git dans Zsh A1.8 Git dans PowerShell A1.9 Résumé A2. Annexe B: Embarquer Git dans vos applications A2.1 Git en ligne de commande A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Commandes Git A3.1 Installation et configuration A3.2 Obtention et création des projets A3.3 Capture d’instantané basique A3.4 Création de branches et fusion A3.5 Partage et mise à jour de projets A3.6 Inspection et comparaison A3.7 Débogage A3.8 Patchs A3.9 Courriel A3.10 Systèmes externes A3.11 Administration A3.12 Commandes de plomberie 2nd Edition 2.1 Les bases de Git - Démarrer un dépôt Git Si vous ne deviez lire qu’un chapitre avant de commencer à utiliser Git, c’est celui-ci. Ce chapitre couvre les commandes de base nécessaires pour réaliser la vaste majorité des activités avec Git. À la fin de ce chapitre, vous devriez être capable de configurer et initialiser un dépôt, commencer et arrêter le suivi de version de fichiers, d’indexer et valider des modifications. Nous vous montrerons aussi comment paramétrer Git pour qu’il ignore certains fichiers ou patrons de fichiers, comment revenir sur les erreurs rapidement et facilement, comment parcourir l’historique de votre projet et voir les modifications entre deux validations, et comment pousser et tirer les modifications avec des dépôts distants. Démarrer un dépôt Git Vous pouvez principalement démarrer un dépôt Git de deux manières. Vous pouvez prendre un répertoire existant et le transformer en dépôt Git. Vous pouvez cloner un dépôt Git existant sur un autre serveur. Dans les deux cas, vous vous retrouvez avec un dépôt Git sur votre machine locale, prêt pour y travailler. Initialisation d’un dépôt Git dans un répertoire existant Si vous commencez à suivre dans Git un projet existant qui n’est pas suivi en gestion de version, vous n’avez qu’à vous positionner dans le répertoire du projet. Si vous ne l’avez jamais fait, cela se présente de différentes manières selon votre système d’exploitation : pour Linux: $ cd /home/user/my_project pour macOS: $ cd /Users/user/my_project pour Windows: $ cd C:/Users/user/my_project et entrez : $ git init Cela crée un nouveau sous-répertoire nommé .git qui contient tous les fichiers nécessaires au dépôt — un squelette de dépôt Git. Pour l’instant, aucun fichier n’est encore versionné. (Cf. Les tripes de Git pour plus d’information sur les fichiers contenus dans le répertoire .git que vous venez de créer.) Si vous souhaitez démarrer le contrôle de version sur des fichiers existants (par opposition à un répertoire vide), vous devrez probablement suivre ces fichiers et faire un commit initial. Vous pouvez le réaliser avec quelques commandes add qui spécifient les fichiers que vous souhaitez suivre, suivies par un git commit  : $ git add *.c $ git add LICENSE $ git commit -m 'initial project version' Nous allons détailler ce que ces commandes font dans quelques instants. À présent, vous avez un dépôt Git avec des fichiers suivis et un commit initial. Cloner un dépôt existant Si vous souhaitez obtenir une copie d’un dépôt Git existant — par exemple, un projet auquel vous aimeriez contribuer — la commande dont vous avez besoin s’appelle git clone . Si vous êtes familier avec d’autres systèmes de gestion de version tels que Subversion, vous noterez que la commande est clone et non checkout . C’est une distinction importante — Git reçoit une copie de quasiment toutes les données dont le serveur dispose. Toutes les versions de tous les fichiers pour l’historique du projet sont téléchargées quand vous lancez git clone . En fait, si le disque du serveur se corrompt, vous pouvez utiliser n’importe quel clone pour remettre le serveur dans l’état où il était au moment du clonage (vous pourriez perdre quelques paramètres du serveur, mais toutes les données sous gestion de version seraient récupérées — cf. Installation de Git sur un serveur pour de plus amples détails). Vous clonez un dépôt avec git clone [url] . Par exemple, si vous voulez cloner la bibliothèque logicielle Git appelée libgit2, vous pouvez le faire de la manière suivante : $ git clone https://github.com/libgit2/libgit2 Ceci crée un répertoire nommé libgit2 , initialise un répertoire .git à l’intérieur, récupère toutes les données de ce dépôt, et extrait une copie de travail de la dernière version. Si vous examinez le nouveau répertoire libgit2 , vous y verrez les fichiers du projet, prêts à être modifiés ou utilisés. Si vous souhaitez cloner le dépôt dans un répertoire nommé différemment, vous pouvez spécifier le nom dans une option supplémentaire de la ligne de commande : $ git clone https://github.com/libgit2/libgit2 monlibgit2 Cette commande réalise la même chose que la précédente, mais le répertoire cible s’appelle monlibgit2 . Git dispose de différents protocoles de transfert que vous pouvez utiliser. L’exemple précédent utilise le protocole https:// , mais vous pouvez aussi voir git:// ou utilisateur@serveur:/chemin.git , qui utilise le protocole de transfert SSH. Installation de Git sur un serveur introduit toutes les options disponibles pour mettre en place un serveur Git, ainsi que leurs avantages et inconvénients. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/workbook/how-sre-relates/
Google SRE - SRE vs DevOps, Similarity and Difference Chapter 1 - How SRE Relates to DevOps Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon How SRE Relates to DevOps class SRE implements interface DevOps By Niall Richard Murphy, Liz Fong-Jones, and Betsy Beyer, with Todd Underwood, Laura Nolan, and Dave Rensin Operations, as a discipline, is hard . 1 Not only is there the generally unsolved question of how to run systems well, but the best practices that have been found to work are highly context-dependent and far from widely adopted. There is also the largely unaddressed question of how to run operations teams well. Detailed analysis of these topics is generally thought to originate with Operational Research devoted to improving processes and output in the Allied military during World War II, but in reality, we have been thinking about how to operate things better for millennia . Yet, despite all this effort and thought, reliable production operations remains elusive—particularly in the domains of information technology and software operability . The enterprise world, for example, often treats operations as a cost center, 2 which makes meaningful improvements in outcomes difficult if not impossible. The tremendous short-sightedness of this approach is not yet widely understood, but dissatisfaction with it has given rise to a revolution in how to organize what we do in IT. That revolution stemmed from trying to solve a common set of problems. The newest solutions to these problems are called by two separate names—DevOps and Site Reliability Engineering (SRE). Although we talk about them individually as if they are totally separate reactions to the enterprise mentality just described, 3 we hope to persuade you that in fact they are much more alike, and practitioners of each have much more in common, than you might assume. But first, some background on the key tenets of each. Background on DevOps DevOps is a loose set of practices, guidelines, and culture designed to break down silos in IT development, operations, networking, and security. Articulated by John Willis, Damon Edwards, and Jez Humble, CA(L)MS —which stands for Culture, Automation, Lean (as in Lean management ; also see continuous delivery ), Measurement, and Sharing—is a useful acronym for remembering the key points of DevOps philosophy. Sharing and collaboration are at the forefront of this movement. In a DevOps approach, you improve something (often by automating it), measure the results, and share those results with colleagues so the whole organization can improve. All of the CALMS principles are facilitated by a supportive culture. DevOps, Agile, and a variety of other business and software reengineering techniques are all examples of a general worldview on how best to do business in the modern world. None of the elements in the DevOps philosophy are easily separable from each other, and this is essentially by design. There are, however, a few key ideas that can be discussed in relative isolation. No More Silos The first key idea is no more silos . This is a reaction to a couple ideas: The historically popular but now increasingly old-fashioned arrangement of separate operations and development teams The fact that extreme siloization of knowledge , incentives for purely local optimization, and lack of collaboration have in many cases been actively bad for business 4 Accidents Are Normal The second key idea is that accidents are not just a result of the isolated actions of an individual, but rather result from missing safeguards for when things inevitably go wrong. 5 For example, a bad interface inadvertently encourages the wrong action under pressure; a system misfeature makes failure inevitable if the (unarticulated) wrong circumstances occur; broken monitoring makes it impossible to know if something is wrong, never mind what is wrong. Some more traditionally minded businesses possess the cultural instinct to root out the mistake maker and punish them. But doing so has its own consequences: most obviously, it creates incentives to confuse issues, hide the truth, and blame others, all of which are ultimately unprofitable distractions. Therefore, it is more profitable to focus on speeding recovery than preventing accidents. Change Should Be Gradual The third key idea is that change is best when it is small and frequent . In environments where change committees meet monthly to discuss thoroughly documented plans to make changes to the mainframe configuration, this is a radical idea. However, this is not a new idea. The notion that all changes must be considered by experienced humans and batched for efficient consideration turns out to be more or less the opposite of best practice. Change is risky, true, but the correct response is to split up your changes into smaller subcomponents where possible. Then you build a steady pipeline of low-risk change out of regular output from product, design, and infrastructure changes. 6 This strategy, coupled with automatic testing of smaller changes and reliable rollback of bad changes, leads to approaches to change management like continuous integration (CI) and continuous delivery or deployment (CD) . Tooling and Culture Are Interrelated Tooling is an important component of DevOps, particularly given the emphasis on managing change correctly—today, change management relies on highly specific tools. Overall, however, proponents of DevOps strongly emphasize organizational culture—rather than tooling—as the key to success in adopting a new way of working. A good culture can work around broken tooling, but the opposite rarely holds true. As the saying goes, culture eats strategy for breakfast . Like operations, change itself is hard. Measurement Is Crucial Finally, measurement is particularly crucial in the overall business context of, for example, breaking down silos and incident resolution. In each of these environments, you establish the reality of what’s happening by means of objective measurement, verify that you’re changing the situation as you expect, and create an objective foundation for conversations that different functions agree upon. (This applies in both business and other contexts, such as on-call.) Background on SRE Site Reliability Engineering (SRE) is a term (and associated job role) coined by Ben Treynor Sloss, a VP of engineering at Google. 7 As we can see in the previous section, DevOps is a broad set of principles about whole-lifecycle collaboration between operations and product development. SRE is a job role, a set of practices (described next) we’ve found to work, and some beliefs that animate those practices. If you think of DevOps as a philosophy and an approach to working, you can argue that SRE implements some of the philosophy that DevOps describes, and is somewhat closer to a concrete definition of a job or role than, say, “DevOps engineer.” 8 So, in a way, class SRE implements interface DevOps . Unlike the DevOps movement, which originated from collaborations between leaders and practitioners at multiple companies, SRE at Google inherited much of its culture from the surrounding company before the term SRE became widely popularized across the industry. Given that trajectory, the discipline as a whole currently does not foreground cultural change by default quite as much as DevOps. (That doesn’t imply anything about whether cultural change is necessary to do SRE in an arbitrary organization, of course.) SRE is defined by the following concrete principles. Operations Is a Software Problem The basic tenet of SRE is that doing operations well is a software problem. SRE should therefore use software engineering approaches to solve that problem. This is across a wide field of view, encompassing everything from process and business change to similarly complicated but more traditional software problems, such as rewriting a stack to eliminate single points of failure in business logic. Manage by Service Level Objectives (SLOs) SRE does not attempt to give everything 100% availability. As discussed in our first book, Site Reliability Engineering , this is the wrong target for a number of reasons. Instead, the product team and the SRE team select an appropriate availability target for the service and its user base, and the service is managed to that SLO. 9 Deciding on such a target requires strong collaboration from the business. SLOs have cultural implications as well: as collaborative decisions among stakeholders, SLO violations bring teams back to the drawing board, blamelessly. Work to Minimize Toil For SRE, any manual, structurally mandated operational task is abhorrent. (That doesn’t mean we don’t have any such operations: we have plenty of them. We just don’t like them.) We believe that if a machine can perform a desired operation, then a machine often should . This is a distinction (and a value) not often seen in other organizations, where toil is the job, and that’s what you’re paying a person to do. For SRE in the Google context, toil is not the job—it can’t be. Any time spent on operational tasks means time not spent on project work—and project work is how we make our services more reliable and scalable. Performing operational tasks does, however, by “the wisdom of production,” provide vital input into decisions. This work keeps us grounded by providing real-time feedback from a given system. Sources of toil need to be identifiable so you can minimize or eliminate them. However, if you find yourself in a position of operational underload, you may need to push new features and changes more often so that engineers remain familiar with the workings of the service you support. The Wisdom of Production A note on “the wisdom of production”: by this phrase, we mean the wisdom you get from something running in production—the messy details of how it actually behaves, and how software should actually be designed, rather than a whiteboarded view of a service isolated from the facts on the ground. All of the pages you get, the tickets the team gets, and so on, are a direct connection with reality that should inform better system design and behavior. Automate This Year’s Job Away The real work in this area is determining what to automate, under what conditions, and how to automate it. SRE as practiced in Google has a hard limit of how much time a team member can spend on toil, as opposed to engineering that produces lasting value: 50%. Many people think of this limit as a cap . In fact, it’s much more useful to think of it as a guarantee —an explicit statement, and enabling mechanism, for taking an engineering-based approach to problems rather than just toiling at them over and over. There is an unintuitive and interesting interaction between this benchmark and how it plays out when we think about automation and toil. Over time, an SRE team winds up automating all that it can for a service, leaving behind things that can’t be automated ( the Murphy-Beyer effect ). Other things being equal, this comes to dominate what an SRE team does unless other actions are taken. In the Google environment, you tend to either add more services, up to some limit that still supports 50% engineering time, or you are so successful at your automation that you can go and do something else completely different instead. Move Fast by Reducing the Cost of Failure One of the main benefits of SRE engagement is not necessarily increased reliability, although obviously that does happen; it is actually improved product development output. Why? Well, a reduced mean time to repair (MTTR) for common faults results in increased product developer velocity, as engineers don’t have to waste time and focus cleaning up after these issues. This follows from the well-known fact that the later in the product lifecycle a problem is discovered, the more expensive it is to fix . SREs are specifically charged with improving undesirably late problem discovery, yielding benefits for the company as a whole. Share Ownership with Developers Rigid boundaries between “application development” and “production” (sometimes called programmers and operators) are counterproductive. This is especially true if the segregation of responsibilities and classification of ops as a cost center leads to power imbalances or discrepancies in esteem or pay. SREs tend to be inclined to focus on production problems rather than business logic problems, but as their approach brings software engineering tools to bear on the problem, they share skill sets with product development teams. In general, an SRE has particular expertise around the availability, latency, performance, efficiency, change management, monitoring, emergency response, and capacity planning of the service(s) they are looking after. Those specific (and usually well-defined) competencies are the bread-and-butter of what SRE does for a product and for the associated product development team. 10 Ideally, both product development and SRE teams should have a holistic view of the stack—the frontend, backend, libraries, storage, kernels, and physical machine—and no team should jealously own single components. It turns out that you can get a lot more done if you “blur the lines” 11 and have SREs instrument JavaScript, or product developers qualify kernels: knowledge of how to make changes and the authority to do so are much more widespread, and incentives to jealously guard any particular function are removed. In Site Reliability Engineering , we did not make it sufficiently clear that product development teams in Google own their service by default. SRE is neither available nor warranted for the bulk of services, although SRE principles still inform how services are managed throughout Google. 12 The ownership model when an SRE team works with a product development team is ultimately a shared model as well. Use the Same Tooling, Regardless of Function or Job Title Tooling is an incredibly important determinant of behavior, and it would be naive to assume that the efficacy of SRE in the Google context has nothing to do with the widely accessible unified codebase , the wide array of software and systems tooling, the highly optimized and proprietary production stack, and so on. Yet we share this absolute assumption with DevOps: teams minding a service 13 should use the same tools, regardless of their role in the organization. There is no good way to manage a service that has one tool for the SREs and another for the product developers, behaving differently (and potentially catastrophically so) in different situations. The more divergence you have, the less your company benefits from each effort to improve each individual tool. Compare and Contrast Looking at the preceding principles, we immediately see quite a lot of commonality in the points outlined: DevOps and SRE are both contingent on an acceptance that change is necessary in order to improve. Without that, there’s not much room for maneuvering. 14 Collaboration is front and center for DevOps work. An effective shared ownership model and partner team relationships are necessary for SRE to function. Like DevOps, SRE also has strong values shared across the organization, which can make climbing out of team-based silos slightly easier. Change management is best pursued as small, continual actions, the majority of which are ideally both automatically tested and applied. The critical interaction between change and reliability makes this especially important for SRE. The right tooling is critically important, and tooling to a certain extent determines the scope of your acts. Yet we must not focus too hard on whether something is achieved using some specific set of tools; at the end of the day, API orientation for system management is a more important philosophy that will outlast any particular implementation of it. Measurement is absolutely key to how both DevOps and SRE work. For SRE, SLOs are dominant in determining the actions taken to improve the service. Of course, you can’t have SLOs without measurement (as well as cross-team debate—ideally among product, infrastructure/SRE, and the business). For DevOps, the act of measurement is often used to understand what the outputs of a process are, what the duration of feedback loops is, and so on. Both DevOps and SRE are data-oriented things, whether they are professions or philosophies. The brute reality of managing production services means that bad things happen occasionally, and you have to talk about why. SRE and DevOps both practice blameless postmortems in order to offset unhelpful, adrenaline-laden reactions. Ultimately, implementing DevOps or SRE is a holistic act; both hope to make the whole of the team (or unit, or organization) better, as a function of working together in a highly specific way. For both DevOps and SRE, better velocity should be the outcome. 15 As you can see, there are many areas of commonality between DevOps and SRE. Yet there are significant differences as well. DevOps is in some sense a wider philosophy and culture. Because it effects wider change than does SRE, DevOps is more context-sensitive. DevOps is relatively silent on how to run operations at a detailed level. For example, it is not prescriptive around the precise management of services. It chooses instead to concentrate on breaking down barriers in the wider organization. This has much value. SRE, on the other hand, has relatively narrowly defined responsibilities and its remit is generally service-oriented (and end-user-oriented) rather than whole-business-oriented. As a result, it brings an opinionated intellectual framework (including concepts like error budgets ) to the problem of how to run systems effectively. Although SRE is, as a profession, highly aware of incentives and their effects, it in turn is relatively silent on topics like siloization and information barriers. It would support CI and CD not necessarily because of the business case, but because of the improved operational practices involved. Or, to put it another way, SRE believes in the same things as DevOps but for slightly different reasons . Organizational Context and Fostering Successful Adoption DevOps and SRE have a very large conceptual overlap in how they operate. As you might expect, they also have a similar set of conditions that have to be true within the organization in order for them to a) be implementable in the first place, and b) obtain the maximum benefit from that implementation. As Tolstoy almost but never quite said , effective operations approaches are all alike, whereas broken approaches are all broken in their own way. Incentives can in part explain why that is. If an organization’s culture values the benefits of a DevOps approach and is willing to bear those costs—typically expressed as difficulties in hiring, the energy required to maintain fluidity in teams and responsibilities, and increased financial resources dedicated to compensating a skill set that is necessarily more rare—then that organization must also make sure the incentives are correct in order to achieve the full benefit of this approach. Specifically, the following should hold true in the context of both DevOps and SRE. Narrow, Rigid Incentives Narrow Your Success Many companies accidentally define formal incentives that undermine collective performance. To avoid this mistake, don’t structure incentives to be narrowly tied to launch-related or reliability-related outcomes. As any economist can tell you, if there is a numeric measure, people will find a way to game it to bad effect, sometimes even in a completely well-intentioned way. 16 Instead, you should allow your people the freedom to find the right tradeoffs. As discussed earlier, DevOps or SRE can act as an accelerant for your product team in general, allowing the rest of the software org to ship features to customers in a continuous and reliable fashion. Such a dynamic also fixes one persistent problem with the traditional and divergent systems/software group approach: the lack of a feedback loop between design and production. A system with early SRE engagement (ideally, at design time) typically works better in production after deployment, regardless of who is responsible for managing the service. (Nothing slows down feature development like losing user data.) It’s Better to Fix It Yourself; Don’t Blame Someone Else Furthermore, avoid any incentives to pass off the blame for production incidents or system failures onto other groups. In many ways, the dynamics of passing off blame is the core problem with the traditional model for engineering operations, as separating operations and software teams allows separate incentives to emerge. Instead, consider adopting the following practices to combat blame passing at an organizational level: Don’t just allow, but actively encourage , engineers to change code and configuration when required for the product. Also allow these teams the authority to be radical within the limits of their mission, thereby eliminating incentives to proceed more slowly. Support blameless postmortems. 17 Doing so eliminates incentives to downplay or cover up a problem. This step is crucial in fully understanding the product and actually optimizing its performance and functionality, and relies on the wisdom of production mentioned previously. Allow support to move away from products that are irredeemably operationally difficult. The threat of support withdrawal motivates product development to fix issues both in the run-up to support and once the product is itself supported, saving everyone time. What it means to be “irredeemably operationally difficult” may differ depending on your context—the dynamic here should be one of mutually understood responsibilities. Pushback to other orgs might be a softer, “We think there are higher-value uses of the time of people with this skill set,” or framed within the limit of, “These people will quit if they’re tasked with too much operational work and aren’t given the opportunity to use their engineering skill set.” At Google, the practice of outright withdrawing support from such products has become institutional. Consider Reliability Work as a Specialized Role At Google, SRE and product development are separate organizations. Each group has its own focus, priorities, and management, and does not have to do the bidding of the other. However, the product development teams effectively fund the growth of SRE with new hires when a product is successful. In this way, product development has a stake in the success of SRE teams, just as SREs have a stake in the success of the product development teams. SRE is also fortunate to receive high-level support from management, which ensures that engineering teams’ objections to supporting services “the SRE way” are generally short-lived. You don’t need to have an org chart to do things differently, though—you just need a different community of practice to emerge. Regardless of whether you fork your organizational chart or use more informal mechanisms, it’s important to recognize that specialization creates challenges. Practitioners of DevOps and SRE benefit from having a community of peers for support and career development, and job ladders that reward them 18 for the unique skills and perspectives they bring to the table. It’s important to note that the organizational structure employed by Google, as well as some of the aforementioned incentives, is somewhat reliant on a sizeable organization. For example, if your 20-person startup has only one (comparatively small) product, there’s not much sense in allowing withdrawal of operational support. It’s still possible to take a DevOps-style approach, 19 but the ability to improve an operationally poor product is undermined if literally all you can do is help it grow. Usually, though, people have more choice than they imagine about how to fulfill those growth needs versus the rate at which technical debt accumulates. 20 When Can Substitute for Whether However, when your organization or product grows beyond a certain size, you can exercise more latitude in what products to support, or how to prioritize that support. If it’s clear that support for system X is going to happen much sooner than support for system Y, the implicit conditionality can play much the same role as the choice to not support services in the SRE world. At Google, SRE’s strong partnership with product development has proven to be critically important: if such a relationship exists at your organization, then the decision to withdraw (or supply) support can be based on objective data about comparative operational characteristics, thereby avoiding otherwise nonproductive conversations. A productive relationship between SRE and product development also helps in avoiding the organizational anti-pattern in which a product development team has to ship a product or feature before it’s quite ready. Instead, SRE can work with a development team to improve the product before the burden of maintenance shifts away from the people with the most expertise to fix it. Strive for Parity of Esteem: Career and Financial Finally, make sure that the career incentives to do the right thing are in place: we want our DevOps/SRE organization to be held in the same esteem as their product development counterparts. Therefore, members of each team should be rated by roughly the same methods and have the same financial incentives. Conclusion In many ways, DevOps and SRE sit, in both practice and philosophy, very close to each other in the overall landscape of IT operations. Both DevOps and SRE require discussion, management support, and buy-in from the people actually doing the work to make serious progress. Implementing either of them is a journey and not a quick fix: the practice of rename-and-shame 21 is a hollow one, unlikely to yield benefit. Given that it is a more opinionated implementation of how to perform operations, SRE has more concrete suggestions on how to change your work practices earlier on in that journey, albeit requiring specific adaptation. DevOps, having a wider focus, is somewhat more difficult to reason about and translate into concrete steps, but precisely because of that wider focus, is likely to meet with weaker initial resistance. But practitioners of each use many of the same tools, the same approaches to change management, and the same data-based decision-making mindset. At the end of the day, we all face the same persistent problem: production, and making it better—no matter what we’re called. For those interested in further reading, the following suggestions should help you develop a wider understanding of the cultural, business, and technical underpinnings of the operations revolution taking place right now: Site Reliability Engineering Effective DevOps The Phoenix Project The Practice of Cloud System Administration: DevOps and SRE Practices for Web Services, Volume 2 Accelerate: The Science of Lean Software and DevOps 1 Note that as this discussion appears in a book about SRE, some of this discussion is specific to software service operations, as opposed to IT operations. 2 Mary Poppendieck has an excellent article on this called “ The Cost Center Trap .” Another way in which this approach fails is when a very large and improbable disaster completely wipes out the cost savings you made by moving to a low-grade operations model (c.f. the British Airways outage in May 2017 ). 3 Of course, there are a number of other potential reactions. For example, ITIL® is another approach to IT management that advocates for better standardization. 4 Note also that because this is a complicated world, there are also positive effects to partitioning , silos, and the like, but the downsides seem to be particularly pernicious in the domain of operations. 5 See https://en.wikipedia.org/wiki/Normal_Accidents . 6 Higher-risk changes, or those unvalidatable by automatic means, should obviously still be vetted by humans, if not enacted by them. 7 The history of SRE at Google is that it sprang from a precursor team, which was more operationally focused, and Ben provided the impetus for treating the problem from an engineering standpoint. 8 This is a misnomer in a large number of ways, perhaps the most fundamental being that you can’t just hire some people, call them “DevOps engineers,” and expect benefits immediately. You have to buy into the whole philosophy of changing how you work in order to benefit. As Andrew Clay Shafer says , “People sell DevOps, but you can’t buy it.” And, as Seth Vargo points out in “The 10 Myths of DevOps” , you can’t “hire a DevOp to fix your organization.” 9 A service level objective is a target for performance of a particular metric (e.g., available 99.9% of the time). 10 Of course, not every team does everything, but those are the most common headings under which SRE works. 11 Perform a layering violation, if you think of this as layered stacks. 12 In fact, there’s a Production Readiness Review for onboarding anything ; SRE won’t just onboard services from a standing start. 13 A service is loosely defined as software running to perform some business need, generally with availability constraints. 14 Within Google, that question is largely settled, and services change state, configuration, ownership, direction, and so on, all the time. To a certain extent, SRE at Google is the beneficiary of the “change is necessary” argument having been fought and won a number of times in the past. But not completely evenly distributed , as William Gibson might say. 15 See relevant research at https://devops-research.com/research.html . 16 See https://en.wikipedia.org/wiki/Goodhart%27s_law and https://skybrary.aero/bookshelf/books/2336.pdf . 17 See, for example, https://codeascraft.com/2012/05/22/blameless-postmortems/ . 18 In orgs that have well-developed cultures of either. Early-stage companies likely do not have established ways to reward these job roles. 19 Indeed, arguably, that’s your only choice unless you outsource operations. 20 For a discussion of how to apply SRE principles in different contexts, see SRE Team Lifecycles . 21 In other words, simply retitling a group DevOps or SRE with no other change in their organizational positioning, resulting in inevitable shaming of the team when promised improvement is not forthcoming. Previous Preface Next Part I - Foundations Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://git-scm.com/book/fa/v2/%d8%b4%d8%b1%d9%88%d8%b9-%d8%a8%d9%87-%da%a9%d8%a7%d8%b1-getting-started-%d8%b3%d8%aa%d8%a7%d9%be-%d8%a7%d9%88%d9%84%db%8c%d9%87-%da%af%db%8c%d8%aa-First-Time-Git-Setup
Git - ستاپ اولیه گیت (First-Time Git Setup) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. شروع به کار (getting started) 1.1 درباره ورژن کنترل (About Version Control) 1.2 تاریخچه کوتاهی از گیت (A Short History of Git) 1.3 گیت چیست؟ (What is Git) 1.4 نصب گیت (Installing Git) 1.5 ستاپ اولیه گیت (First-Time Git Setup) 1.6 دریافت کمک (Getting Help) 1.7 خلاصه (summary) 2. مقدمات گیت (git basics chapter) 2.1 گرفتن یک مخزن گیت (Getting a Git Repository) 2.2 ثبت تغییرات در مخزن (Recording Changes to the Repository) 2.3 مشاهده تاریخچه کامیت‌ها (Viewing the Commit History) 2.4 بازگرداندن تغییرات (Undoing Things) 2.5 کار کردن با ریموت ها (Working with Remotes) 2.6 تگ کردن (Tagging) 2.7 نام مستعار گیت (Git Aliases) 2.8 خلاصه (summary) 3. انشعاب‌گیری در گیت (Git Branching) 3.1 شاخه‌ها در یک نگاه (Branches in a Nutshell) 3.2 شاخه‌بندی و ادغام پایه‌ای (Basic Branching and Merging) 3.3 مدیریت شاخه‌ها (Branch Management) 3.4 روندهای کاری شاخه‌ها (Branching Workflows) 3.5 شاخه‌های راه دور (Remote Branches) 3.6 بازپایه‌گذاری (Rebasing) 3.7 خلاصه (Summary) 4. گیت روی سرور (Git on the server) 4.1 پروتکل‌ها (The Protocols) 4.2 راه‌اندازی گیت روی یک سرور (Getting Git on a Server) 4.3 ایجاد کلید عمومی SSH شما (Generating Your SSH Public Key) 4.4 نصب و راه‌اندازی سرور (Setting up server) 4.5 سرویس‌دهنده گیت (Git Daemon) 4.6 HTTP هوشمند (Smart HTTP) 4.7 گیت‌وب (GitWeb) 4.8 گیت‌لب (GitLab) 4.9 گزینه‌های میزبانی شخص ثالث (Third Party Hosted Options) 4.10 خلاصه (Summary) 5. گیت توزیع‌شده (Distributed git) 5.1 جریان‌های کاری توزیع‌شده (Distributed Workflows) 5.2 مشارکت در یک پروژه (Contributing to a Project) 5.3 نگهداری یک پروژه (Maintaining a Project) 5.4 خلاصه (Summary) 6. گیت هاب (GitHub) 6.1 راه‌اندازی و پیکربندی حساب کاربری (Account Setup and Configuration) 6.2 مشارکت در یک پروژه (Contributing to a Project) 6.3 نگهداری یک پروژه (Maintaining a Project) 6.4 مدیریت یک سازمان (Managing an organization) 6.5 اسکریپتنویسی در گیتهاب (Scripting GitHub) 6.6 خلاصه (Summary) 7. ابزارهای گیت (Git Tools) 7.1 انتخاب بازبینی (Revision Selection) 7.2 مرحله‌بندی تعاملی (Interactive Staging) 7.3 ذخیره موقت و پاک‌سازی (Stashing and Cleaning) 7.4 امضای کارهای شما (Signing Your Work) 7.5 جستجو (Searching) 7.6 بازنویسی تاریخچه (Rewriting History) 7.7 بازنشانی به زبان ساده (Reset Demystified) 7.8 ادغام پیشرفته (Advanced Merging) 7.9 بازاستفاده خودکار از حل تضادها (Rerere) 7.10 اشکال‌زدایی با گیت (Debugging with Git) 7.11 سابماژول ها (Submodules) 7.12 بسته‌بندی (Bundling) 7.13 جایگزینی (Replace) 7.14 ذخیره‌سازی اطلاعات ورود (Credential Storage) 7.15 خلاصه (Summary) 8. سفارشی‌سازی Git (Customizing Git) 8.1 پیکربندی گیت (Git Configuration) 8.2 ویژگی‌های گیت (Git Attributes) 8.3 هوک‌های گیت (Git Hooks) 8.4 یک نمونه سیاست اعمال شده توسط گیت (An Example Git-Enforced Policy) 8.5 خلاصه (Summary) 9. گیت و سیستم‌های دیگر (Git and Other Systems) 9.1 گیت به‌عنوان کلاینت (Git as a Client) 9.2 مهاجرت به گیت (Migrating to Git) 9.3 خلاصه (Summary) 10. مباحث درونی گیت (Git Internals) 10.1 ابزارها و دستورات سطح پایین (Plumbing and Porcelain) 10.2 اشیا گیت (Git Objects) 10.3 مراجع گیت (Git References) 10.4 فایل‌های بسته (Packfiles) 10.5 نگاشت (The Refspec) 10.6 پروتکل‌های انتقال (Transfer Protocols) 10.7 نگهداری و بازیابی داده‌ها (Maintenance and Data Recovery) 10.8 متغیرهای محیطی (Environment Variables) 10.9 خلاصه (Summary) A1. پیوست A: گیت در محیط‌های دیگر (Git in Other Environments) A1.1 رابط های گرافیکی (Graphical Interfaces) A1.2 گیت در ویژوال استودیو (Git in Visual Studio) A1.3 گیت در Visual Studio Code (Git in Visual Studio Code) A1.4 گیت در IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine (Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine) A1.5 گیت در Sublime Text (Git in Sublime Text) A1.6 گیت در بش (Git in Bash) A1.7 گیت در Zsh (Git in Zsh) A1.8 گیت در PowerShell (Git in PowerShell) A1.9 خلاصه (Summary) A2. پیوست B: گنجاندن گیت در برنامه‌های شما (Embedding Git in your Applications) A2.1 خط فرمان گیت (Command-line Git) A2.2 کتابخانهٔ گیت به زبان سی (Libgit2) A2.3 کتابخانه گیت برای زبان جاوا (JGit) A2.4 کتابخانه گیت برای زبان گو (go-git) A2.5 کتابخانه گیت پایتون (Dulwich) A3. پیوست C: دستورات گیت (Git Commands) A3.1 تنظیم و پیکربندی (Setup and Config) A3.2 گرفتن و ایجاد پروژه‌ها (Getting and Creating Projects) A3.3 نمونه‌برداری پایه‌ای (Basic Snapshotting) A3.4 انشعاب‌گیری و ادغام (Branching and Merging) A3.5 به‌اشتراک‌گذاری و به‌روزرسانی پروژه‌ها (Sharing and Updating Projects) A3.6 بازرسی و مقایسه (Inspection and Comparison) A3.7 عیب‌یابی (Debugging) A3.8 اعمال تغییرات به صورت پچ (Patching) A3.9 ایمیل (Email) A3.10 سیستم‌های خارجی (External Systems) A3.11 مدیریت (Administration) A3.12 دستورات سطح پایین گیت (Plumbing Commands) 2nd Edition 1.5 شروع به کار (getting started) - ستاپ اولیه گیت (First-Time Git Setup) ستاپ اولیه گیت (First-Time Git Setup) حالا که Git را روی سیستم خود نصب کرده‌اید، می‌خواهید چند کار برای شخصی‌سازی محیط Git خود انجام دهید. این تنظیمات را فقط یک‌بار روی هر کامپیوتر باید انجام دهید؛ و این تنظیمات بین آپدیت‌ها حفظ می‌شوند. همچنین می‌توانید در هر زمانی با اجرای مجدد دستورات، این تنظیمات را تغییر دهید. Git همراه با ابزاری به نام “git config” ارائه می‌شود که به شما امکان می‌دهد متغیرهای پیکربندی را دریافت و تنظیم کنید؛ این متغیرها کنترل می‌کنند که Git چگونه ظاهر شود و چگونه کار کند. این متغیرها می‌توانند در سه جای مختلف ذخیره شوند: فایل [path]/etc/gitconfig : شامل مقادیری است که برای همه‌ی کاربران سیستم و تمام مخازن آن‌ها اعمال می‌شود. اگر گزینه‌ی --system را به "git config" بدهید، این فایل را به‌طور خاص می‌خواند و در آن می‌نویسد. از آنجا که این یک فایل پیکربندی سیستم است، برای اعمال تغییرات به دسترسی مدیریتی یا دسترسی سوپریوزر نیاز دارید. فایل ~/.gitconfig یا ~/.config/git/config : شامل مقادیری است که به‌صورت شخصی و مخصوص شما، کاربر، است. می‌توانید با استفاده از گزینه‌ی --global به "git config" بگویید که فقط این فایل را بخواند و در آن بنویسد، و این تنظیمات روی تمام مخازنی که در سیستم خود با آن‌ها کار می‌کنید تأثیر می‌گذارد. فایل config در پوشه ی گیت (that is, .git/config ) مربوط به هر مخزنی که در حال حاضر استفاده می‌کنید: مخصوص همان مخزن است. می‌توانید با گزینه‌ی --local به Git دستور دهید که فقط از این فایل بخواند و در آن بنویسد، اما این گزینه در واقع حالت پیش‌فرض است. بدیهی است که برای استفاده‌ی درست از این گزینه باید در داخل یک مخزن Git باشید. هر سطح، مقادیر سطح قبلی را نادیده می‌گیرد، بنابراین مقادیر موجود در فایل ".git/config" اولویت بیشتری نسبت به مقادیر در فایل "[path]/etc/gitconfig" دارند. در سیستم‌های ویندوز، Git به دنبال فایل ".gitconfig" در دایرکتوری "$HOME" می‌گردد (که برای اکثر کاربران معمولاً "C:\Users\$USER" است). همچنین همچنان به دنبال "[path]/etc/gitconfig" است، هرچند این مسیر نسبت به ریشه‌ی MSys است؛ یعنی هر جایی که هنگام نصب Git روی ویندوز خود آن را نصب کرده‌اید. اگر از نسخه‌ی ۲.x یا جدیدتر Git برای ویندوز استفاده می‌کنید، یک فایل پیکربندی در سطح سیستم نیز وجود دارد: در "C:\Documents and Settings\All Users\Application Data\Git\config" برای ویندوز XP، و در "C:\ProgramData\Git\config" برای ویندوز ویستا و نسخه‌های جدیدتر. این فایل پیکربندی تنها با اجرای دستور "git config -f <file>" به صورت مدیر (admin) قابل تغییر است. می‌توانید همه‌ی تنظیمات خود و منبع آن‌ها را با استفاده از دستور زیر مشاهده کنید: $ git config --list --show-origin هویت شما (Your Identity) اولین کاری که بعد از نصب Git باید انجام دهید، تنظیم نام کاربری و آدرس ایمیل خود است. این موضوع اهمیت زیادی دارد چون هر commit در Git از این اطلاعات استفاده می‌کند و این اطلاعات به‌صورت غیرقابل تغییر در کامیت‌هایی که ایجاد می‌کنید ثبت می‌شوند: $ git config --global user.name "John Doe" $ git config --global user.email johndoe@example.com دوباره تأکید می‌کنیم که اگر گزینه‌ی --global را استفاده کنید، فقط یک‌بار باید این کار را انجام دهید، زیرا در این صورت Git همیشه از آن اطلاعات برای کاربر شما روی آن سیستم استفاده خواهد کرد. اگر بخواهید برای پروژه‌های خاص نام یا آدرس ایمیل متفاوتی تعیین کنید، می‌توانید دستور را بدون گزینه‌ی --global در پوشه‌ی آن پروژه اجرا کنید. بسیاری از ابزارهای رابط کاربری گرافیکی (GUI) هنگام اولین اجرا به شما در انجام این تنظیمات کمک می‌کنند. ویرایشگر شما (Your Editor) حالا که هویت خود را تنظیم کرده‌اید، می‌توانید ویرایشگر متنی پیش‌فرض را تنظیم کنید که هنگام نیاز Git به نوشتن پیام از آن استفاده شود. اگر این تنظیم انجام نشود، Git از ویرایشگر پیش‌فرض سیستم شما استفاده می‌کند. اگر می‌خواهید از ویرایشگر متنی متفاوتی مثل Emacs استفاده کنید، می‌توانید به این صورت عمل کنید: $ git config --global core.editor emacs در سیستم ویندوز، اگر بخواهید از ویرایشگر متنی متفاوتی استفاده کنید، باید مسیر کامل فایل اجرایی (executable) آن را مشخص کنید. این مسیر بسته به نحوه بسته‌بندی ویرایشگر ممکن است متفاوت باشد. در مورد Notepad++، یک ویرایشگر محبوب برنامه‌نویسی، معمولاً تمایل دارید نسخه‌ی ۳۲ بیتی آن را استفاده کنید، چون در زمان نگارش این متن، نسخه‌ی ۶۴ بیتی هنوز تمام پلاگین‌ها را پشتیبانی نمی‌کند. اگر روی سیستم ویندوز ۳۲ بیتی هستید، یا نسخه‌ی ۶۴ بیتی ویرایشگر را روی سیستم ۶۴ بیتی دارید، چیزی شبیه به این را تایپ خواهید کرد: $ git config --global core.editor "'C:/Program Files/Notepad++/notepad++.exe' -multiInst -notabbar -nosession -noPlugin" یادداشت ویرایشگرهای متنی محبوبی مانند Vim، Emacs و Notepad++ معمولاً توسط توسعه‌دهندگان روی سیستم‌های مبتنی بر یونیکس مانند لینوکس و مک‌اواس یا سیستم ویندوز استفاده می‌شوند. اگر از ویرایشگر دیگری استفاده می‌کنید، یا نسخه‌ی ۳۲ بیتی آن را دارید، لطفاً دستورالعمل‌های خاص تنظیم ویرایشگر مورد علاقه‌تان با Git را در دستورات ویرایشگر اصلی گیت (git config core.editor commands) بیابید. هشدار ممکن است اگر ویرایشگر خود را به این شکل تنظیم نکنید، هنگام تلاش Git برای باز کردن آن با وضعیت بسیار گیج‌کننده‌ای مواجه شوید. به‌عنوان مثال، در سیستم ویندوز ممکن است عملیات Git که برای ویرایش آغاز شده بود به‌صورت زودهنگام و ناگهانی متوقف شود. نام پیشفرض برنچ شما (Your default branch name) به‌طور پیش‌فرض، Git هنگام ایجاد یک مخزن جدید با دستور git init یک شاخه به نام main می‌سازد. از نسخه‌ی 2.28 به بعد، می‌توانید نام متفاوتی برای شاخه‌ی اولیه تنظیم کنید. برای تنظیم نام master به‌عنوان شاخه‌ی پیش‌فرض، این کار را انجام دهید: $ git config --global init.defaultBranch master بررسی تنظیمات شما (Checking Your Settings) اگر می‌خواهید تنظیمات پیکربندی خود را بررسی کنید، می‌توانید از دستور "git config --list" استفاده کنید تا تمام تنظیماتی که Git در آن لحظه می‌تواند پیدا کند را لیست کند: $ git config --list user.name=John Doe user.email=johndoe@example.com color.status=auto color.branch=auto color.interactive=auto color.diff=auto ... ممکن است بعضی کلیدها را بیشتر از یک بار ببینید، زیرا Git همان کلید را از فایل‌های مختلفی می‌خواند (مثلاً "[path]/etc/gitconfig" و "~/.gitconfig"). در این حالت، Git آخرین مقدار هر کلید یکتا را که می‌بیند استفاده می‌کند. همچنین می‌توانید با تایپ دستور "git config <key>" مقدار در نظر گرفته شده برای یک کلید خاص را مشاهده کنید: $ git config user.name John Doe یادداشت از آنجایی که Git ممکن است مقدار یک متغیر پیکربندی را از بیش از یک فایل بخواند، ممکن است با مقداری غیرمنتظره مواجه شوید و ندانید چرا. در چنین مواردی، می‌توانید از Git بپرسید که منشأ (origin) آن مقدار کجا بوده است، و Git به شما خواهد گفت که کدام فایل پیکربندی آخرین تصمیم را در تعیین آن مقدار گرفته است: $ git config --show-origin rerere.autoUpdate file:/home/johndoe/.gitconfig false prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/fr/v2/Les-branches-avec-Git-Rebaser-Rebasing
Git - Rebaser (Rebasing) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Démarrage rapide 1.1 À propos de la gestion de version 1.2 Une rapide histoire de Git 1.3 Rudiments de Git 1.4 La ligne de commande 1.5 Installation de Git 1.6 Paramétrage à la première utilisation de Git 1.7 Obtenir de l’aide 1.8 Résumé 2. Les bases de Git 2.1 Démarrer un dépôt Git 2.2 Enregistrer des modifications dans le dépôt 2.3 Visualiser l’historique des validations 2.4 Annuler des actions 2.5 Travailler avec des dépôts distants 2.6 Étiquetage 2.7 Les alias Git 2.8 Résumé 3. Les branches avec Git 3.1 Les branches en bref 3.2 Branches et fusions : les bases 3.3 Gestion des branches 3.4 Travailler avec les branches 3.5 Branches de suivi à distance 3.6 Rebaser (Rebasing) 3.7 Résumé 4. Git sur le serveur 4.1 Protocoles 4.2 Installation de Git sur un serveur 4.3 Génération des clés publiques SSH 4.4 Mise en place du serveur 4.5 Démon (Daemon) Git 4.6 HTTP intelligent 4.7 GitWeb 4.8 GitLab 4.9 Git hébergé 4.10 Résumé 5. Git distribué 5.1 Développements distribués 5.2 Contribution à un projet 5.3 Maintenance d’un projet 5.4 Résumé 6. GitHub 6.1 Configuration et paramétrage d’un compte 6.2 Contribution à un projet 6.3 Maintenance d’un projet 6.4 Gestion d’un regroupement 6.5 Écriture de scripts pour GitHub 6.6 Résumé 7. Utilitaires Git 7.1 Sélection des versions 7.2 Indexation interactive 7.3 Remisage et nettoyage 7.4 Signer votre travail 7.5 Recherche 7.6 Réécrire l’historique 7.7 Reset démystifié 7.8 Fusion avancée 7.9 Rerere 7.10 Déboguer avec Git 7.11 Sous-modules 7.12 Empaquetage (bundling) 7.13 Replace 7.14 Stockage des identifiants 7.15 Résumé 8. Personnalisation de Git 8.1 Configuration de Git 8.2 Attributs Git 8.3 Crochets Git 8.4 Exemple de politique gérée par Git 8.5 Résumé 9. Git et les autres systèmes 9.1 Git comme client 9.2 Migration vers Git 9.3 Résumé 10. Les tripes de Git 10.1 Plomberie et porcelaine 10.2 Les objets de Git 10.3 Références Git 10.4 Fichiers groupés 10.5 La refspec 10.6 Les protocoles de transfert 10.7 Maintenance et récupération de données 10.8 Les variables d’environnement 10.9 Résumé A1. Annexe A: Git dans d’autres environnements A1.1 Interfaces graphiques A1.2 Git dans Visual Studio A1.3 Git dans Visual Studio Code A1.4 Git dans IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git dans Sublime Text A1.6 Git dans Bash A1.7 Git dans Zsh A1.8 Git dans PowerShell A1.9 Résumé A2. Annexe B: Embarquer Git dans vos applications A2.1 Git en ligne de commande A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Commandes Git A3.1 Installation et configuration A3.2 Obtention et création des projets A3.3 Capture d’instantané basique A3.4 Création de branches et fusion A3.5 Partage et mise à jour de projets A3.6 Inspection et comparaison A3.7 Débogage A3.8 Patchs A3.9 Courriel A3.10 Systèmes externes A3.11 Administration A3.12 Commandes de plomberie 2nd Edition 3.6 Les branches avec Git - Rebaser (Rebasing) Rebaser ( Rebasing ) Dans Git, il y a deux façons d’intégrer les modifications d’une branche dans une autre : en fusionnant ( merge ) et en rebasant ( rebase ). Dans ce chapitre, vous apprendrez la signification de rebaser, comment le faire, pourquoi c’est un outil incroyable et dans quels cas il est déconseillé de l’utiliser. Les bases Si vous revenez à un exemple précédent du chapitre Fusions ( Merges ) , vous remarquerez que votre travail a divergé et que vous avez ajouté des commits sur deux branches différentes. Figure 35. Historique divergeant simple Comme nous l’avons déjà expliqué, le moyen le plus simple pour intégrer ces branches est la fusion via la commande merge . Cette commande réalise une fusion à trois branches entre les deux derniers instantanés ( snapshots ) de chaque branche (C3 et C4) et l’ancêtre commun le plus récent (C2), créant un nouvel instantané (et un commit ). Figure 36. Fusion pour intégrer des travaux aux historiques divergeant Cependant, il existe un autre moyen : vous pouvez prendre le patch de la modification introduite en C4 et le réappliquer sur C3 . Dans Git, cette action est appelée "rebaser" ( rebasing ). Avec la commande rebase , vous pouvez prendre toutes les modifications qui ont été validées sur une branche et les rejouer sur une autre. Dans cet exemple, vous lanceriez les commandes suivantes : $ git checkout experiment $ git rebase master First, rewinding head to replay your work on top of it... Applying: added staged command Cela fonctionne en cherchant l’ancêtre commun le plus récent des deux branches (celle sur laquelle vous vous trouvez et celle sur laquelle vous rebasez), en récupérant toutes les différences introduites par chaque commit de la branche courante, en les sauvant dans des fichiers temporaires, en réinitialisant la branche courante sur le même commit que la branche de destination et en appliquant finalement chaque modification dans le même ordre. Figure 37. Rebasage des modifications introduites par C4 sur C3 À ce moment, vous pouvez retourner sur la branche master et réaliser une fusion en avance rapide ( fast-forward merge ). $ git checkout master $ git merge experiment Figure 38. Avance rapide de la branche master À présent, l’instantané pointé par C4' est exactement le même que celui pointé par C5 dans l’exemple de fusion. Il n’y a pas de différence entre les résultats des deux types d’intégration, mais rebaser rend l’historique plus clair. Si vous examinez le journal de la branche rebasée, elle est devenue linéaire : toutes les modifications apparaissent en série même si elles ont eu lieu en parallèle. Vous aurez souvent à faire cela pour vous assurer que vos commits s’appliquent proprement sur une branche distante — par exemple, sur un projet où vous souhaitez contribuer mais que vous ne maintenez pas. Dans ce cas, vous réaliseriez votre travail dans une branche puis vous rebaseriez votre travail sur origin/master quand vous êtes prêt à soumettre vos patchs au projet principal. De cette manière, le mainteneur n’a pas à réaliser de travail d’intégration — juste une avance rapide ou simplement une application propre. Il faut noter que l’instantané pointé par le commit final, qu’il soit le dernier des commits d’une opération de rebasage ou le commit final issu d’une fusion, sont en fait le même instantané — c’est juste que l’historique est différent. Rebaser rejoue les modifications d’une ligne de commits sur une autre dans l’ordre d’apparition, alors que la fusion joint et fusionne les deux têtes. Rebases plus intéressants Vous pouvez aussi faire rejouer votre rebasage sur autre chose qu’une branche. Prenez un historique tel que Un historique avec deux branches thématiques qui sortent l’une de l’autre par exemple. Vous avez créé une branche thématique ( server ) pour ajouter des fonctionnalités côté serveur à votre projet et avez réalisé un commit . Ensuite, vous avez créé une branche pour ajouter des modifications côté client ( client ) et avez validé plusieurs fois. Finalement, vous avez rebasculé sur la branche server et avez réalisé quelques commits supplémentaires. Figure 39. Un historique avec deux branches thématiques qui sortent l’une de l’autre Supposons que vous décidez que vous souhaitez fusionner vos modifications du côté client dans votre ligne principale pour une publication ( release ) mais vous souhaitez retenir les modifications de la partie serveur jusqu’à ce qu’elles soient un peu mieux testées. Vous pouvez récupérer les modifications du côté client qui ne sont pas sur le serveur ( C8 et C9 ) et les rejouer sur la branche master en utilisant l’option --onto de git rebase  : $ git rebase --onto master server client Cela signifie en substance "Extraire la branche client, déterminer les patchs depuis l’ancêtre commun des branches client et server puis les rejouer sur master ". C’est assez complexe, mais le résultat est assez impressionnant. Figure 40. Rebaser deux branches thématiques l’une sur l’autre Maintenant, vous pouvez faire une avance rapide sur votre branche master (cf. Avance rapide sur votre branche master pour inclure les modifications de la branche client ): $ git checkout master $ git merge client Figure 41. Avance rapide sur votre branche master pour inclure les modifications de la branche client Supposons que vous décidiez de tirer ( pull ) votre branche server aussi. Vous pouvez rebaser la branche server sur la branche master sans avoir à l’extraire avant en utilisant git rebase [branchedebase] [branchethematique] — qui extrait la branche thématique (dans notre cas, server ) pour vous et la rejoue sur la branche de base ( master ) : $ git rebase master server Cette commande rejoue les modifications de server sur le sommet de la branche master , comme indiqué dans Rebasage de la branche server sur le sommet de la branche master . Figure 42. Rebasage de la branche server sur le sommet de la branche master Vous pouvez ensuite faire une avance rapide sur la branche de base ( master ) : $ git checkout master $ git merge server Vous pouvez effacer les branches client et server une fois que tout le travail est intégré et que vous n’en avez plus besoin, éliminant tout l’historique de ce processus, comme visible sur Historique final des commits  : $ git branch -d client $ git branch -d server Figure 43. Historique final des commits Les dangers du rebasage Ah… mais les joies de rebaser ne viennent pas sans leurs contreparties, qui peuvent être résumées en une ligne : Ne rebasez jamais des commits qui ont déjà été poussés sur un dépôt public. Si vous suivez ce conseil, tout ira bien. Sinon, de nombreuses personnes vont vous haïr et vous serez méprisé par vos amis et votre famille. Quand vous rebasez des données, vous abandonnez les commits existants et vous en créez de nouveaux qui sont similaires mais différents. Si vous poussez des commits quelque part, que d’autres les tirent et se basent dessus pour travailler, et qu’après coup, vous réécrivez ces commits à l’aide de git rebase et les poussez à nouveau, vos collaborateurs devront re-fusionner leur travail et les choses peuvent rapidement devenir très désordonnées quand vous essaierez de tirer leur travail dans votre dépôt. Examinons un exemple expliquant comment rebaser un travail déjà publié sur un dépôt public peut générer des gros problèmes. Supposons que vous clonez un dépôt depuis un serveur central et réalisez quelques travaux dessus. Votre historique de commits ressemble à ceci : Figure 44. Cloner un dépôt et baser du travail dessus À présent, une autre personne travaille et inclut une fusion, puis elle pousse ce travail sur le serveur central. Vous le récupérez et vous fusionnez la nouvelle branche distante dans votre copie, ce qui donne l’historique suivant : Figure 45. Récupération de commits et fusion dans votre copie Ensuite, la personne qui a poussé le travail que vous venez de fusionner décide de faire marche arrière et de rebaser son travail. Elle lance un git push --force pour forcer l’écrasement de l’historique sur le serveur. Vous récupérez alors les données du serveur, qui vous amènent les nouveaux commits . Figure 46. Quelqu’un pousse des commits rebasés, en abandonnant les commits sur lesquels vous avez fondé votre travail Vous êtes désormais tous les deux dans le pétrin. Si vous faites un git pull , vous allez créer un commit de fusion incluant les deux historiques et votre dépôt ressemblera à ça : Figure 47. Vous fusionnez le même travail une nouvelle fois dans un nouveau commit de fusion Si vous lancez git log lorsque votre historique ressemble à ceci, vous verrez deux commits qui ont la même date d’auteur et les mêmes messages, ce qui est déroutant. De plus, si vous poussez cet historique sur le serveur, vous réintroduirez tous ces commits rebasés sur le serveur central, ce qui va encore plus dérouter les autres développeurs. C’est plutôt logique de présumer que l’autre développeur ne souhaite pas voir apparaître C4 et C6 dans l’historique. C’est la raison pour laquelle il avait effectué un rebasage initialement. Rebaser quand vous rebasez Si vous vous retrouvez effectivement dans une situation telle que celle-ci, Git dispose d’autres fonctions magiques qui peuvent vous aider. Si quelqu’un de votre équipe pousse de force des changements qui écrasent des travaux sur lesquels vous vous êtes basés, votre défi est de déterminer ce qui est à vous et ce qui a été réécrit. Il se trouve qu’en plus de l’empreinte SHA du commit , Git calcule aussi une empreinte qui est uniquement basée sur le patch introduit avec le commit. Ceci est appelé un "identifiant de patch" ( patch-id ). Si vous tirez des travaux qui ont été réécrits et les rebasez au-dessus des nouveaux commits de votre collègue, Git peut souvent déterminer ceux qui sont uniquement les vôtres et les réappliquer au sommet de votre nouvelle branche. Par exemple, dans le scénario précédent, si au lieu de fusionner quand nous étions à l’étape Quelqu’un pousse des commits rebasés, en abandonnant les commits sur lesquels vous avez fondé votre travail nous exécutons la commande git rebase teamone/master , Git va : Déterminer quels travaux sont uniques à notre branche (C2, C3, C4, C6, C7) Déterminer ceux qui ne sont pas des commits de fusion (C2, C3, C4) Déterminer ceux qui n’ont pas été réécrits dans la branche de destination (uniquement C2 et C3 puisque C4 est le même patch que C4') Appliquer ces commits au sommet de teamone/master Ainsi, au lieu du résultat que nous avons observé au chapitre Vous fusionnez le même travail une nouvelle fois dans un nouveau commit de fusion , nous aurions pu finir avec quelque chose qui ressemblerait davantage à Rebaser au-dessus de travaux rebasés puis que l’on a poussé en forçant . Figure 48. Rebaser au-dessus de travaux rebasés puis que l’on a poussé en forçant Cela fonctionne seulement si les commits C4 et C4' de votre collègue correspondent presque exactement aux mêmes modifications. Autrement, le rebasage ne sera pas capable de déterminer qu’il s’agit d’un doublon et va ajouter un autre patch similaire à C4 (ce qui échouera probablement puisque les changements sont au moins partiellement déjà présents). Vous pouvez également simplifier tout cela en lançant un git pull --rebase au lieu d’un git pull normal. Vous pouvez encore le faire manuellement à l’aide d’un git fetch suivi d’un git rebase team1/master dans le cas présent. Si vous utilisez git pull et voulez faire de --rebase le traitement par défaut, vous pouvez changer la valeur du paramètre de configuration pull.rebase par git config --global pull.rebase true . Si vous considérez le fait de rebaser comme un moyen de nettoyer et réarranger des commits avant de les pousser et si vous vous en tenez à ne rebaser que des commits qui n’ont jamais été publiés, tout ira bien. Si vous tentez de rebaser des commits déjà publiés sur lesquels les gens ont déjà basé leur travail, vous allez au devant de gros problèmes et votre équipe vous en tiendra rigueur. Si vous ou l’un de vos collègues y trouve cependant une quelconque nécessité, assurez-vous que tout le monde sache lancer un git pull --rebase pour essayer de rendre les choses un peu plus faciles. Rebaser ou Fusionner Maintenant que vous avez vu concrètement ce que signifient rebaser et fusionner, vous devez vous demander ce qu’il est préférable d’utiliser. Avant de pouvoir répondre à cela, revenons quelque peu en arrière et parlons un peu de ce que signifie un historique. On peut voir l’historique des commits de votre dépôt comme un enregistrement de ce qu’il s’est réellement passé . Il s’agit d’un document historique qui a une valeur en tant que tel et ne doit pas être altéré. Sous cet angle, modifier l’historique des commits est presque blasphématoire puisque vous mentez sur ce qu’il s’est réellement passé. Dans ce cas, que faire dans le cas d’une série de commits de fusions désordonnés ? Cela reflète ce qu’il s’est passé et le dépôt devrait le conserver pour la postérité. Le point de vue inverse consiste à considérer que l’historique des commits est le reflet de la façon dont votre projet a été construit . Vous ne publieriez jamais le premier brouillon d’un livre et le manuel de maintenance de votre projet mérite une révision attentive. Ceci constitue le camp de ceux qui utilisent des outils tels que le rebasage et les branches filtrées pour raconter une histoire de la meilleure des manières pour les futurs lecteurs. Désormais, nous espérons que vous comprenez qu’il n’est pas si simple de répondre à la question portant sur le meilleur outil entre fusion et rebasage. Git est un outil puissant et vous permet beaucoup de manipulations sur et avec votre historique mais chaque équipe et chaque projet sont différents. Maintenant que vous savez comment fonctionnent ces deux outils, c’est à vous de décider lequel correspond le mieux à votre situation en particulier. De manière générale, la manière de profiter au mieux des deux mondes consiste à rebaser des modifications locales que vous avez effectuées mais qui n’ont pas encore été partagées avant de les pousser de manière à obtenir un historique propre mais sans jamais rebaser quoi que ce soit que vous ayez déjà poussé quelque part. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/mobaa/
Google SRE - Get glimpse into the mobaa visual patterns Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content Welcome to the Museum of Borgmon Abstract Art! These images offer a glimpse into the visual patterns that appear in our variables and time-series, and the beauty that emerges from chaos. Some of the images in these galleries appeared during difficult rollouts, and some even during production incidents. All come from graphs generated by Google’s monitoring systems. 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Interested in joining SRE? Google strives to cultivate an inclusive workplace. We believe diversity of perspectives and ideas leads to better discussions, decisions, and outcomes for everyone. Explore SRE opportunities at Google Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
https://www.linkedin.com/products/zohocorp-zoho-crm/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
Zoho CRM | LinkedIn Skip to main content LinkedIn Zoho in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Zoho CRM Customer Relationship Management (CRM) Software by Zoho See who's skilled in this Add as skill Learn more Report this product About Zoho CRM is an online customer relationship management software for managing your sales, marketing, support in a single system. Zoho CRM helps businesses of all sizes build excellent customer relationships, with features like an inbuilt AI sales assistant, sales pipeline management, marketing automation, analytics, and more. Zoho CRM empowers a global network of over 150,000 businesses in 180 countries to convert more leads, engage with customers, and grow their revenue. This product is intended for Chief Executive Officer Sales Manager Regional Sales Manager Customer Service Representative Sales Executive Director of Sales Marketing Sales And Marketing Specialist Salesperson Marketer Customer Experience Manager Media Products media viewer No more previous content Get to know Zoho CRM Zoho CRM is an online Customer Relationship Management (CRM) software for managing your sales, marketing & support in a single system. Zoho CRM homepage with dashboards Dashboard is a summarized view of the custom report data in Zoho CRM. It provides a platform for a real-time analysis of the sales stages and business operations. You can create a dashboard and share it with all the users or few selected users. You can also create private dashboards that will be accessible only to you. Customize Zoho CRM to fit your business A truly customizable solution that can be modified as needed to fit your business. Use custom fields, buttons, and layouts to capture various information from your prospects, create custom views, relate information, test customization before rolling out, and even localize the language and currency to fit your users. Analytics and AI In the world of sales, there's no better motivation than numbers. A clear understanding of your company's current performance and its future potential will help you successfully meet changing industry trends and protect you against any adverse economic conditions. Stay on top of your business wherever you go When it comes to business, time is money. If your sales team can't respond quickly, you could be missing out on a deal or be left with an unhappy customer. With Zoho CRM mobile apps, your sales team can keep in touch with their leads on the move, log and access important prospect information, keep track of their KPIs, and make data-driven decisions on the move. You can also find prospects near you so you can keep selling. No more next content Featured customers of Zoho CRM ReBiz Technology, Information and Internet 2,694 followers H+K International Mechanical Or Industrial Engineering 9,267 followers The Streaming Network Technology, Information and Internet 482 followers Arctic Spas (Blue Falls Manufacturing) Wellness and Fitness Services 1,402 followers Call Center Sales Pro Telecommunications 834 followers Recover Health, Inc Hospitals and Health Care 2,584 followers Selectra Technology, Information and Internet 79,939 followers Ecolab, Purolite Resins Manufacturing 31,434 followers Benseron Hospitality - Profit Improvement Specialist for Hospitality Industry Software Development 38 followers Show more Show less Similar products Sales Cloud Sales Cloud Customer Relationship Management (CRM) Software Bigin by Zoho CRM Bigin by Zoho CRM Customer Relationship Management (CRM) Software Experian DataShare Experian DataShare Customer Relationship Management (CRM) Software Odoo CRM Odoo CRM Customer Relationship Management (CRM) Software Freshsales Freshsales Customer Relationship Management (CRM) Software Clickup Clickup Customer Relationship Management (CRM) Software Sign in to see more Show more Show less Zoho products Bigin by Zoho CRM Bigin by Zoho CRM Customer Relationship Management (CRM) Software Zoho Analytics Zoho Analytics Business Intelligence (BI) Software Zoho Commerce Zoho Commerce E-Commerce Platforms Zoho DataPrep Zoho DataPrep Data Preparation Tools Zoho Forms Zoho Forms Online Form Builder Software Zoho FSM Zoho FSM Field Service Management (FSM) Software Zoho One Zoho One Business Management Software Zoho People Zoho People Human Resources Management Systems (HRMS) Zoho RPA Zoho RPA Robotic Process Automation (RPA) Software Zoho Webinar Zoho Webinar Webinar Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
http://www.trello.com/webinars
Trello Webinars: On-Demand Webinars for Productivity & Collaboration | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Trello Webinars Enjoy our free Trello webinars and become a productivity professional. On-Demand Webinars Beyond The Board: Trello's New Views And Card Remote Work Playbook for Team Success Transform The Way Your Company Works With Trello Enterprise Trello Agile Series: The Sprint Board Trello Agile Series: Retrospectives And Roadmaps Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://sre.google/workbook/foreword-I/
Google SRE - SRE Handbook: Insights, Case Studies and IT Foreword I Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Foreword I Mark Burgess Having introduced the first SRE book for O’Reilly, I am honored to be invited back for the sequel. In this book, the writing team is leaving the history of the first book to speak for itself and reaching out to a broader audience, offering direct experiences, case studies, and informal guidance. The broad themes will be familiar to anyone in IT, perhaps relabeled and reprioritized, and with a modern sense of business awareness. In place of technical descriptions, here we have user-facing services and their promises or objectives. We see human-computer systems originate from within the evolving business, intrinsic to its purpose, rather than as foreign meteorites impacting an unsuspecting and pristine infrastructure. Cooperation of all human-computer parts is the focus. Indeed, the book might be summarized as follows: Commit to clear promises that set service objectives, expectations, and levels. Assess those promises continuously, with metrics and budgetary limits. React quickly to keep and repair promises, be on-call, and guard autonomy to avoid new gatekeepers. Keeping promises reliably (to all stakeholders) depends on the stability of all their dependencies, of intent, and of the lives of the people involved (e.g., see Thinking in Promises ). Remarkably, the human aspects of human-computer systems only grow alongside the perceived menace of scale: it turns out that automation doesn’t eliminate humans, after all; rather, it challenges us to reassert human needs across all scales, from the genesis of an individual idea to the massive deployments on behalf of a global user base. Teaching these lessons is a service challenge in its own right—and, like any service, hard-won knowledge is an iterative process. We make these lessons our own by questioning, trying, failing, rehearsing, and perfecting them. There’s a wealth of material to ponder and adapt in the book, so let’s go. Next Foreword II Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://scholar.google.com/citations?hl=ko&oe=ASCII&user=5HoF_9oAAAAJ&view_op=list_works&sortby=title
:root{--gm3-sys-color-on-surface-rgb:31,31,31}@media screen and (prefers-color-scheme:dark){:root{--gm3-sys-color-on-surface-rgb:227,227,227}}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.q4Wquf,.nfoC7c{display:block;height:25vh;position:relative}@media (min-width:600px){.q4Wquf,.nfoC7c{height:150px}}.q4Wquf.Irjbwb{height:auto}@media screen and (prefers-color-scheme:dark){.q4Wquf:not(.GtvzYd){display:none}}.nfoC7c{margin:0;overflow:hidden}.PwpMUe,.lVUmD{display:block;height:100%;margin:0 auto;width:100%}.St9mde{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.St9mde{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.PwpMUe,.lVUmD,.St9mde{-o-object-fit:contain;object-fit:contain}.wsArZ[data-ss-mode="1"] .q4Wquf,.wsArZ[data-ss-mode="1"] .St9mde{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .St9mde{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .q4Wquf,.NQ5OL .St9mde{height:auto;width:100%}.NQ5OL .St9mde{max-width:400px}}.q4Wquf.NWba7e,.q4Wquf.NWba7e .St9mde{height:auto}.q4Wquf.NWba7e .St9mde{height:auto;max-width:312px;width:100%}.q4Wquf.NWba7e.zpCp3 .St9mde{max-width:unset}.q4Wquf.IiQozc .St9mde{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.q4Wquf.Irjbwb .St9mde{height:auto;width:100%}.q4Wquf.EEeaqf .St9mde{max-height:144px;max-width:144px}.SnAaEd{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.SnAaEd{display:none}}.SnAaEd::after,.SnAaEd::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.SnAaEd::after,.SnAaEd::before{width:150px}}.SnAaEd::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.SnAaEd::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .SnAaEd~.St9mde{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .SnAaEd~.St9mde{width:auto}}.RHNWk .St9mde{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .RHNWk .St9mde{width:115px}}.cf660d .St9mde{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.tUhwwc .St9mde{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.Jkvqxd .St9mde{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.onc8Ic .St9mde{-webkit-transform:translate(var( --c-ps-s,24px ),0);-ms-transform:translate(var( --c-ps-s,24px ),0);transform:translate(var( --c-ps-s,24px ),0)}.WA89Yb .St9mde{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .XEN8Yb .St9mde{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .XEN8Yb .St9mde{max-width:115px}}@media (min-width:1240px) and (orientation:landscape),all and (min-width:1600px){.KyPKed .NQ5OL .XEN8Yb .St9mde{max-width:115px}}.IsSr6b .St9mde{max-width:300px}.mmskdd .St9mde{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@-webkit-keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@-webkit-keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@-webkit-keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}:root{--wf-tfs:calc(var(--c-tfs,32)/16*1rem);--wf-tfs-bp2:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp3:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp5:calc(var(--c-tfs,44)/16*1rem);--wf-stfs:calc(var(--c-stfs,16)/16*1rem);--wf-stfs-bp5:calc(var(--c-stfs,16)/16*1rem)}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.Dzz9Db,.GpMPBe{display:block;height:25vh;position:relative}@media (min-width:600px){.Dzz9Db,.GpMPBe{height:150px}}@media screen and (prefers-color-scheme:dark){.Dzz9Db:not(.GtvzYd){display:none}}.Dzz9Db.Irjbwb{height:auto}.GpMPBe{margin:0;overflow:hidden}.UFQPDd,.JNOvdd{display:block;height:100%;margin:0 auto;-o-object-fit:contain;object-fit:contain;width:100%}.f4ZpM{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.f4ZpM{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.wsArZ[data-ss-mode="1"] .Dzz9Db,.wsArZ[data-ss-mode="1"] .f4ZpM{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .f4ZpM{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .Dzz9Db,.NQ5OL .f4ZpM{height:auto;width:100%}.NQ5OL .f4ZpM{max-width:400px}}.Dzz9Db.utFBGf,.Dzz9Db.utFBGf .f4ZpM{height:auto}.Dzz9Db.utFBGf .f4ZpM{height:auto;max-width:312px;width:100%}.Dzz9Db.utFBGf.zpCp3 .f4ZpM{max-width:unset}.Dzz9Db.IiQozc .f4ZpM{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.Dzz9Db.Irjbwb .f4ZpM{height:auto;width:100%}.Dzz9Db.EEeaqf .f4ZpM{max-height:144px;max-width:144px}.nPt1pc{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.nPt1pc{display:none}}.nPt1pc::after,.nPt1pc::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.nPt1pc::after,.nPt1pc::before{width:150px}}.nPt1pc::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.nPt1pc::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .nPt1pc~.f4ZpM{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .nPt1pc~.f4ZpM{width:auto}}.ZS7CGc .f4ZpM{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .ZS7CGc .f4ZpM{width:115px}}.qiRZ5e .f4ZpM{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.vIv7Gf .f4ZpM{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.nvYXVd .f4ZpM{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.uOhnzd .f4ZpM{-webkit-transform:translate(24px,0);-ms-transform:translate(24px,0);transform:translate(24px,0)}.MsYMaf .f4ZpM{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .YIi9qf .f4ZpM{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .YIi9qf .f4ZpM{max-width:115px}}.QG3Xbe .f4ZpM{max-width:300px}.F6gtje .f4ZpM{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@-webkit-keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@-webkit-keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}@keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}.VfPpkd-ksKsZd-XxIAqe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity;position:relative;outline:none;overflow:hidden}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-ksKsZd-XxIAqe::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-ksKsZd-XxIAqe::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded],.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd{overflow:visible}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:0;left:0;width:100%;height:100%}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-ksKsZd-XxIAqe:hover::before,.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe::before,.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):focus::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d)::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):active::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe{font-size:24px;width:48px;height:48px;padding:12px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-Jh9lGc{width:40px;height:40px;margin-top:4px;margin-bottom:4px;margin-right:4px;margin-left:4px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-J1Ukfc-LhBDec{max-height:40px;max-width:40px}.VfPpkd-Bz112c-LgbsSe:disabled{color:rgba(0,0,0,.38);color:var(--mdc-theme-text-disabled-on-light,rgba(0,0,0,.38))}.VfPpkd-Bz112c-LgbsSe svg,.VfPpkd-Bz112c-LgbsSe img{width:24px;height:24px}.VfPpkd-Bz112c-LgbsSe{display:inline-block;position:relative;-webkit-box-sizing:border-box;box-sizing:border-box;border:none;outline:none;background-color:transparent;fill:currentColor;color:inherit;text-decoration:none;cursor:pointer;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;z-index:0;overflow:visible}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-RLmnJb{position:absolute;top:50%;height:48px;left:50%;width:48px;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-J1Ukfc-LhBDec,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-J1Ukfc-LhBDec{display:block}}.VfPpkd-Bz112c-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-Bz112c-LgbsSe[hidden]{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-KVuj8d-Q3DXx{-webkit-box-align:center;-webkit-align-items:center;align-items:center;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center}.VfPpkd-Bz112c-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:100%;width:100%;display:none}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{border-color:CanvasText}}.VfPpkd-Bz112c-kBDsod{display:inline-block}.VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd,.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd{display:inline-block}.VfPpkd-Bz112c-mRLv6{height:100%;left:0;outline:none;position:absolute;top:0;width:100%}.VfPpkd-Bz112c-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;width:100%;height:100%}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-Bz112c-LgbsSe:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe:disabled:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc{height:100%;left:0;pointer-events:none;position:absolute;top:0;width:100%;z-index:-1}.VfPpkd-dgl2Hf-ppHlrf-sM5MNb{display:inline}.VfPpkd-LgbsSe{position:relative;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-align:center;-webkit-align-items:center;align-items:center;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center;-webkit-box-sizing:border-box;box-sizing:border-box;min-width:64px;border:none;outline:none;line-height:inherit;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;-webkit-appearance:none;overflow:visible;vertical-align:middle;background:transparent}.VfPpkd-LgbsSe .VfPpkd-BFbNVe-bF1uUb{width:100%;height:100%;top:0;left:0}.VfPpkd-LgbsSe::-moz-focus-inner{padding:0;border:0}.VfPpkd-LgbsSe:active{outline:none}.VfPpkd-LgbsSe:hover{cursor:pointer}.VfPpkd-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-LgbsSe[hidden]{display:none}.VfPpkd-LgbsSe .VfPpkd-kBDsod{margin-left:0;margin-right:8px;display:inline-block;position:relative;vertical-align:top}[dir=rtl] .VfPpkd-LgbsSe .VfPpkd-kBDsod,.VfPpkd-LgbsSe .VfPpkd-kBDsod[dir=rtl]{margin-left:8px;margin-right:0}.VfPpkd-LgbsSe .VfPpkd-UdE5de-uDEFge{font-size:0;position:absolute;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);top:50%;left:50%;line-height:normal}.VfPpkd-LgbsSe .VfPpkd-vQzf8d{position:relative}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px);display:none}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-color:CanvasText}}@media screen and (forced-colors:active){.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-J1Ukfc-LhBDec,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-J1Ukfc-LhBDec{display:block}}.VfPpkd-LgbsSe .VfPpkd-RLmnJb{position:absolute;top:50%;height:48px;left:0;right:0;-webkit-transform:translateY(-50%);-ms-transform:translateY(-50%);transform:translateY(-50%)}.VfPpkd-vQzf8d+.VfPpkd-kBDsod{margin-left:8px;margin-right:0}[dir=rtl] .VfPpkd-vQzf8d+.VfPpkd-kBDsod,.VfPpkd-vQzf8d+.VfPpkd-kBDsod[dir=rtl]{margin-left:0;margin-right:8px}svg.VfPpkd-kBDsod{fill:currentColor}.VfPpkd-LgbsSe-OWXEXe-dgl2Hf{margin-top:6px;margin-bottom:6px}.VfPpkd-LgbsSe{-moz-osx-font-smoothing:grayscale;-webkit-font-smoothing:antialiased;text-decoration:none}.VfPpkd-LgbsSe{padding:0 8px 0 8px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-INsAgc{border-style:solid;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-style:solid;border-color:transparent}.VfPpkd-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{z-index:0}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Jh9lGc{position:absolute;-webkit-box-sizing:content-box;box-sizing:content-box;overflow:hidden;z-index:0;top:0;left:0;bottom:0;right:0}.VfPpkd-LgbsSe{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-text-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-filled-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;-webkit-box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12);box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-protected-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled):active{-webkit-box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12);box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{-webkit-box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12);box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;padding:0 15px 0 15px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-outlined-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 11px 0 15px}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 15px 0 11px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{top:-1px;left:-1px;bottom:-1px;right:-1px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-RLmnJb{left:-1px;width:calc(100% + 2px)}.nCP5yc{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.nCP5yc .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.nCP5yc:not(:disabled){background-color:rgb(26,115,232);background-color:var(--gm-fillbutton-container-color,rgb(26,115,232))}.nCP5yc:not(:disabled){color:#fff;color:var(--gm-fillbutton-ink-color,#fff)}.nCP5yc:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-fillbutton-disabled-container-color,rgba(60,64,67,.12))}.nCP5yc:disabled{color:rgba(60,64,67,.38);color:var(--gm-fillbutton-disabled-ink-color,rgba(60,64,67,.38))}.nCP5yc .VfPpkd-Jh9lGc::before,.nCP5yc .VfPpkd-Jh9lGc::after{background-color:rgb(32,33,36);background-color:var(--gm-fillbutton-state-color,rgb(32,33,36))}.nCP5yc:hover .VfPpkd-Jh9lGc::before,.nCP5yc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.16;opacity:var(--mdc-ripple-hover-opacity,.16)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24;opacity:var(--mdc-ripple-focus-opacity,.24)}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.2;opacity:var(--mdc-ripple-press-opacity,.2)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.2)}.nCP5yc .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:#fff}@media (-ms-high-contrast:active),screen and (forced-colors:active){.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.nCP5yc:hover{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:hover .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:active{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:active .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:disabled{-webkit-box-shadow:none;box-shadow:none}.nCP5yc:disabled:hover .VfPpkd-Jh9lGc::before,.nCP5yc:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.Rj2Mlf .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.Rj2Mlf:not(:disabled){color:rgb(26,115,232);color:var(--gm-hairlinebutton-ink-color,rgb(26,115,232))}.Rj2Mlf:not(:disabled){border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled):hover{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled).VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.Rj2Mlf:not(:disabled):not(.VfPpkd-ksKsZd-mWPk3d):focus{border-color:rgb(23,78,166);border-color:var(--gm-hairlinebutton-outline-color--stateful,rgb(23,78,166))}.Rj2Mlf:not(:disabled):active,.Rj2Mlf:not(:disabled):focus:active{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:disabled{color:rgba(60,64,67,.38);color:var(--gm-hairlinebutton-disabled-ink-color,rgba(60,64,67,.38))}.Rj2Mlf:disabled{border-color:rgba(60,64,67,.12);border-color:var(--gm-hairlinebutton-disabled-outline-color,rgba(60,64,67,.12))}.Rj2Mlf:hover:not(:disabled),.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.Rj2Mlf:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-hairlinebutton-ink-color--stateful,rgb(23,78,166))}.Rj2Mlf .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.Rj2Mlf .VfPpkd-Jh9lGc::before,.Rj2Mlf .VfPpkd-Jh9lGc::after{background-color:rgb(26,115,232);background-color:var(--gm-hairlinebutton-state-color,rgb(26,115,232))}.Rj2Mlf:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.Rj2Mlf:disabled:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.b9hyVd{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);border-width:0;-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15))}.b9hyVd .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.b9hyVd:not(:disabled){background-color:#fff;background-color:var(--gm-protectedbutton-container-color,#fff)}.b9hyVd:not(:disabled){color:rgb(26,115,232);color:var(--gm-protectedbutton-ink-color,rgb(26,115,232))}.b9hyVd:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-protectedbutton-disabled-container-color,rgba(60,64,67,.12))}.b9hyVd:disabled{color:rgba(60,64,67,.38);color:var(--gm-protectedbutton-disabled-ink-color,rgba(60,64,67,.38))}.b9hyVd:hover:not(:disabled),.b9hyVd.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.b9hyVd:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.b9hyVd:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-protectedbutton-ink-color--stateful,rgb(23,78,166))}.b9hyVd .VfPpkd-BFbNVe-bF1uUb{opacity:0}.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-B
2026-01-13T09:29:19
https://www.linkedin.com/products/solarwinds-service-desk/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
SolarWinds Service Desk | LinkedIn Skip to main content LinkedIn SolarWinds in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in SolarWinds Service Desk Service Desk Software by SolarWinds See who's skilled in this Add as skill Learn more Report this product About An IT service management (ITSM) solution that understands what it takes to successfully manage your employee services. Try Solarwinds Service Desk for free! Media Products media viewer No more previous content No more next content Similar products Jira Service Management Jira Service Management Service Desk Software Freshservice Freshservice Service Desk Software Intercom Intercom Service Desk Software Atomicwork Atomicwork Service Desk Software Service Desk Service Desk Service Desk Software TOPdesk TOPdesk Service Desk Software Sign in to see more Show more Show less SolarWinds products NetFlow Traffic Analyzer NetFlow Traffic Analyzer Network Traffic Analysis (NTA) Tools Security Event Manager Security Event Manager Security Information & Event Management (SIEM) Software Serv-U File Transfer Protocol Server Serv-U File Transfer Protocol Server File Transfer Protocol (FTP) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software SolarWinds AppOptics APM SolarWinds AppOptics APM Application Performance Monitoring (APM) Software Solarwinds Mail Assure Solarwinds Mail Assure Email Security Software SolarWinds Papertrail SolarWinds Papertrail Log Management Software SolarWinds Web Performance Monitor SolarWinds Web Performance Monitor Website Monitoring Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/de/v2/Git-auf-dem-Server-Smart-HTTP
Git - Smart HTTP About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Erste Schritte 1.1 Was ist Versionsverwaltung? 1.2 Kurzer Überblick über die Historie von Git 1.3 Was ist Git? 1.4 Die Kommandozeile 1.5 Git installieren 1.6 Git Basis-Konfiguration 1.7 Hilfe finden 1.8 Zusammenfassung 2. Git Grundlagen 2.1 Ein Git-Repository anlegen 2.2 Änderungen nachverfolgen und im Repository speichern 2.3 Anzeigen der Commit-Historie 2.4 Ungewollte Änderungen rückgängig machen 2.5 Mit Remotes arbeiten 2.6 Taggen 2.7 Git Aliases 2.8 Zusammenfassung 3. Git Branching 3.1 Branches auf einen Blick 3.2 Einfaches Branching und Merging 3.3 Branch-Management 3.4 Branching-Workflows 3.5 Remote-Branches 3.6 Rebasing 3.7 Zusammenfassung 4. Git auf dem Server 4.1 Die Protokolle 4.2 Git auf einem Server einrichten 4.3 Erstellung eines SSH-Public-Keys 4.4 Einrichten des Servers 4.5 Git-Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Von Drittanbietern gehostete Optionen 4.10 Zusammenfassung 5. Verteiltes Git 5.1 Verteilter Arbeitsablauf 5.2 An einem Projekt mitwirken 5.3 Ein Projekt verwalten 5.4 Zusammenfassung 6. GitHub 6.1 Einrichten und Konfigurieren eines Kontos 6.2 Mitwirken an einem Projekt 6.3 Ein Projekt betreuen 6.4 Verwalten einer Organisation 6.5 Skripte mit GitHub 6.6 Zusammenfassung 7. Git Tools 7.1 Revisions-Auswahl 7.2 Interaktives Stagen 7.3 Stashen und Bereinigen 7.4 Deine Arbeit signieren 7.5 Suchen 7.6 Den Verlauf umschreiben 7.7 Reset entzaubert 7.8 Fortgeschrittenes Merging 7.9 Rerere 7.10 Debuggen mit Git 7.11 Submodule 7.12 Bundling 7.13 Replace (Ersetzen) 7.14 Anmeldeinformationen speichern 7.15 Zusammenfassung 8. Git einrichten 8.1 Git Konfiguration 8.2 Git-Attribute 8.3 Git Hooks 8.4 Beispiel für Git-forcierte Regeln 8.5 Zusammenfassung 9. Git und andere VCS-Systeme 9.1 Git als Client 9.2 Migration zu Git 9.3 Zusammenfassung 10. Git Interna 10.1 Basisbefehle und Standardbefehle (Plumbing and Porcelain) 10.2 Git Objekte 10.3 Git Referenzen 10.4 Packdateien (engl. Packfiles) 10.5 Die Referenzspezifikation (engl. Refspec) 10.6 Transfer Protokolle 10.7 Wartung und Datenwiederherstellung 10.8 Umgebungsvariablen 10.9 Zusammenfassung A1. Anhang A: Git in anderen Umgebungen A1.1 Grafische Schnittstellen A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Zusammenfassung A2. Anhang B: Git in deine Anwendungen einbetten A2.1 Die Git-Kommandozeile A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Anhang C: Git Kommandos A3.1 Setup und Konfiguration A3.2 Projekte importieren und erstellen A3.3 Einfache Snapshot-Funktionen A3.4 Branching und Merging A3.5 Projekte gemeinsam nutzen und aktualisieren A3.6 Kontrollieren und Vergleichen A3.7 Debugging A3.8 Patchen bzw. Fehlerkorrektur A3.9 E-mails A3.10 Externe Systeme A3.11 Administration A3.12 Basisbefehle 2nd Edition 4.6 Git auf dem Server - Smart HTTP Smart HTTP Wir haben jetzt authentifizierten Zugriff über SSH und nicht authentifizierten Zugriff über git:// , aber es gibt auch ein Protokoll, das beides gleichzeitig kann. Die Einrichtung von Smart HTTP ist im Grunde genommen nur die Aktivierung eines CGI-Skripts, das zusammen mit Git namens git-http-backend auf dem Server bereitgestellt wird. Dieses CGI liest den Pfad und die Header, die von einem git fetch oder git push an eine HTTP-URL gesendet werden, und bestimmt, ob der Client über HTTP kommunizieren kann (was für jeden Client ab Version 1.6.6 gilt). Wenn das CGI sieht, dass der Client intelligent ist, kommuniziert es intelligent mit ihm; andernfalls fällt es auf das dumme Verhalten zurück (also ist es rückwärtskompatibel für Lesezugriffe mit älteren Clients). Lass uns durch ein sehr einfaches Setup gehen. Wir werden Apache als CGI-Server verwenden. Wenn du kein Apache-Setup hast, kannst du dies auf einem Linux-System, wie nachfolgend beschrieben einrichten: $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env Dadurch werden auch die Module mod_cgi , mod_alias , und mod_env aktiviert, die alle benötigt werden, damit das Ganze ordnungsgemäß funktioniert. Du solltest auch die Unix-Benutzergruppe im Verzeichnis /srv/git auf www-data setzen, damit dein Webserver auf die Repositorys lesend und schreibend zugreifen kann. Die Apache-Instanz, auf der das CGI-Skript läuft, wird standardmäßig unter dieser Benutzer laufen: $ chgrp -R www-data /srv/git Als nächstes müssen wir der Apache-Konfiguration einige Dinge hinzufügen, um das git-http-backend als Handler für alles, was im`/git` Pfad dein Webservers liegt, auszuführen. SetEnv GIT_PROJECT_ROOT /srv/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/lib/git-core/git-http-backend/ Wenn du die Umgebungsvariable GIT_HTTP_EXPORT_ALL nicht setzt, wird Git unauthentifizierten Clients nur die Repositorys, die die Datei git-daemon-export-ok enthalten, zur Verfügung stellen. Das Verhalten ist dann wie beim Git-Daemon. Abschließend konfigurierst du Apache so, dass er Anfragen an das git-http-backend zulassen soll, um Schreibvorgänge zu authentifiziert. Dazu kannst du folgenden Code nutzen: <Files "git-http-backend"> AuthType Basic AuthName "Git Access" AuthUserFile /srv/git/.htpasswd Require expr !(%{QUERY_STRING} -strmatch '*service=git-receive-pack*' || %{REQUEST_URI} =~ m#/git-receive-pack$#) Require valid-user </Files> Dazu musst du eine .htpasswd Datei erstellen, die die Passwörter aller gültigen Benutzer enthält. Hier ein Beispiel für das Hinzufügen eines Benutzers „schacon“: $ htpasswd -c /srv/git/.htpasswd schacon Es gibt unzählige Möglichkeiten, Benutzer mit Apache zu authentifizieren. Du musst eine von ihnen auswählen und implementieren. Dies ist ein einfaches Beispiel. Du wirst dies wahrscheinlich über SSL konfigurieren wollen, damit auch alle Daten verschlüsselt werden. Wir wollen nicht zu weit in das Konzept der Apache-Konfigurationsspezifikationen eindringen, da du möglicherweise einen anderen Server verwendst oder andere Authentifizierungsanforderungen hast. Die Idee ist, dass Git mit einem CGI mit dem Namen git-http-backend daherkommt, das beim Senden und Empfangen von Daten die Kommunikation über HTTP aushandelt. Es implementiert selbst keine Authentifizierung, aber diese kann über den Webservers gesteuert werden. Du kannst das mit fast jedem CGI-fähigen Webserver tun. Am besten wählst du denjenigen, den du am besten kennst. Anmerkung Weitere Informationen zur Konfiguration der Authentifizierung in Apache findest du in den Apache-Dokumenten unter: https://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/books/
Google SRE book- Comprehensive guide to site reliability Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content SRE Books Read online View details Read online View details Book updates Read online View details Building Secure & Reliable Systems By: Heather Adkins, Betsy Beyer, Paul Blankinship, Ana Oprea, Piotr Lewandowski, Adam Stubblefield Can a system be considered truly reliable if it isn't fundamentally secure? Or can it be considered secure if it's unreliable? Security is crucial to the design and operation of scalable systems in production, as it plays an important part in product quality, performance, and availability. In this book, experts from Google share best practices to help your organization design scalable and reliable systems that are fundamentally secure. Buy from Google Books Read online The Site Reliability Workbook Edited by: Betsy Beyer, Niall Richard Murphy, David K. Rensin, Kent Kawahara and Stephen Thorne The Site Reliability Workbook is the hands-on companion to the bestselling Site Reliability Engineering book and uses concrete examples to show how to put SRE principles and practices to work. This book contains practical examples from Google’s experiences and case studies from Google’s Cloud Platform customers. Evernote, The Home Depot, The New York Times, and other companies outline hard-won experiences of what worked for them and what didn’t. Buy from Google Books Read online Site Reliability Engineering Edited by: Betsy Beyer, Chris Jones, Jennifer Petoff and Niall Richard Murphy Members of the SRE team explain how their engagement with the entire software lifecycle has enabled Google to build, deploy, monitor, and maintain some of the largest software systems in the world. Buy from Google Books Read online Interested in joining SRE? Google strives to cultivate an inclusive workplace. We believe diversity of perspectives and ideas leads to better discussions, decisions, and outcomes for everyone. Explore SRE opportunities at Google Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
https://git-scm.com/book/bg/v2/GitHub-%d0%9e%d0%b1%d0%be%d0%b1%d1%89%d0%b5%d0%bd%d0%b8%d0%b5
Git - Обобщение About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Начало 1.1 За Version Control системите 1.2 Кратка история на Git 1.3 Какво е Git 1.4 Конзолата на Git 1.5 Инсталиране на Git 1.6 Първоначална настройка на Git 1.7 Помощна информация в Git 1.8 Обобщение 2. Основи на Git 2.1 Създаване на Git хранилище 2.2 Запис на промени в хранилището 2.3 Преглед на историята на действията 2.4 Възстановяване на направени действия 2.5 Работа с отдалечени хранилища 2.6 Тагове в Git 2.7 Псевдоними в Git 2.8 Обобщение 3. Клонове в Git 3.1 Накратко за разклоненията 3.2 Основи на клоновете код и сливането 3.3 Управление на клонове 3.4 Стратегии за работа с клонове код 3.5 Отдалечени клонове 3.6 Управление на проект 3.7 Обобщение 4. GitHub 4.1 Създаване и настройка на акаунт 4.2 Как да сътрудничим в проект 4.3 Управление на проект 4.4 Управление на организация 4.5 Автоматизиране с GitHub 4.6 Обобщение 5. Git инструменти 5.1 Избор на къмити 5.2 Интерактивно индексиране 5.3 Stashing и Cleaning 5.4 Подписване на вашата работа 5.5 Търсене 5.6 Манипулация на историята 5.7 Мистерията на командата Reset 5.8 Сливане за напреднали 5.9 Rerere 5.10 Дебъгване с Git 5.11 Подмодули 5.12 Пакети в Git (Bundling) 5.13 Заместване 5.14 Credential Storage система 5.15 Обобщение 6. Настройване на Git 6.1 Git конфигурации 6.2 Git атрибути 6.3 Git Hooks 6.4 Примерна Git-Enforced политика 6.5 Обобщение 7. Git и други системи 7.1 Git като клиент 7.2 Миграция към Git 7.3 Обобщение 8. Git на ниско ниво 8.1 Plumbing и Porcelain команди 8.2 Git обекти 8.3 Git референции 8.4 Packfiles 8.5 Refspec спецификации 8.6 Транспортни протоколи 8.7 Поддръжка и възстановяване на данни 8.8 Environment променливи 8.9 Обобщение 9. Приложение A: Git в други среди 9.1 Графични интерфейси 9.2 Git във Visual Studio 9.3 Git във Visual Studio Code 9.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 9.5 Git в Sublime Text 9.6 Git в Bash 9.7 Git в Zsh 9.8 Git в PowerShell 9.9 Обобщение 10. Приложение B: Вграждане на Git в приложения 10.1 Git от команден ред 10.2 Libgit2 10.3 JGit 10.4 go-git 10.5 Dulwich A1. Приложение C: Git команди A1.1 Настройки и конфигурация A1.2 Издърпване и създаване на проекти A1.3 Snapshotting A1.4 Клонове и сливане A1.5 Споделяне и обновяване на проекти A1.6 Инспекция и сравнение A1.7 Дебъгване A1.8 Patching A1.9 Email команди A1.10 Външни системи A1.11 Административни команди A1.12 Plumbing команди 2nd Edition 4.6 GitHub - Обобщение Обобщение Сега вече сте GitHub потребител. Знаете как се създава акаунт, как се управлява организация, как се създават и актуализират хранилища, как да допринасяте към проектите на друго потребители и да приемате код от тях във вашите проекти. В следващата глава ще разгледаме някои мощни инструменти и съвети за справяне със сложни ситуации, които ще ви направят майстори на Git. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-arbor-cloud-ddos-protection/?trk=products_seo_search
Arbor Cloud DDoS Protection | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Arbor Cloud DDoS Protection DDoS Protection Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Arbor Cloud combines on-premise DDoS defense with cloud-based traffic scrubbing services that are tightly integrated via an automated cloud signal. This multi-layered, adaptive DDoS protection approach is a proven industry best practice and is the only way to mitigate today's full spectrum of DDoS threats for both Service Providers and Enterprises, all from a single cloud DDoS protection provider. This product is intended for Cyber Security Engineer Chief Executive Officer Chief Information Officer Network Operations Center Head of Information Technology Network Security Engineer Security Engineer Director of Security Information Technology Specialist Director of Information Technology Media Products media viewer No more previous content Arbor Cloud Interface Arbor Cloud Arbor Cloud Infrastructure No more next content Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less NETSCOUT products Arbor Edge Defense Arbor Edge Defense DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/zh/v2/%e6%9c%8d%e5%8a%a1%e5%99%a8%e4%b8%8a%e7%9a%84-Git-Smart-HTTP
Git - Smart HTTP About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. 起步 1.1 关于版本控制 1.2 Git 简史 1.3 Git 是什么? 1.4 命令行 1.5 安装 Git 1.6 初次运行 Git 前的配置 1.7 获取帮助 1.8 总结 2. Git 基础 2.1 获取 Git 仓库 2.2 记录每次更新到仓库 2.3 查看提交历史 2.4 撤消操作 2.5 远程仓库的使用 2.6 打标签 2.7 Git 别名 2.8 总结 3. Git 分支 3.1 分支简介 3.2 分支的新建与合并 3.3 分支管理 3.4 分支开发工作流 3.5 远程分支 3.6 变基 3.7 总结 4. 服务器上的 Git 4.1 协议 4.2 在服务器上搭建 Git 4.3 生成 SSH 公钥 4.4 配置服务器 4.5 Git 守护进程 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 第三方托管的选择 4.10 总结 5. 分布式 Git 5.1 分布式工作流程 5.2 向一个项目贡献 5.3 维护项目 5.4 总结 6. GitHub 6.1 账户的创建和配置 6.2 对项目做出贡献 6.3 维护项目 6.4 管理组织 6.5 脚本 GitHub 6.6 总结 7. Git 工具 7.1 选择修订版本 7.2 交互式暂存 7.3 贮藏与清理 7.4 签署工作 7.5 搜索 7.6 重写历史 7.7 重置揭密 7.8 高级合并 7.9 Rerere 7.10 使用 Git 调试 7.11 子模块 7.12 打包 7.13 替换 7.14 凭证存储 7.15 总结 8. 自定义 Git 8.1 配置 Git 8.2 Git 属性 8.3 Git 钩子 8.4 使用强制策略的一个例子 8.5 总结 9. Git 与其他系统 9.1 作为客户端的 Git 9.2 迁移到 Git 9.3 总结 10. Git 内部原理 10.1 底层命令与上层命令 10.2 Git 对象 10.3 Git 引用 10.4 包文件 10.5 引用规范 10.6 传输协议 10.7 维护与数据恢复 10.8 环境变量 10.9 总结 A1. 附录 A: 在其它环境中使用 Git A1.1 图形界面 A1.2 Visual Studio 中的 Git A1.3 Visual Studio Code 中的 Git A1.4 IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 中的 Git A1.5 Sublime Text 中的 Git A1.6 Bash 中的 Git A1.7 Zsh 中的 Git A1.8 PowerShell 中的 Git A1.9 总结 A2. 附录 B: 在你的应用中嵌入 Git A2.1 命令行 Git 方式 A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. 附录 C: Git 命令 A3.1 设置与配置 A3.2 获取与创建项目 A3.3 快照基础 A3.4 分支与合并 A3.5 项目分享与更新 A3.6 检查与比较 A3.7 调试 A3.8 补丁 A3.9 邮件 A3.10 外部系统 A3.11 管理 A3.12 底层命令 2nd Edition 4.6 服务器上的 Git - Smart HTTP Smart HTTP 我们一般通过 SSH 进行授权访问,通过 git:// 进行无授权访问,但是还有一种协议可以同时实现以上两种方式的访问。 设置 Smart HTTP 一般只需要在服务器上启用一个 Git 自带的名为 git-http-backend 的 CGI 脚本。 该 CGI 脚本将会读取由 git fetch 或 git push 命令向 HTTP URL 发送的请求路径和头部信息, 来判断该客户端是否支持 HTTP 通信(不低于 1.6.6 版本的客户端支持此特性)。 如果 CGI 发现该客户端支持智能(Smart)模式,它将会以智能模式与它进行通信, 否则它将会回落到哑(Dumb)模式下(因此它可以对某些老的客户端实现向下兼容)。 在完成以上简单的安装步骤后, 我们将用 Apache 来作为 CGI 服务器。 如果你没有安装 Apache,你可以在 Linux 环境下执行如下或类似的命令来安装: $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env 该操作将会启用 mod_cgi , mod_alias 和 mod_env 等 Apache 模块, 这些模块都是使该功能正常工作所必须的。 你还需要将 /srv/git 的 Unix 用户组设置为 www-data ,这样 Web 服务器才能读写该仓库, 因为运行 CGI 脚本的 Apache 实例默认会以该用户的权限运行: $ chgrp -R www-data /srv/git 接下来我们要向 Apache 配置文件添加一些内容,来让 git-http-backend 作为 Web 服务器对 /git 路径请求的处理器。 SetEnv GIT_PROJECT_ROOT /srv/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/lib/git-core/git-http-backend/ 如果留空 GIT_HTTP_EXPORT_ALL 这个环境变量,Git 将只对无授权客户端提供带 git-daemon-export-ok 文件的版本库,就像 Git 守护进程一样。 最后,如果想让 Apache 允许 git-http-backend 请求并实现写入操作的授权验证,使用如下授权屏蔽配置即可: <Files "git-http-backend"> AuthType Basic AuthName "Git Access" AuthUserFile /srv/git/.htpasswd Require expr !(%{QUERY_STRING} -strmatch '*service=git-receive-pack*' || %{REQUEST_URI} =~ m#/git-receive-pack$#) Require valid-user </Files> 这需要你创建一个包含所有合法用户密码的 .htpasswd 文件。 以下是一个添加 “schacon” 用户到此文件的例子: $ htpasswd -c /srv/git/.htpasswd schacon 你可以通过许多方式添加 Apache 授权用户,选择使用其中一种方式即可。 以上仅仅只是我们可以找到的最简单的一个例子。 如果愿意的话,你也可以通过 SSL 运行它,以保证所有数据是在加密状态下进行传输的。 我们不想深入去讲解 Apache 配置文件,因为你可能会使用不同的 Web 服务器,或者可能有不同的授权需求。 它的主要原理是使用一个 Git 附带的,名为 git-http-backend 的 CGI。它被引用来处理协商通过 HTTP 发送和接收的数据。 它本身并不包含任何授权功能,但是授权功能可以在 Web 服务器层引用它时被轻松实现。 你可以在任何所有可以处理 CGI 的 Web 服务器上办到这点,所以随便挑一个你最熟悉的 Web 服务器试手吧。 Note 欲了解更多的有关配置 Apache 授权访问的信息,请通过以下链接浏览 Apache 文档: https://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/resources/practices-and-processes/anatomy-of-an-incident/#page-content
Google SRE - Anatomy of an incident management handbook Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content Anatomy of an Incident - Google - Site Reliability Engineering When it comes to system design, failure is inevitable. Scientists and engineers implement solutions based on the available information, without a complete knowledge of the future. You can’t always anticipate the next zero-day event, viral media trend, weather disaster, or shift in technology. But you can be prepared to respond when incidents like these affect your systems. With this report, SRE and DevOps practitioners, IT managers, and engineering leaders will explore methods to help your organization prepare for, respond to, and recover from incidents. With advice from Ayelet Sachto, Adrienne Walcer, and Jessie Yang, you’ll learn how to be prepared to handle failure if and when it happens. Learn the stages of the incident management lifecycle : preparedness, response, recovery, and mitigation Deal proactively with incidents: issues that escalate beyond metrics and alerts Be prepared: practice disaster role playing and incident response exercises Learn the characteristics of the incident-response organizational structure Examine steps to recovery and mitigation after an incident has occurred Conduct postmortems to analyze what went wrong Explore a real-world example from Google: The Mayan Apocalypse Learn how to measure and reduce incidents impact Use postmortems as a tool for prevention and psychological safety PDF EPUB MOBI Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
https://sre.google/workbook/monitoring/
Google SRE - Monitoring Systems with Advanced Analytics Chapter 4 - Monitoring Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Monitoring By Jess Frame, Anthony Lenton, Steven Thurgood, Anton Tolchanov, and Nejc Trdin with Carmela Quinito Monitoring can include many types of data, including metrics, text logging, structured event logging, distributed tracing, and event introspection. While all of these approaches are useful in their own right, this chapter mostly addresses metrics and structured logging. In our experience, these two data sources are best suited to SRE’s fundamental monitoring needs. At the most basic level, monitoring allows you to gain visibility into a system, which is a core requirement for judging service health and diagnosing your service when things go wrong. Chapter 6 in the first SRE book provides some basic monitoring definitions and explains that SREs monitor their systems in order to: Alert on conditions that require attention. Investigate and diagnose those issues. Display information about the system visually. Gain insight into trends in resource usage or service health for long-term planning. Compare the behavior of the system before and after a change, or between two groups in an experiment. The relative importance of these use cases might lead you to make tradeoffs when selecting or building a monitoring system. This chapter talks about how Google manages monitoring systems and provides some guidelines for questions that may arise when you’re choosing and running a monitoring system. Desirable Features of a Monitoring Strategy When choosing a monitoring system, it is important to understand and prioritize the features that matter to you. If you’re evaluating different monitoring systems, the attributes in this section can help guide your thinking about which solution(s) best suits your needs. If you already have a monitoring strategy, you might consider using some additional capabilities of your current solution. Depending on your needs, one monitoring system may address all of your use cases, or you may want to use a combination of systems. Speed Different organizations will have different needs when it comes to the freshness of data and the speed of data retrieval . Data should be available when you need it: freshness impacts how long it will take your monitoring system to page you when something goes wrong. Additionally, slow data might lead you to accidentally act on incorrect data. For example, during incident response, if the time between cause (taking an action) and effect (seeing that action reflected in your monitoring) is too long, you might assume a change had no effect or deduce a false correlation between cause and effect. Data more than four to five minutes stale might significantly impact how quickly you can respond to an incident. If you’re selecting a monitoring system based upon this criteria, you need to figure out your speed requirements ahead of time. Speed of data retrieval is mostly a problem when you’re querying vast amounts of data. It might take some time for a graph to load if it has to tally up a lot of data from many monitored systems. To speed up your slower graphs, it’s helpful if the monitoring system can create and store new time series based on incoming data; then it can precompute answers to common queries. Calculations Support for calculations can span a variety of use cases, across a range of complexities. At a minimum, you’ll probably want your system to retain data over a multimonth time frame . Without a long-term view of your data, you cannot analyze long-term trends like system growth. In terms of granularity, summary data (i.e., aggregated data that you can’t drill down into) is sufficient to facilitate growth planning. Retaining all detailed individual metrics may help with answering questions like, “Has this unusual behavior happened before?” However, the data might be expensive to store or impractical to retrieve. The metrics you retain about events or resource consumption should ideally be monotonically incrementing counters. Using counters, your monitoring system can calculate windowed functions over time—for example, to report the rate of requests per second from that counter. Computing these rates over a longer window (up to a month) allows you to implement the building blocks for SLO burn-based alerting (see Alerting on SLOs ). Finally, support for a more complete range of statistical functions can be useful because trivial operations may mask bad behavior. A monitoring system that supports computing percentiles (i.e., 50th, 95th, 99th percentiles) when recording latency will let you see if 50%, 5%, or 1% of your requests are too slow, whereas the arithmetic mean can only tell you—without specifics—that the request time is slower. Alternatively, if your system doesn’t support computing percentiles directly, you can achieve this by: Obtaining a mean value by summing the seconds spent in requests and dividing by the number of requests Logging every request and computing the percentile values by scanning or sampling the log entries You might want to record your raw metric data in a separate system for offline analysis—for example, to use in weekly or monthly reports, or to perform more intricate calculations that are too difficult to compute in your monitoring system. Interfaces A robust monitoring system should allow you to concisely display time-series data in graphs, and also to structure data in tables or a range of chart styles. Your dashboards will be primary interfaces for displaying monitoring, so it’s important that you choose formats that most clearly display the data you care about. Some options include heatmaps, histograms, and logarithmic scale graphs. You’ll likely need to offer different views of the same data based upon audience; high-level management may want to view quite different information than SREs. Be specific about creating dashboards that make sense to the people consuming the content. For each set of dashboards, displaying the same types of data consistently is valuable for communication. You might need to graph information across different aggregations of a metric—such as machine type, server version, or request type—in real time. It’s a good idea for your team to be comfortable with performing ad hoc drill-downs on your data. By slicing your data according to a variety of metrics, you can look for correlations and patterns in the data when you need it. Alerts It’s helpful to be able to classify alerts: multiple categories of alerts allow for proportional responses. The ability to set different severity levels for different alerts is also useful: you might file a ticket to investigate a low rate of errors that lasts more than an hour, while a 100% error rate is an emergency that deserves immediate response. Alert suppression functionality lets you avoid unnecessary noise from distracting on-call engineers. For example: When all nodes are experiencing the same high rate of errors, you can alert just once for the global error rate instead of sending an individual alert for every single node. When one of your service dependencies has a firing alert (e.g., a slow backend), you don’t need to alert for error rates of your service. You also need to be able to ensure alerts are no longer suppressed once the event is over. The level of control you require over your system will dictate whether you use a third-party monitoring service or deploy and run your own monitoring system. Google developed its own monitoring system in-house, but there are plenty of open source and commercial monitoring systems available. Sources of Monitoring Data Your choice of monitoring system(s) will be informed by the specific sources of monitoring data you’ll use. This section discusses two common sources of monitoring data: logs and metrics. There are other valuable monitoring sources that we won’t cover here, like distributed tracing and runtime introspection. Metrics are numerical measurements representing attributes and events, typically harvested via many data points at regular time intervals. Logs are an append-only record of events. This chapter’s discussion focuses on structured logs that enable rich query and aggregation tools as opposed to plain-text logs. Google’s logs-based systems process large volumes of highly granular data. There’s some inherent delay between when an event occurs and when it is visible in logs. For analysis that’s not time-sensitive, these logs can be processed with a batch system, interrogated with ad hoc queries, and visualized with dashboards. An example of this workflow would be using Cloud Dataflow to process logs, BigQuery for ad hoc queries, and Data Studio for the dashboards. By contrast, our metrics-based monitoring system, which collects a large number of metrics from every service at Google, provides much less granular information, but in near real time. These characteristics are fairly typical of other logs- and metrics-based monitoring systems, although there are exceptions, such as real-time logs systems or high-cardinality metrics. Our alerts and dashboards typically use metrics. The real-time nature of our metrics-based monitoring system means that engineers can be notified of problems very rapidly. We tend to use logs to find the root cause of an issue , as the information we need is often not available as a metric. When reporting isn’t time-sensitive, we often generate detailed reports using logs processing systems because logs will nearly always produce more accurate data than metrics. If you’re alerting based on metrics, it might be tempting to add more alerting based on logs—for example, if you need to be notified when even a single exceptional event happens. We still recommend metrics-based alerting in such cases: you can increment a counter metric when a particular event happens, and configure an alert based on that metric’s value. This strategy keeps all alert configuration in one place, making it easier to manage (see Managing Your Monitoring System ). Examples The following real-world examples illustrate how to reason through the process of choosing between monitoring systems. Move information from logs to metrics Problem. The HTTP status code is an important signal to App Engine customers debugging their errors. This information was available in logs, but not in metrics. The metrics dashboard could provide only a global rate of all errors, and did not include any information about the exact error code or the cause of the error. As a result, the workflow to debug an issue involved: Looking at the global error graph to find a time when an error occurred. Reading log files to look for lines containing an error. Attempting to correlate errors in the log file to the graph. The logging tools did not give a sense of scale, making it hard to know if an error seen in one log line was occurring frequently. The logs also contained many other irrelevant lines, making it hard to track down the root cause. Proposed solution. The App Engine dev team chose to export the HTTP status code as a label on the metric (e.g., requests_total{status=404} versus requests_total​{status=500} ). Because the number of different HTTP status codes is relatively limited, this did not increase the volume of metric data to an impractical size, but did make the most pertinent data available for graphing and alerting. Outcome. This new label meant the team could upgrade the graphs to show separate lines for different error categories and types. Customers could now quickly form conjectures about possible problems based on the exposed error codes. We could now also set different alerting thresholds for client and server errors, making the alerts trigger more accurately. Improve both logs and metrics Problem. One Ads SRE team maintained ~50 services, which were written in a number of different languages and frameworks. The team used logs as the canonical source of truth for SLO compliance. To calculate the error budget, each service used a logs processing script with many service-specific special cases. Here’s an example script to process a log entry for a single service: If the HTTP status code was in the range (500, 599) AND the 'SERVER ERROR' field of the log is populated AND DEBUG cookie was not set as part of the request AND the url did not contain '/reports' AND the 'exception' field did not contain 'com.google.ads.PasswordException' THEN increment the error counter by 1 These scripts were hard to maintain and also used data that wasn’t available to the metrics-based monitoring system. Because metrics drove alerts, sometimes the alerts would not correspond to user-facing errors. Every alert required an explicit triage step to determine if it was user-facing, which slowed down response time. Proposed solution. The team created a library that hooked into the logic of the framework languages of each application. The library decided if the error was impacting users at request time. The instrumentation wrote this decision in logs and exported it as a metric at the same time, improving consistency. If the metric showed that the service had returned an error, the logs contained the exact error, along with request-related data to help reproduce and debug the issue. Correspondingly, any SLO-impacting error that manifested in the logs also changed the SLI metrics, which the team could then alert on. Outcome. By introducing a uniform control surface across multiple services, the team reused tooling and alerting logic instead of implementing multiple custom solutions. All services benefited from removing the complicated, service-specific logs processing code, which resulted in increased scalability. Once alerts were directly tied to SLOs, they were more clearly actionable, so the false-positive rate decreased significantly. Keep logs as the data source Problem. While investigating production issues, one SRE team would often look at the affected entity IDs to determine user impact and root cause. As with the earlier App Engine example, this investigation required data that was available only in logs. The team had to perform one-off log queries for this while they were responding to incidents. This step added time to incident recovery: a few minutes to correctly put together the query, plus the time to query the logs. Proposed solution. The team initially debated whether a metric should replace their log tools. Unlike in the App Engine example, the entity ID could take on millions of different values, so it would not be practical as a metric label. Ultimately, the team decided to write a script to perform the one-off log queries they needed, and documented which script to run in the alert emails. They could then copy the command directly into a terminal if necessary. Outcome. The team no longer had the cognitive load of managing the correct one-off log query. Accordingly, they could get the results they needed faster (although not as quickly as a metrics-based approach). They also had a backup plan: they could run the script automatically as soon as an alert triggered, and use a small server to query the logs at regular intervals to constantly retrieve semifresh data. Managing Your Monitoring System Your monitoring system is as important as any other service you run. As such, it should be treated with the appropriate level of care and attention. Treat Your Configuration as Code Treating system configuration as code and storing it in the revision control system are common practices that provide some obvious benefits: change history, links from specific changes to your task tracking system, easier rollbacks and linting checks, 1 and enforced code review procedures. We strongly recommend also treating monitoring configuration as code (for more on configuration, see Configuration Design and Best Practices ). A monitoring system that supports intent-based configuration is preferable to systems that only provide web UIs or CRUD-style APIs. This configuration approach is standard for many open source binaries that only read a configuration file. Some third-party solutions like grafanalib enable this approach for components that are traditionally configured with a UI. Encourage Consistency Large companies with multiple engineering teams who use monitoring need to strike a fine balance: a centralized approach provides consistency, but on the other hand, individual teams may want full control over the design of their configuration. The right solution depends on your organization. Google’s approach has evolved over time toward convergence on a single framework run centrally as a service. This solution works well for us for a few reasons. A single framework enables engineers to ramp up faster when they switch teams, and makes collaboration during debugging easier. We also have a centralized dashboarding service, where each team’s dashboards are discoverable and accessible. If you easily understand another team’s dashboard, you can debug both your issues and theirs more quickly. If possible, make basic monitoring coverage effortless. If all your services 2 export a consistent set of basic metrics, you can automatically collect those metrics across your entire organization and provide a consistent set of dashboards. This approach means that any new component you launch automatically has basic monitoring. Many teams across your company—even nonengineering teams—can use this monitoring data. Prefer Loose Coupling Business requirements change, and your production system will look different a year from now. Similarly, your monitoring system needs to evolve over time as the services it monitors evolve through different patterns of failure. We recommend keeping the components of your monitoring system loosely coupled. You should have stable interfaces for configuring each component and passing monitoring data. Separate components should be in charge of collecting, storing, alerting, and visualizing your monitoring. Stable interfaces make it easier to swap out any given component for a better alternative. Splitting functionality into individual components is becoming popular in the open source world. A decade ago, monitoring systems like Zabbix combined all functions into a single component. Modern design usually involves separating collection and rule evaluation (with a solution like Prometheus server ), long-term time series storage ( InfluxDB ), alert aggregation ( Alertmanager ), and dashboarding ( Grafana ). As of this writing, there are at least two popular open standards for instrumenting your software and exposing metrics: statsd The metric aggregation daemon initially written by Etsy and now ported to a majority of programming languages. Prometheus An open source monitoring solution with a flexible data model, support for metric labels, and robust histogram functionality. Other systems are now adopting the Prometheus format, and it is being standardized as OpenMetrics . A separate dashboarding system that can use multiple data sources provides a central and unified overview of your service. Google recently saw this benefit in practice: our legacy monitoring system (Borgmon 3 ) combined dashboards in the same configuration as alerting rules. While migrating to a new system ( Monarch ), we decided to move dashboarding into a separate service ( Viceroy ). Because Viceroy was not a component of Borgmon or Monarch, Monarch had fewer functional requirements. Since users could use Viceroy to display graphs based on data from both monitoring systems, they could gradually migrate from Borgmon to Monarch. Metrics with Purpose Alerting on SLOs covers how to monitor and alert using SLI metrics when a system’s error budget is under threat. SLI metrics are the first metrics you want to check when SLO-based alerts trigger. These metrics should appear prominently in your service’s dashboard, ideally on its landing page. When investigating the cause of an SLO violation, you will most likely not get enough information from the SLO dashboards. These dashboards show that you are violating the SLO, but not necessarily why. What other data should the monitoring dashboards display? We’ve found the following guidelines helpful in implementing metrics. These metrics should provide reasonable monitoring that allows you to investigate production issues and also provide a broad range of information about your service. Intended Changes When diagnosing an SLO-based alert, you need to be able to move from alerting metrics that notify you of user-impacting issues to metrics that tell you what is causing these issues. Recent intended changes to your service might be at fault. Add monitoring that informs you of any changes in production. 4 To determine the trigger, we recommend the following: Monitor the version of the binary. Monitor the command-line flags, especially when you use these flags to enable and disable features of the service. If configuration data is pushed to your service dynamically, monitor the version of this dynamic configuration. If any of these pieces of the system aren’t versioned, you should be able to monitor the timestamp at which it was last built or packaged. When you’re trying to correlate an outage with a rollout, it’s much easier to look at a graph/dashboard linked from your alert than to trawl through your CI/CD (continuous integration/continuous delivery) system logs after the fact. Dependencies Even if your service didn’t change, any of its dependencies might change or have problems, so you should also monitor responses coming from direct dependencies. It’s reasonable to export the request and response size in bytes, latency, and response codes for each dependency. When choosing the metrics to graph, keep the four golden signals in mind. You can use additional labels on the metrics to break them down by response code, RPC (remote procedure call) method name, and peer job name. Ideally, you can instrument the lower-level RPC client library to export these metrics once, instead of asking each RPC client library to export them. 5 Instrumenting the client library provides more consistency and allows you to monitor new dependencies for free. You will sometimes come across dependencies that offer a very narrow API, where all functionality is available via a single RPC called Get , Query , or something equally unhelpful, and the actual command is specified as arguments to this RPC. A single instrumentation point in the client library falls short with this type of dependency: you will observe a high variance in latency and some percentage of errors that may or may not indicate that some part of this opaque API is failing entirely. If this dependency is critical, you have a couple of options to monitor it well: Export separate metrics to tailor for the dependency, so that the metrics can unpack requests they receive to get at the actual signal. Ask the dependency owners to perform a rewrite to export a broader API that supports separate functionality split across separate RPC services and methods. Saturation Aim to monitor and track the usage of every resource the service relies upon. Some resources have hard limits you cannot exceed, like RAM, disk, or CPU quota allocated to your application. Other resources—like open file descriptors, active threads in any thread pools, waiting times in queues, or the volume of written logs—may not have a clear hard limit but still require management. Depending on the programming language in use, you should monitor additional resources: In Java: The heap and metaspace size, and more specific metrics depending on what type of garbage collection you’re using In Go: The number of goroutines The languages themselves provide varying support to track these resources. In addition to alerting on significant events as described in Alerting on SLOs you might also need to set up alerting that fires when you approach exhaustion for specific resources, such as: When the resource has a hard limit When crossing a usage threshold causes performance degradation You should have monitoring metrics to track all resources—even resources that the service manages well. These metrics are vital in capacity and resource planning. Status of Served Traffic It’s a good idea to add metrics or metric labels that allow the dashboards to break down served traffic by status code (unless the metrics your service uses for SLI purposes already include this information). Here are some recommendations: For HTTP traffic, monitor all response codes, even if they don’t provide enough signal for alerting, because some can be triggered by incorrect client behavior. If you apply rate limits or quota limits to your users, monitor aggregates of how many requests were denied due to lack of quota. Graphs of this data can help you identify when the volume of errors changes noticeably during a production change. Implementing Purposeful Metrics Each exposed metric should serve a purpose. Resist the temptation of exporting a handful of metrics just because they are easy to generate. Instead, think about how these metrics will be used. Metric design, or lack thereof, has implications. Ideally, metric values used for alerting change dramatically only when the system enters a problem state, and don’t change when a system is operating normally. On the other hand, metrics for debugging don’t have these requirements—they’re intended to provide insight about what is happening when alerts are triggered. Good debugging metrics will point at some aspect of the system that’s potentially causing the issues. When you write a postmortem, think about which additional metrics would have allowed you to diagnose the issue faster. Testing Alerting Logic In an ideal world, monitoring and alerting code should be subject to the same testing standards as code development. While Prometheus developers are discussing developing unit tests for monitoring , there is currently no broadly adopted system that allows you to do this. At Google, we test our monitoring and alerting using a domain-specific language that allows us to create synthetic time series. We then write assertions based upon the values in a derived time series , or the firing status and label presence of specific alerts. Monitoring and alerting is often a multistage process, which therefore calls for multiple families of unit tests. While this area remains largely underdeveloped, should you want to implement monitoring testing at some point, we recommend a three-tiered approach, as shown in Figure 4-1 . Figure 4-1. Monitoring testing environment tiers Binary reporting: Check that the exported metric variables change in value under certain conditions as expected. Monitoring configurations: Make sure that rule evaluation produces expected results, and that specific conditions produce the expected alerts. Alerting configurations: Test that generated alerts are routed to a predetermined destination, based on alert label values. If you can’t test your monitoring via synthetic means, or there’s a stage of your monitoring you simply can’t test, consider creating a running system that exports well-known metrics, like number of requests and errors. You can use this system to validate derived time series and alerts. It’s very likely that your alerting rules will not fire for months or years after you configure them, and you need to have confidence that when the metric passes a certain threshold, the correct engineers will be alerted with notifications that make sense. Conclusion Because the SRE role is responsible for the reliability of systems in production, SREs are often required to be intimately familiar with a service’s monitoring system and its features. Without this knowledge, SREs might not know where to look, how to identify abnormal behavior, or how to find the information they need during an emergency. We hope that by pointing out monitoring system features we find useful and why, we can help you evaluate how well your monitoring strategy fits your needs, explore some additional features you might be able to leverage, and consider changes you might want to make. You’ll probably find it useful to combine some source of metrics and logging in your monitoring strategy; the exact mix you need is highly context-dependent. Make sure to collect metrics that serve a particular purpose. That purpose may be to enable better capacity planning, assist in debugging, or directly notify you about problems. Once you have monitoring in place, it needs to be visible and useful. To this end, we also recommend testing your monitoring setup. A good monitoring system pays dividends. It is well worth the investment to put substantial thought into what solutions best meet your needs, and to iterate until you get it right. 1 For example, using promtool to verify that your Prometheus config is syntactically correct. 2 You can export basic metrics via a common library: an instrumentation framework like OpenCensus , or a service mesh like Istio . 3 See Chapter 10 of Site Reliability Engineering for Borgmon’s concepts and structure. 4 This is one case where monitoring via logs is appealing, particularly because production changes are relatively infrequent. Whether you use logs or metrics, these changes should be surfaced in your dashboards so they’re easily accessible for debugging production issues. 5 See https://opencensus.io/ for a set of libraries that provides this. Previous Chapter 3 - SLO Engineering Case Studies Next Chapter 5 - Alerting on SLOs Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://fr-fr.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT317rgIlXz-cdvSK7V3LNAW1GKivKoxP492C7CHQSWKr68LB-a6DQY1S5UOUfUGRFieCiBqleSokpucOfQfOC7nQYePlwBPeZ9TfjNHiOTH6yXDb7ru42jbguKFKkBPmym9lwctwh7bOFDY
Facebook Facebook Adresse e-mail ou téléphone Mot de passe Informations de compte oubliées ? Créer un compte Cette fonction est temporairement bloquée Cette fonction est temporairement bloquée Il semble que vous ayez abusé de cette fonctionnalité en l’utilisant trop vite. Vous n’êtes plus autorisé à l’utiliser. Back Français (France) 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Deutsch S’inscrire Se connecter Messenger Facebook Lite Vidéo Meta Pay Boutique Meta Meta Quest Ray-Ban Meta Meta AI Plus de contenu Meta AI Instagram Threads Centre d’information sur les élections Politique de confidentialité Centre de confidentialité À propos Créer une publicité Créer une Page Développeurs Emplois Cookies Choisir sa publicité Conditions générales Aide Importation des contacts et non-utilisateurs Paramètres Historique d’activité Meta © 2026
2026-01-13T09:29:19
https://git-scm.com/book/bg/v2/%d0%9a%d0%bb%d0%be%d0%bd%d0%be%d0%b2%d0%b5-%d0%b2-Git-%d0%a3%d0%bf%d1%80%d0%b0%d0%b2%d0%bb%d0%b5%d0%bd%d0%b8%d0%b5-%d0%bd%d0%b0-%d0%bf%d1%80%d0%be%d0%b5%d0%ba%d1%82
Git - Управление на проект About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Начало 1.1 За Version Control системите 1.2 Кратка история на Git 1.3 Какво е Git 1.4 Конзолата на Git 1.5 Инсталиране на Git 1.6 Първоначална настройка на Git 1.7 Помощна информация в Git 1.8 Обобщение 2. Основи на Git 2.1 Създаване на Git хранилище 2.2 Запис на промени в хранилището 2.3 Преглед на историята на действията 2.4 Възстановяване на направени действия 2.5 Работа с отдалечени хранилища 2.6 Тагове в Git 2.7 Псевдоними в Git 2.8 Обобщение 3. Клонове в Git 3.1 Накратко за разклоненията 3.2 Основи на клоновете код и сливането 3.3 Управление на клонове 3.4 Стратегии за работа с клонове код 3.5 Отдалечени клонове 3.6 Управление на проект 3.7 Обобщение 4. GitHub 4.1 Създаване и настройка на акаунт 4.2 Как да сътрудничим в проект 4.3 Управление на проект 4.4 Управление на организация 4.5 Автоматизиране с GitHub 4.6 Обобщение 5. Git инструменти 5.1 Избор на къмити 5.2 Интерактивно индексиране 5.3 Stashing и Cleaning 5.4 Подписване на вашата работа 5.5 Търсене 5.6 Манипулация на историята 5.7 Мистерията на командата Reset 5.8 Сливане за напреднали 5.9 Rerere 5.10 Дебъгване с Git 5.11 Подмодули 5.12 Пакети в Git (Bundling) 5.13 Заместване 5.14 Credential Storage система 5.15 Обобщение 6. Настройване на Git 6.1 Git конфигурации 6.2 Git атрибути 6.3 Git Hooks 6.4 Примерна Git-Enforced политика 6.5 Обобщение 7. Git и други системи 7.1 Git като клиент 7.2 Миграция към Git 7.3 Обобщение 8. Git на ниско ниво 8.1 Plumbing и Porcelain команди 8.2 Git обекти 8.3 Git референции 8.4 Packfiles 8.5 Refspec спецификации 8.6 Транспортни протоколи 8.7 Поддръжка и възстановяване на данни 8.8 Environment променливи 8.9 Обобщение 9. Приложение A: Git в други среди 9.1 Графични интерфейси 9.2 Git във Visual Studio 9.3 Git във Visual Studio Code 9.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 9.5 Git в Sublime Text 9.6 Git в Bash 9.7 Git в Zsh 9.8 Git в PowerShell 9.9 Обобщение 10. Приложение B: Вграждане на Git в приложения 10.1 Git от команден ред 10.2 Libgit2 10.3 JGit 10.4 go-git 10.5 Dulwich A1. Приложение C: Git команди A1.1 Настройки и конфигурация A1.2 Издърпване и създаване на проекти A1.3 Snapshotting A1.4 Клонове и сливане A1.5 Споделяне и обновяване на проекти A1.6 Инспекция и сравнение A1.7 Дебъгване A1.8 Patching A1.9 Email команди A1.10 Външни системи A1.11 Административни команди A1.12 Plumbing команди 2nd Edition 3.6 Клонове в Git - Управление на проект Управление на проект След като разгледахме как се допринася в проект по ефективен начин, вероятно ще ви интересува и обратната страна, как да поддържаме собствен такъв. Това може да включва приемане и прилагане на пачове генерирани през format-patch и изпратени до вас или пък интегриране на промени в отдалечени клонове за хранилища добавени като remotes към проекта ви. Независимо дали поддържате canonical хранилище или искате да помогнете проверявайки и одобрявайки пачове, трябва да знаете как да приемате работа от колегите ви по начин ясен за тях и удобен за вас във времето. Работа в Topic клонове Когато мислите да интегрирате новополучени промени, добра идея е да ги изпробвате в topic клон  — временен такъв създаден специално за теста. По такъв начин е лесно да поправите пач индивидуално и да го зарежете, ако той не работи, докато имате време да го разгледате по-подробно. Ако създадете клон с име съответстващо на темата на изпратената работа, например ruby_client или нещо подобно, можете лесно да го запомните и по-късно да се върнете в него. Поддръжникът на Git проекта например, дори се стреми да използва namespaces за тези имена — като sc/ruby_client , където sc е съкращение за човека, който е изпратил работата. Както помним, създаването на клон базиран на master се прави така: $ git branch sc/ruby_client master Или, ако искате да превключите към него веднага, може да използвате checkout -b варианта: $ git checkout -b sc/ruby_client master Сега сте готови да добавите новата работа, която сте получили, в този нов topic клон и да решите дали искате да я слеете в long-term клоновете си. Прилагане на пачове от Email Ако сте получили пач по имейл, ще трябва да го приложите в topic клона и да го изпробвате. Това се прави с помощта на една от двете команди git apply или git am . Прилагане на пач с apply Ако сте получили пача от някого, който го е генерирал с git diff или някакъв вариант на Unix diff инструмента (което не е препоръчително, вижте следващата секция), можете да го приложите с командата git apply . Ако сте записали пача в /tmp/patch-ruby-client.patch : $ git apply /tmp/patch-ruby-client.patch Това модифицира файловете в работната директория. Командата е почти идентична на patch -p1 , въпреки че е по-параноична и приема по-малко fuzzy matches от patch. Тя също така обработва добавянето, изтриването и преименуването на файлове, ако тези процеси са правилно описани в git diff формата, нещо което patch няма да направи. Освен това git apply използва модела “apply all or abort all”, където или всичко се прилага успешно или нищо не се прилага. За разлика от нея, patch може да прилага частично patchfiles, оставяйки работната ви директория в странен статус. git apply като цяло е много по-консервативна от patch . Процесът няма автоматично да ви направи къмит — след като завърши ще трябва сами да индексирате и къмитнете промените. Можете също да използвате git apply за да видите дали пачът ще се приложи коректно преди в действителност да се изпълни, командата е git apply --check с името на пача: $ git apply --check 0001-see-if-this-helps-the-gem.patch error: patch failed: ticgit.gemspec:1 error: ticgit.gemspec: patch does not apply Ако няма изход, тогава пачът би следвало да се приложи чисто. В допълнение, командата завършва с код за грешка, ако проверката установи неуспех, така че можете да я използвате и в скриптове, ако желаете. Прилагане на пач с am Ако работата идва от напреднал Git потребител, който е наясно с format-patch , тогава работата ви се улеснява, защото пачът ще съдържа също и информация за автора и къмит съобщение. Ако можете, окуражавайте сътрудниците ви да използват format-patch вместо diff , когато генерират пачове за вас. Използвайте git apply само за legacy пачове и неща като това ( diff генерирани). За да приложите пач генериран с format-patch , по-добрият вариант е да използвате git am (наречена е am понеже се използва за "apply a series of patches from a mailbox"). Технически, git am е проектирана да чете mbox файл, който е plain-text формат за съхранение на едно или повече имейл съобщения в един файл. Изглежда по подобен начин: From 330090432754092d704da8e76ca5c05c198e71a8 Mon Sep 17 00:00:00 2001 From: Jessica Smith <jessica@example.com> Date: Sun, 6 Apr 2008 10:17:23 -0700 Subject: [PATCH 1/2] Add limit to log function Limit log functionality to the first 20 Това е началото на изхода от командата git format-patch , който видяхме в предната секция, също валиден mbox имейл формат. Ако някой ви е изпратил пача по пощата коректно с git send-email и го запишете в mbox формат, тогава можете да подадете на git am въпросния mbox файл и тя ще започне да прилага всички пачове, които намери. Ако имате имейл клиент способен да записва множество съобщения в mbox формат, тогава можете да запишете цялата серия пачове в един файл и след това да пуснете git am , която да ги приложи последователно. Обаче, ако някой е качил пач файл генериран с git format-patch към ticketing система или нещо подобно, можете да запишете файла локално и след това да го подадете на git am да го приложи: $ git am 0001-limit-log-function.patch Applying: Add limit to log function Може да видите, че той се е приложил чисто и автоматично се създава нов къмит за вас. Информацията за автора е взета от хедърите From и Date на имейла, а къмит съобщението от Subject хедъра и тялото му (частта преди пача). Например, ако този пач беше приложен от mbox примера отгоре, генерираният къмит би изглеждал така: $ git log --pretty=fuller -1 commit 6c5e70b984a60b3cecd395edd5b48a7575bf58e0 Author: Jessica Smith <jessica@example.com> AuthorDate: Sun Apr 6 10:17:23 2008 -0700 Commit: Scott Chacon <schacon@gmail.com> CommitDate: Thu Apr 9 09:19:06 2009 -0700 Add limit to log function Limit log functionality to the first 20 Commit реда индикира кой е човекът, който е приложил пача и времето, когато това е станало. Author от своя страна носи информация за създателя на пача и кога е създаден първоначално. Възможно е обаче пачът да не се прилага чисто. Може главният ви клон да се е разделил твърде много от клона, от който е направен пача или пък самият пач да зависи от друг такъв, който все още не сте приложили. В подобен случай git am ще прекъсне и ще ви попита как искате да продължите: $ git am 0001-see-if-this-helps-the-gem.patch Applying: See if this helps the gem error: patch failed: ticgit.gemspec:1 error: ticgit.gemspec: patch does not apply Patch failed at 0001. When you have resolved this problem run "git am --resolved". If you would prefer to skip this patch, instead run "git am --skip". To restore the original branch and stop patching run "git am --abort". Командата поставя маркери за конфликт в съответните файлове, както това става при конфликтно сливане или пребазиране. Разрешаването на конфликтите също е аналогично — редактирате файла, индексирате го и след това изпълнявате git am --resolved за да продължите цикъла със следващия пач: $ (fix the file) $ git add ticgit.gemspec $ git am --resolved Applying: See if this helps the gem Ако искате Git да се опита да разреши конфликта по малко по-интелигентен начин, може да подадете флага -3 и Git ще се опита да направи three-way сливане. Тази опция е забранена по подразбиране, защото няма да работи в случай, че пачът рапортува, че е базиран на къмит, който не присъства в хранилището ви. Ако обаче имате този къмит (ако пачът например е базиран на публичен такъв), тогава -3 опцията е много по-гъвкава в прилагането на конфликтен пач: $ git am -3 0001-see-if-this-helps-the-gem.patch Applying: See if this helps the gem error: patch failed: ticgit.gemspec:1 error: ticgit.gemspec: patch does not apply Using index info to reconstruct a base tree... Falling back to patching base and 3-way merge... No changes -- Patch already applied. В този случай, без -3 флага пачът щеше се счита за конфликтен. Но понеже той е подаден, пачът се прилага чисто. Ако прилагате множество от пачове от mbox, можете също така да пуснете am командата в интерактивен режим, при което тя ще спира на всеки пач и ще ви пита дали желаете да го приложи: $ git am -3 -i mbox Commit Body is: -------------------------- See if this helps the gem -------------------------- Apply? [y]es/[n]o/[e]dit/[v]iew patch/[a]ccept all Това е полезно при много пачове, защото можете да видите всеки от тях, ако сте забравили за какво е или да го откажете, ако вече е приложен. Когато всички пачове са приложени и къмитнати в topic клона, може да изберете дали и кога да ги интегрирате в long-term клон. Извличане от отделечени клонове Работата на вашите колеги може и да не идва по имейл. Те могат да имат свои собствени онлайн хранилища, да са извършили много промени и да са ви пратили URL до хранилището и клона, където промените се пазят. В случаи като този, можете да добавите отдалечените хранилища и да сливате локално вместо да пачвате. Ако Jessica ви съобщи, че има нова функционалност в клона ruby-client на хранилището ѝ, можете да тествате бързо добавяйки го като remote референция и извличайки клона локално: $ git remote add jessica git://github.com/jessica/myproject.git $ git fetch jessica $ git checkout -b rubyclient jessica/ruby-client Ако впоследствие тя ви съобщи за нова функционалност в отделен клон, можете директно да направите fetch и checkout , защото вече отдалеченото хранилище е конфигурирано при вас. Това е най-полезно, ако работите сравнително често с даден колега. Ако някой иска да ви изпрати само единичен пач и няма намерение за продължително сътрудничество, тогава имейл методът вероятно е по-бързо решение и няма нужда колегата ви да поддържа онлайн хранилище. Освен това едва ли бихте искали да имате стотици remotes, всяко от които допринася само с един-два пача. Обаче, скриптовете и хостнатите публични услуги могат да улеснят това — зависи най-вече от това как разработвате вие и как колегите ви. Друго предимство на този подход е, че получавате историята на къмитите. Въпреки, че може да имате merge проблеми, вие все пак знаете на коя точка от историята ви е базирана работата на колегата и правилното three-way сливане е по подразбиране вместо да трябва да подавате -3 и да се надявате, че пачът е бил генериран на публичен къмит, до който имате достъп. Ако не работите често с определен колега, но въпреки това искате да получавате работата му по този начин, можете да подадете на git pull директно адреса на отдалеченото хранилище. По този начин правите one-time pull и не съхранявате URL-а като remote референция: $ git pull https://github.com/onetimeguy/project From https://github.com/onetimeguy/project * branch HEAD -> FETCH_HEAD Merge made by the 'recursive' strategy. Изследване на промените Сега имате topic клон с новата работа от колега. На този етап може да определите какво бихте искали да правите с нея. Тази секция преглежда няколко команди, които ви помагат да разберете какво точно ще бъде въведено в главния клон, ако решите да направите сливане на topic клона. Често е полезно да имате представа за всички къмити, които са налични във временния клон, но все още не са в главния. Можете да извадите къмитите в master клона добавяйки опцията --not преди името му. Това прави същото като формата master..contrib , който видяхме по-рано. Например, ако вашият колега ви изпрати два пача и създадете клон contrib , в който сте ги приложили, може да изпълните това: $ git log contrib --not master commit 5b6235bd297351589efc4d73316f0a68d484f118 Author: Scott Chacon <schacon@gmail.com> Date: Fri Oct 24 09:53:59 2008 -0700 See if this helps the gem commit 7482e0d16d04bea79d0dba8988cc78df655f16a0 Author: Scott Chacon <schacon@gmail.com> Date: Mon Oct 22 19:38:36 2008 -0700 Update gemspec to hopefully work better Ако ви трябват и промените, които въвежда всеки къмит, можете да подадете флага -p към git log и тя ще ви изведе в допълнение diff информацията за всеки от къмитите. За да видите пълен diff на това какво ще се случи, ако слеете този topic клон с друг, може да се наложи да използвате малък трик, за да получите коректните резултати. Може би си мислите за това: $ git diff master Командата действително извежда diff, но той може да е заблуждаващ. Ако master клонът се е придвижил напред след като сте създали topic клона от него, тогава ще получите изглеждащи странно резултати. Това се случва, защото Git директно сравнява snapshot-а на последния къмит от topic клона, в който сте и snapshot-а на най-новия къмит от master клона. Ако например сте добавили ред във файл от master след създаването на topic клона, директното сравнение на snapshot-ите ще изглежда така сякаш при сливане topic клона ще изтрие този ред. Ако master е директен родител на topic клона това не е проблем. Но ако двете истории са се разклонили, тогава diff изходът ще показва, че добавяте всичките нови промени от topic и изтривате всичко уникално за master клона. Но това, което наистина искате да видите, са промените добавени в topic — работата, която ще бъде въведена при сливането му в master . Начинът да получите този резултат е да накарате Git да сравни последния къмит в topic клона с първия общ предшественик от master . Технически това може да направите като изрично установите кой е този предшественик и след това изпълните diff към него: $ git merge-base contrib master 36c7dba2c95e6bbb78dfa822519ecfec6e1ca649 $ git diff 36c7db или по-съкратено: $ git diff $(git merge-base contrib master) И понеже нито един от тези два начина не е достатъчно удобен, Git осигурява още едно съкратено изписване за същия резултат: triple-dot синтаксиса. В контекста на командата git diff , можете да поставите три точки между имената на два клона и тогава Git ще изпълни diff между последния къмит на клона отдясно в израза (contrib) и най-актуалния му общ предшественик от клона вляво (master): $ git diff master...contrib Така получавате само работата, която текущия topic клон въвежда от последния общ родителски къмит в master . Това е един много полезен синтаксис и си заслужава да се запомни наизуст. Интегриране на получена работа Когато работата в topic клона е готова да бъде изпратена към един или няколко mainline клонове, възниква въпроса как да го направите. Дръг въпрос е какъв да е принципния работен процес, който да използвате за поддръжка на проекта си? Има множество варианти и ще разгледаме някои от тях. Merging работни процеси Един простичък процес за действие е просто да слеете цялата нова работа директно в master клона. В такъв сценарий се предполага, че имате master клон, който съдържа стабилен код. Когато получите промени в topic клон и смятате, че той е готов за интегриране или пък когато сте получили и проверили работата на колега, просто сливате нещата в master клона, изтривате topic клона и това се повтаря във времето. Например, ако имаме хранилище с работа в два клона, ruby_client и php_client , изглеждащи като в История с няколко topic клона и слеем ruby_client последван от php_client , историята накрая ще изглежда като След сливане на topic клон . Фигура 72. История с няколко topic клона Фигура 73. След сливане на topic клон Вероятно това е най-простият работен процес, но може да бъде проблематичен, ако си имате работа с по-големи или по-стабилни проекти, където искате да сте много внимателни с това, което въвеждате като промени. Ако проектът е особено важен, може да искате да използвате двустъпков цикъл на сливане. При такъв сценарий, имате два long-running клона, master и develop , при които master се обновява само, когато излезе много стабилна версия на проекта и промените преди това са интегрирани успешно в develop . И двата клона редовно се публикуват в публично хранилище. Всеки път, когато имате нов topic клон за сливане ( Преди сливане на topic клон ), вие го сливате в develop ( След сливане на topic клон ), след което тагвате нова версия и правите fast-forward на master към мястото, където сега е стабилния develop ( След нова версия на проекта ). Фигура 74. Преди сливане на topic клон Фигура 75. След сливане на topic клон Фигура 76. След нова версия на проекта По такъв начин, когато някой клонира хранилището, може да извлече или master и да компилира последната стабилна версия, или develop , който съдържа най-пресните новости. Можете допълнително да разширите тази концепция с отделен integrate клон, където цялата работа е слята заедно. След това, когато кодът в него е стабилен и преминава тестовете, го сливате в develop клона и ако времето покаже, че и той е наистина надежден, правите fast-forward на master . Large-Merging работни процеси Проектът Git има четири long-running клона: master , next , и seen (преди това известен като 'pu' — proposed updates) за нови промени, и maint за maintenance backports. Когато сътрудниците изпратят нова работа, тя се събира в topic клонове в хранилище на поддържащия проекта в маниер подобен на този, който описахме (вижте Управление на сложни серии от parallel contributed topic клонове. ). В този момент, topic клоновете се изпитват, за да се определи дали са безопасни и готови за използване или трябва да се поправят допълнително. Ако са добре, те се сливат в next и този клон се публикува, така че всички могат да изпробват новите промени както са интегрирани заедно. Фигура 77. Управление на сложни серии от parallel contributed topic клонове. Ако промените не са задоволителни, те вместо това се сливат в клона seen . Когато се установи, че са окончателно надеждни, промените биват интегрирани в master . След това next и seen се построяват отново от master . Това значи, че master винаги се мести напред, next от време на време се пребазира, а seen се пребазира още по-често: Фигура 78. Сливане на topic клонове в long-term integration клонове. Когато даден topic клон най-сетне се слее в master , той се изтрива от хранилището. Проектът Git също има и клон maint , който се прави от последната версия за да осигури backported пачове в случай, че се изисква maintenance release. Така при клониране на Git хранилището, имате четири клона, които може да използвате за да пробвате проекта в различните му състояния на разработка, в зависимост от това колко актуален искате да бъде или как искате да допринасяте към него. А поддържащият проекта има структуриран работен процес за управление на новите промени. Все пак, Git проектът е със специализиран работен процес. За да го разберете напълно, можете да погледнете Git Maintainer’s guide . Rebasing и Cherry-Picking работни процеси Други мениджъри на проекти предпочитат да пребазират или cherry-pick-ват нови промени на базата на master клона си, вместо да ги сливат. Целта е да поддържат почти линейна история. Ако се установи, че промените в topic клона са подходящи за интегриране, превключвате към него и изпълнявате rebase командата, за да възпроизведете промените на базата на текущия master клон (или от develop и т.н.). Ако това работи добре, можете да направите fast-forward на master клона и да запазите линейната история на проекта. Другият начин да преместите нова работа от един клон в друг е да я cherry-pick-нете. Процесът cherry-pick в Git е подобен на rebase за единичен къмит. Той взема пача, който е бил въведен в даден къмит и се опитва да го приложи повторно в текущия клон. Това е полезно, ако имате множество къмити в topic клон и искате да интегрирате само един от тях или ако къмитът е само един, но предпочитате да не изпълнявате rebase. Да кажем, че имате проект изглеждащ така: Фигура 79. Примерна история преди cherry-pick Ако искате да издърпате къмита e43a6 в master клона, можете да направите следното: $ git cherry-pick e43a6 Finished one cherry-pick. [master]: created a0a41a9: "More friendly message when locking the index fails." 3 files changed, 17 insertions(+), 3 deletions(-) Това въвежда същата промяна от e43a6 , но получавате нов SHA-1 хеш за къмита, защото приложената дата е различна. Сега историята изглежда така: Фигура 80. История след cherry-picking на къмит в topic клон Сега можете да премахнете topic клона и да изоставите къмитите, които не желаете да въвеждате. Rerere Ако ви се налага да правите често сливания и пребазирания или ако поддържате topic клон с по-дълъг живот, Git предлага във ваша услуга опцията известна като “rerere”. Терминът Rerere идва от “reuse recorded resolution” — това е начин за бързо прилагане на ръчно разрешаване на конфликти. Когато rerere е активна, Git ще съхранява набор от pre- и post-images от успешни сливания и ако установи, че е налице конфликт, който изглежда подобен на такъв, който е бил разрешен в миналото, автоматично ще използва записания начин за разрешаването му без да ви занимава с него. Тази функционалност идва в две части: конфигурационна настройка и команда. Настройката е rerere.enabled и е достатъчно да имате следното в глобалната конфигурация: $ git config --global rerere.enabled true Сега след като направите ръчно разрешаване на конфликт, резолюцията ще бъде запомнена в кеша и ще бъде използвана в бъдеще. Ако е необходимо, може да комуникирате с rerere кеша с командата git rerere . Когато тя се използва самостоятелно, Git проверява базата си данни с решения за конфликти и се опитва да намери съответствие (въпреки, че това се прави автоматично, ако rerere.enabled е true ). Също така има подкоманди за разглеждане на това какво ще се запише, за изтриване на конкретна резолюция от кеша и за изтриване на целия кеш. Ще разгледаме детайлно rerere в Rerere . Тагване на версии Ако сте решили да обявите стабилна версия на проекта, вероятно бихте желали да присъедините таг към нея, така че ако се налага по-късно да можете да я пресъздадете отново. Как да създадете таг видяхме в Основи на Git . Ако решите и да подпишете тага, тогава командата може да изглежда така: $ git tag -s v1.5 -m 'my signed 1.5 tag' You need a passphrase to unlock the secret key for user: "Scott Chacon <schacon@gmail.com>" 1024-bit DSA key, ID F721C45A, created 2009-02-09 Ако правите подписване, може да се наложи да се погрижите за евентуални проблеми свързани с разпространението на публичния PGP ключ, който се използва в процеса. Поддържащият проекта на Git е решил това включвайки публичния си ключ директно като blob в хранилището и добавяйки таг, който сочи към него. За да направите и вие нещо подобно, може да установите кой е желания ключ изпълнявайки gpg --list-keys : $ gpg --list-keys /Users/schacon/.gnupg/pubring.gpg --------------------------------- pub 1024D/F721C45A 2009-02-09 [expires: 2010-02-09] uid Scott Chacon <schacon@gmail.com> sub 2048g/45D02282 2009-02-09 [expires: 2010-02-09] След това можете директно да импортирате ключа в базата данни на Git като го експортирате и прекарате през git hash-object , което ще създаде и запише нов blob обект със съдържанието в Git и ще ви върне обратно SHA-1 хеша му: $ gpg -a --export F721C45A | git hash-object -w --stdin 659ef797d181633c87ec71ac3f9ba29fe5775b92 След като вече имате съдържанието на ключа в Git, можете да създадете таг, който сочи директно към него посредством въпросната SHA-1 стойност: $ git tag -a maintainer-pgp-pub 659ef797d181633c87ec71ac3f9ba29fe5775b92 Ако изпълните git push --tags , тогава тагът maintainer-pgp-pub ще бъде публикуван и споделен с всички останали. Ако някой от колегите ви желае да провери таг, може директно да извлече вашия PGP ключ като издърпа blob обекта от базата данни и го импортира в PGP: $ git show maintainer-pgp-pub | gpg --import Впоследствие този ключ може да се използва за проверка на всички подписани от вас тагове. Също така, в таг съобщението може да включите инструкции за това как да се проверява тага и всеки би могъл да ги прочете изпълнявайки git show <tag> . Генериране на номера на версии Git не увеличава монотонно числата в стил 'v123' или нещо подобно с всеки къмит. Ето защо, ако искате да имате нещо описателно с къмитите, бихте могли да изпълните git describe върху съответните къмити. В отговор, Git генерира стринг състоящ се от името на последния таг, който е по-ранен от този къмит, последвано от броя къмити след този таг и от частичната SHA-1 стойност на описвания къмит (добавя се и префиксът "g", означаващ Git): $ git describe master v1.6.2-rc1-20-g8c5b85c Така можете да експортирате snapshot или компилирана версия и да я именувате с разбираемо за хората описание. В действителност, ако компилирате Git от изходен код клониран от Git хранилището, git --version ще ви връща нещо като горното. Ако описвате къмит, който е бил тагнат директно, ще получите просто името на тага. Командата git describe по подразбиране изисква анотирани тагове (тези създадени с флаговете -a или -s ). Ако искате да се възползвате от lightweight (не-анотирани) такива, подайте ѝ опцията --tags . Може да използвате този стринг с командите git checkout или git show , въпреки че това зависи от SHA-1 стойността в края и може да не е валидно завинаги. Например, Linux kernel проекта напоследък премина от 8 на 10 символа за да гарантира SHA-1 уникалността на обектите и по този начин по-старите git describe изходни имена станаха невалидни. Подготовка за издаване на Release Допускаме, че искате да публикувате готова версия на продукта. Едно от нещата, които ще се наложи да направите, е да генерирате архив на най-новия snapshot за потребителите, които не ползват Git. Командата за това е git archive : $ git archive master --prefix='project/' | gzip > `git describe master`.tar.gz $ ls *.tar.gz v1.6.2-rc1-20-g8c5b85c.tar.gz Ако някой отвори този tarball ще получи актуалния snapshot на проекта ви в директория project . Можете също да съдадете zip архив по подобен начин, подавайки --format=zip опцията към git archive : $ git archive master --prefix='project/' --format=zip > `git describe master`.zip Сега имате tarball и zip версии на проекта, които можете да качите в уебсайт или да изпратите по имейла. Shortlog Време е да информирате хората от мейлинг листата ви за промените в проекта ви. Един удобен начин да получите бързо changelog на добавеното в проекта от последния път, когато сте изпратили известие, е командата git shortlog . Тя резюмира всички къмити в подадения ѝ обхват. Например, следното ще ви изведе обобщение на всички къмити от последния release насам, ако този release е бил наречен v1.0.1: $ git shortlog --no-merges master --not v1.0.1 Chris Wanstrath (6): Add support for annotated tags to Grit::Tag Add packed-refs annotated tag support. Add Grit::Commit#to_patch Update version and History.txt Remove stray `puts` Make ls_tree ignore nils Tom Preston-Werner (4): fix dates in history dynamic version method Version bump to 1.0.2 Regenerated gemspec for version 1.0.2 Както се вижда, получавате доклад за всички къмити от v1.0.1 групирани по автор и можете да го изпратите на когото трябва. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/resources/
Google SRE: Digital Library, Books and Resources Hub Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content The SRE resource library SRE represents a mindset, engineering practices, and a job function. Here you will find articles, videos, and guides to help you implement SRE principles and run reliable production systems. Explore All Resources Machine learning Start your journey by exploring Machine Learning in Production Read now Continue your journey by reading Efficient Machine Learning Inference Read now Extend your journey by watching Machine Learning at Scale Watch now SRE books Building Secure & Reliable Systems Can a system be considered truly reliable if it isn't fundamentally secure? Or can it be considered secure if it's unreliable? Security is crucial to the design and operation of scalable systems in production, as it plays an important part in product quality, performance, and availability. In this book, experts from Google share best practices to help your organization design scalable and reliable systems that are fundamentally secure. Edited by: Heather Adkins, Betsy Beyer, Paul Blankinship, Ana Oprea, Piotr Lewandowski, Adam Stubblefield View details The Site Reliability Workbook The Site Reliability Workbook is the hands-on companion to the bestselling Site Reliability Engineering book and uses concrete examples to show how to put SRE principles and practices to work. This book contains practical examples from Google’s experiences and case studies from Google’s Cloud Platform customers. Evernote, The Home Depot, The New York Times, and other companies outline hard-won experiences of what worked for them and what didn’t. Edited by: Betsy Beyer, Niall Richard Murphy, David K. Rensin, Kent Kawahara and Stephen Thorne View details Site Reliability Engineering Members of the SRE team explain how their engagement with the entire software lifecycle has enabled Google to build, deploy, monitor, and maintain some of the largest software systems in the world. Edited by: Betsy Beyer, Chris Jones, Jennifer Petoff and Niall Richard Murphy View details 1 2 3 Building Secure & Reliable Systems By: Heather Adkins, Betsy Beyer, Paul Blankinship, Ana Oprea, Piotr Lewandowski, Adam Stubblefield Can a system be considered truly reliable if it isn't fundamentally secure? Or can it be considered secure if it's unreliable? Security is crucial to the design and operation of scalable systems in production, as it plays an important part in product quality, performance, and availability. In this book, experts from Google share best practices to help your organization design scalable and reliable systems that are fundamentally secure. Buy from Google Books Read online The Site Reliability Workbook Edited by: Betsy Beyer, Niall Richard Murphy, David K. Rensin, Kent Kawahara and Stephen Thorne The Site Reliability Workbook is the hands-on companion to the bestselling Site Reliability Engineering book and uses concrete examples to show how to put SRE principles and practices to work. This book contains practical examples from Google’s experiences and case studies from Google’s Cloud Platform customers. Evernote, The Home Depot, The New York Times, and other companies outline hard-won experiences of what worked for them and what didn’t. Buy from Google Books Read online Site Reliability Engineering Edited by: Betsy Beyer, Chris Jones, Jennifer Petoff and Niall Richard Murphy Members of the SRE team explain how their engagement with the entire software lifecycle has enabled Google to build, deploy, monitor, and maintain some of the largest software systems in the world. Buy from Google Books Read online Service level objectives Begin by reading Implementing SLOs Read now Dig deeper by exploring Alerting on SLOs Read now Build your skills with Art of SLOs View Details Systems engineering Learn the basics by reading Introducing Non-Abstract Large System Design Read now Develop fundamentals by exploring SRE Classroom: Distribued ImageServer Read now Build advanced skills with this video workshop How to Design a Distributed System Watch now Explore resources Filter by: SLO Training NALSD Systems Engineering Program Management Incident Management Enterprise Security Distributed Systems Capacity Planning Toil Machine Learning Software Development Clear filters Article Why Heroism is Bad Read now Online Training SRE Fundamentals online course Sign up now Online Training SRE Classroom: Distributed PubSub View Details Online Training USENIX SREcon Conferences View Details Article Anatomy of an Incident Read now Article Enterprise Roadmap to SRE Read now Article Efficient Machine Learning Inference Read now Article Incident Metrics in SRE Read now Book Practical Guide to Cloud Migration Read now Article Reliable Data Processing with Minimal Toil Read now Online Training The Art of SLOs View Details Book Building Secure and Reliable Systems Read now Video Complexities of Capacity Management for Distributed Services Watch now Article SRE As A Team Sport Read now Article SRE Best Practices for Capacity Management Read now Article A Case Study in Community-Driven Software Adoption Read now Article Cloud CRE Production Maturity Assessment Read now Video Deploying SRE Training Best Practices to Production: How We SRE'ed Our SRE Education Program Watch now Article Engineering Reliable Mobile Applications: Strategies for Developing Resilient Native Mobile Applications Read now Article How to Get Started with Site Reliability Engineering Read now Article Multi-single-tenant architectures in Cloud Read now Article Reduce Toil Through Better Alerting Read now Article Taming Chaos: Preparing for Your Next Incident Read now Article The Calculus of Service Availability Read now Article A Brief Guide to Running ML Systems in Production Read now Article Canary Analysis Service Read now Article Case Studies in Infrastructure Change Management Read now Article Creating a Production Launch Plan Read now Video Google's Production Environment Tech Talk Watch now Video SRE YouTube Playlist Watch now Article The Cybersecurity Canon Hall of Famers Read now Book The Site Reliability Workbook Read now Article Training Site Reliability Engineers: What Your Organization Needs to Create a Learning Program Read now Article Postmortem Action Items: Plan the Work and Work the Plan Read now Article SLO Adoption and Usage in SRE Read now Book Site Reliability Engineering Read now Article Interrupt Reduction Projects Read now Article Invent More, Toil Less Read now Article (Un)Reliability Budgets: Finding Balance between Innovation and Reliability Read now Video 10 Years of Crashing Google Watch now Video Bad Machinery— Managing Interrupts Under Load Watch now Article Being an On-Call Engineer: A Google SRE Perspective Read now Article Capacity Planning Read now Video Continuous Pipelines at Google Watch now Video Distributed Consensus Algorithms for Extreme Reliability Watch now Video Error Budgets and Risks Watch now Article Hiring Site Reliability Engineers Read now Video How Container Clusters Like Kubernetes Change Operations Watch now Video Incident Analysis Watch now Article Managing Incidents Read now Article Reliable Cron across the Planet Read now Article The System Engineering Side of Site Reliability Engineering Read now Article BeyondCorp: A New Approach to Enterprise Security Read now Video Keys to SRE Watch now Article Making “Push on Green” a Reality Read now Video PostOps: A Non-Surgical Tale of Software, Fragility, and Reliability Watch now Article Weathering the Unexpected Read now Sorry, no available at the moment. Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
https://sre.google/workbook/preface/
Google SRE Handbook - Enacting the Principles of SRE Preface Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Preface When we wrote the original Site Reliability Engineering book , we had a goal: explain the philosophy and the principles of production engineering and operations at Google. The book was our attempt to share our teams’ best practices and lessons with the rest of the computing world. We assumed that the SRE book might appeal to a modest number of engineers working in large, reliability-conscious endeavors, and that both the quantity and the focus of the content would tend to limit the book’s appeal. As it turned out, we were happily mistaken on both counts. To our surprise and delight, the SRE book was a best-seller in computing for an exhilarating period after its release, and it was not just being sold or downloaded; it was being read . We received questions from around the world about the book, the team, the practices, and the outcomes. We were asked to speak about chapters, approaches, and incidents. We found ourselves in the unexpected position of having to turn down outside requests because we were out of cycles. Like most success disasters, the SRE book created an opportunity to respond with human effort (“Hire more people! Do more speaking engagements!”) or with something more scalable. And being SREs, it will surprise few readers that we gravitated toward the latter approach. We decided to write a second SRE book—one that expanded on the content we were most frequently being asked to speak about, and that addressed the most common questions readers had about the first book. Out of the many different questions, requests, and comments we received about the first SRE book, two themes were particularly interesting to us; if left unaddressed, they were barriers to putting SRE’s lessons to productive use. These themes are colloquially summarized as: Principles are interesting, but how do I turn them into practice in my project/team/company? SRE’s approach would not work for me; it is feasible only in Google’s culture, and makes sense only at Google’s scale. The purpose of this second SRE book is (a) to add more implementation detail to the principles outlined in the first volume, and (b) to dispel the idea that SRE is implementable only at “Google scale” or in “Google culture.” This volume is a companion to the previous work—not a new version. The two books should be taken together as a pair. You will get the most from this book if you’re already familiar with its predecessor. The first SRE book is available online for free . By design, the structure of this book roughly follows the structure of the first volume. We want you to be able to read the chapters in tandem. Each chapter in this volume assumes you’re familiar with its counterpart from the previous work; our goal is to allow you to jump back and forth between principle and practice as you go. That way, you can use both volumes as ongoing references. Next, a word about ethos: We heard from some readers that while describing Google’s journey toward better operations we concentrated too much on just us . Some readers suggested that we were too removed from the practicalities of the world outside Google, and failed to address the interaction of our ideas with the principles of DevOps. That’s an entirely fair criticism that we’ve tried to take to heart in this volume. However, we do think that the highly opinionated nature of SRE contributes to its usefulness as a discipline. To us that’s a feature, not a bug. We do not advocate that SRE is the only way (or even universally the best way) to build and operate highly reliable systems. It’s just the way that has been most successful for us. We’ll also spend a few words talking about how SRE and DevOps relate to each other. The important point to keep in mind is that they are not in conflict. We’d like to acknowledge up front that this volume is necessarily incomplete. The SRE discipline is a broad field even inside the confines of Google, and it is evolving even faster now that it’s practiced widely outside of Google. Rather than go broad and superficial, we focused this volume to answer the most requested implementation details from the first volume. Finally, this volume and its predecessor are not intended to be gospel. Please don’t treat them that way. Even after all these years, we’re still finding conditions and cases that cause us to tweak (or in some cases, replace) previously firmly held beliefs. SRE is a journey as much as it is a discipline. We hope that you enjoy what you read in these pages and find the book useful. Assembling it has been a labor of love. We’re delighted that there’s a growing and skilled community of SRE professionals with whom we can learn and improve. As always, your direct feedback is much appreciated. It teaches us something valuable every time you contribute it. How to Read This Book This book is the companion volume to Google’s first book, Site Reliability Engineering . To get the most out of this volume, we recommend that you have read, or can refer to, the first SRE book (available to read online for free at https://sre.google ).The two works complement each other in the following ways: The previous work was an introduction to principles and philosophy. This volume concentrates on how those principles are applied. (In a few areas—particularly configuration management and canarying—we also cover some new ground to provide background for the practical treatment of other subjects.) The earlier volume concentrated exclusively on how SRE is practiced at Google. This work includes perspectives from a number of other firms—from traditional enterprises (including The Home Depot and the New York Times ) to digital natives (Evernote, Spotify, and others). The first book didn’t directly refer to the larger operations community—especially DevOps—whereas this book speaks directly to how SRE and DevOps relate to each other. This volume assumes that you will bounce between this volume and its predecessor. You might, for example, read Chapter 4, “Service Level Objectives” in the first book and then read its implementation complement ( Implementing SLOs ) in this volume. This book assumes that every chapter is just the starting point for a longer discussion and journey. Accordingly, this book is intended to be a conversation starter rather than the last word. —The Editors Conventions Used in This Book The following typographical conventions are used in this book: Italic Indicates new terms, URLs, email addresses, filenames, and file extensions. Constant width Used for program listings, as well as within paragraphs to refer to program elements such as variable or function names, databases, data types, environment variables, statements, and keywords. Constant width bold Shows commands or other text that should be typed literally by the user. Constant width italic Shows text that should be replaced with user-supplied values or by values determined by context. Tip This element signifies a tip or suggestion. Note This element signifies a general note. Warning This element indicates a warning or caution. Using Code Examples Supplemental material (code examples, exercises, etc.) is available for download at https://g.co/SiteReliabilityWorkbookMaterials . This book is here to help you get your job done. In general, if example code is offered with this book, you may use it in your programs and documentation. You do not need to contact us for permission unless you’re reproducing a significant portion of the code. For example, writing a program that uses several chunks of code from this book does not require permission. Selling or distributing a CD-ROM of examples from O’Reilly books does require permission. Answering a question by citing this book and quoting example code does not require permission. Incorporating a significant amount of example code from this book into your product’s documentation does require permission. We appreciate, but do not require, attribution. An attribution usually includes the title, author, publisher, and ISBN. For example: “ The Site Reliability Workbook , edited by Betsy Beyer, Niall Richard Murphy, David K. Rensin, Kent Kawahara, and Stephen Thorne (O’Reilly). Copyright 2018 Google LLC, 978-1-492-02950-2.” If you feel your use of code examples falls outside fair use or the permission given above, feel free to contact us at permissions@oreilly.com . O'Reilly Safari Note Safari (formerly Safari Books Online) is a membership-based training and reference platform for enterprise, government, educators, and individuals. Members have access to thousands of books, training videos, Learning Paths, interactive tutorials, and curated playlists from over 250 publishers, including O’Reilly Media, Harvard Business Review, Prentice Hall Professional, Addison-Wesley Professional, Microsoft Press, Sams, Que, Peachpit Press, Adobe, Focal Press, Cisco Press, John Wiley & Sons, Syngress, Morgan Kaufmann, IBM Redbooks, Packt, Adobe Press, FT Press, Apress, Manning, New Riders, McGraw-Hill, Jones & Bartlett, and Course Technology, among others. For more information, please visit https://oreilly.com/safari . How to Contact Us Please address comments and questions concerning this book to the publisher: O’Reilly Media, Inc. 1005 Gravenstein Highway North Sebastopol, CA 95472 800-998-9938 (in the United States or Canada) 707-829-0515 (international or local) 707-829-0104 (fax) We have a web page for this book, where we list errata, examples, and any additional information. You can access this page at https://bit.ly/siteReliabilityWkbk . To comment or ask technical questions about this book, send email to bookquestions@oreilly.com For more information about our books, courses, conferences, and news, see our website at https://www.oreilly.com . Find us on Facebook: https://facebook.com/oreilly Follow us on Twitter: https://twitter.com/oreillymedia Watch us on YouTube: https://www.youtube.com/oreillymedia Acknowledgments This book is the product of the enthusiastic and generous contributions of more than 100 people, including authors, tech writers, and reviewers. Each chapter has a byline for the individual authors and tech writers. We’d also like to take a moment to thank everyone not listed there. We would like to thank the following reviewers for providing valuable (and sometimes pointed) feedback: Abe Hassan, Alex Perry, Cara Donnelly, Chris Jones, Cody Smith, Dermot Duffy, Jarrod Todd, Jay Judkowitz, John T. Reese, Liz Fong-Jones, Mike Danese, Murali Suriar, Narayan Desai, Niccolò Cascarano, Ralph Pearson, Salim Virji, Todd Underwood, Vivek Rau, and Zoltan Egyed. We would like to express our deepest appreciation to the following people for serving as our overall quality bar for this volume. They made substantial contributions throughout the entire volume: Alex Matey, Max Luebbe, Matt Brown, and JC van Winkel. As the leaders of Google SRE, Benjamin Treynor Sloss and Ben Lutch were this book’s primary executive sponsors within Google; their strong and unwavering belief in a follow-up project that was a worthy companion of the first SRE book was essential to making this book happen. While the authors and technical writers are specifically acknowledged in each chapter, we’d like to recognize those that contributed to each chapter by providing thoughtful input, discussion, and review. In chapter order, they are: Implementing SLOs: Javier Kohen, Patrick Eaton, Richard Bondi, Yaniv Aknin Monitoring: Alex Matey, Clint Pauline, Cody Smith, JC van Winkel, Ola Kłapcińska, Štěpán Davidovič Alerting on SLOs: Alex Matey, Clint Pauline, Cody Smith, Iain Cooke, JC van Winkel, Štěpán Davidovič Eliminating Toil: Dermot Duffy, James O'Keeffe, Stephen Thorne Simplicity: Mark Brody On-Call: Alex Perry, Alex Hidalgo, David Huska, Sebastian Kirsch, Sabrina Farmer, Steven Carstensen, Liz Fong-Jones, Nandu Shah (Evernote), Robert Holley (Evernote) Incident Response: Alex Hidalgo, Alex Matey, Alex Perry, Dave Rensin, Matt Brown, Tor Gunnar Houeland, Trevor Strohman Postmortem Culture: Learning from Failure: John T. Reese Managing Load: Daniel E. Eisenbud, Dave Rensin, Dmitry Nefedkin, Dževad Trumić, Edward Wu (Niantic), JC van Winkel, Lucas Pereira, Luke Stone, Matt Brown, Natalia Sakowska, Niall Richard Murphy, Phil Keslin (Niantic), Rita Sodt, Scott Devoid, Simon Donovan, Tomasz Kulczyński Introducing Non-Abstract Large System Design: Ivo Krka, Matt Brown, Nicky Nicolosi, Tanya Reilly Data Processing Pipelines: Bartosz Janota (Spotify), Cara Donnelly, Chris Farrar, Johannes Rußek (Spotify), Max Charas, Max Luebbe, Michelle Duffy, Nelson Arapé (Spotify), Riccardo Petrocco (Spotify), Rickard Zwahlen (Spotify), Robert Stephenson (Spotify), Steven Thurgood Configuration Design and Best Practices: Charlene Perez, Dave Cunningham, Dave Rensin, JC van Winkel, John Reese, Stephen Thorne Configuration Specifics: Alex Matey, Bo Shi, Charlene Perez, Dave Rensin, Eric Johnson, Juliette Benton, Lars Wander, Mike Danese, Narayan Desai, Niall Richard Murphy, Štěpán Davidovič, Stephen Thorne Canarying Releases: Alex Matey, Liz Fong-Jones, Max Luebbe Identifying and Recovering from Overload: Andrew Harvey, Aleksander Szymanek, Brad Kratochvil, Ed Wehrwein, Duncan Sargeant, Jessika Reissland, Matt Brown, Piotr Sieklucki and Thomas Adamcik SRE Engagement Model: Brian Balser ( New York Times ), Deep Kapadia ( New York Times ), Michelle Duffy, Xavier Llorà SRE: Reaching Beyond Your Walls: Matt Brown SRE Team Lifecycles: Brian Balser ( New York Times ), Christophe Kalt, Daniel Rogers, Max Luebbe, Niall Richard Murphy, Ramón Medrano Llamas, Richard Bondi, Steven Carstensen, Stephen Thorne, Steven Thurgood, Thomas Wright Organizational Change Management in SRE: Dave Rensin, JC Van Winkel, Max Luebbe, Ronen Louvton, Stephen Thorne, Tom Feiner, Tsiki Rosenman We are also grateful to the following contributors, who supplied significant expertise or resources, or had some otherwise excellent effect on this work: Caleb Donaldson, Charlene Perez, Evan Leonard, Jennifer Petoff, Juliette Benton, and Lea Miller. We very much appreciate the thoughtful and in-depth feedback that we received from industry reviewers: Mark Burgess, David Blank-Edelman, John Looney, Jennifer Davis, Björn Rabenstein, Susan Fowler, Thomas A. Limoncelli, James Meickle, Theo Schlossangle, Jez Humble, Alice Goldfuss, Arup Chakrabarti, John Allspaw, Angus Lees, Eric Liang, Brendan Gregg, and Bryan Liles. We would like to extend a special thanks to Shylaja Nukala, who generously committed the time and skills of the SRE Technical Writing Team. She enthusiastically supported their necessary and valued efforts. Thanks also to the O’Reilly Media team—Virginia Wilson, Kristen Brown, Rachel Monaghan, Nikki McDonald, Melanie Yarbrough, and Gloria Lukos—for their help and support making the book a reality in our ambitious timeline. And an extra special thanks to Niall Richard Murphy: despite the fact that he moved on from Google before this book hit the shelves, his continual insights and dedication were crucial for getting a goodly portion of meaningful content over the finish line. His leadership, thoughtfulness, tenacity, and wit are nothing short of inspirational! Finally, the editors would also like to personally thank the following people: Betsy Beyer: To Grandmother, my go-to source for encouragement, inspiration, popcorn, pep, and puzzling. You made both this book and my everyday life better! To Duzzie, Hammer, Joan, Kiki, and Mini (note the alphabetical order—ha!) who helped shape me into the obsessive writer slash person I am today. And of course, Riba, for providing the DMD and other provisions necessary to fuel this effort. Niall Richard Murphy: To Léan, Oisín, Fiachra, and Kay, north stars. To someone whose protestations of self-interest are entirely out of odds with how he acts. To Sharon, more influential than she knows. To Alex, in a light-filled sitting room, with a cup of tea, a book, a box of dice, and thou. Stephen Thorne: To my mum and dad, who have always encouraged me to push myself. To my wife, Elspeth. To my colleagues who have given me more respect and encouragement than I think I deserve: Ola, Štěpán, Perry, and David. Dave Rensin: After I wrote my first book, I swore I’d never write another. That was six books ago and I say exactly the same thing each time. To my wife, Lia, who gives me the space to do it and never says “I told you so.” (Even though she tells me so.) To my colleagues at Google—and particularly to the family of SRE—who have taught me more these last few years about production engineering at scale than I had learned in the previous 20. Finally, to Benjamin Treynor Sloss, who interviewed me and convinced me to come to Google in the first place. Kent Kawahara: To my parents, Denby and Setsuko, and my Aunt Asako for helping me get to where I am. To my siblings, Randy and Patti, for their support over the years. To my wife, Angela, and my sons, Ryan, Ethan, and Brady, for their love and support. Finally, to the core team of Dave, Betsy, Niall, Juliette, and Stephen, I feel honored to have worked with you on this project. Previous Foreword II Next Chapter 1 - How SRE Relates to DevOps Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://git-scm.com/book/es/v2/Ramificaciones-en-Git-Reorganizar-el-Trabajo-Realizado
Git - Reorganizar el Trabajo Realizado About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Inicio - Sobre el Control de Versiones 1.1 Acerca del Control de Versiones 1.2 Una breve historia de Git 1.3 Fundamentos de Git 1.4 La Línea de Comandos 1.5 Instalación de Git 1.6 Configurando Git por primera vez 1.7 ¿Cómo obtener ayuda? 1.8 Resumen 2. Fundamentos de Git 2.1 Obteniendo un repositorio Git 2.2 Guardando cambios en el Repositorio 2.3 Ver el Historial de Confirmaciones 2.4 Deshacer Cosas 2.5 Trabajar con Remotos 2.6 Etiquetado 2.7 Alias de Git 2.8 Resumen 3. Ramificaciones en Git 3.1 ¿Qué es una rama? 3.2 Procedimientos Básicos para Ramificar y Fusionar 3.3 Gestión de Ramas 3.4 Flujos de Trabajo Ramificados 3.5 Ramas Remotas 3.6 Reorganizar el Trabajo Realizado 3.7 Recapitulación 4. Git en el Servidor 4.1 Los Protocolos 4.2 Configurando Git en un servidor 4.3 Generando tu clave pública SSH 4.4 Configurando el servidor 4.5 El demonio Git 4.6 HTTP Inteligente 4.7 GitWeb 4.8 GitLab 4.9 Git en un alojamiento externo 4.10 Resumen 5. Git en entornos distribuidos 5.1 Flujos de trabajo distribuidos 5.2 Contribuyendo a un Proyecto 5.3 Manteniendo un proyecto 5.4 Resumen 6. GitHub 6.1 Creación y configuración de la cuenta 6.2 Participando en Proyectos 6.3 Mantenimiento de un proyecto 6.4 Gestión de una organización 6.5 Scripting en GitHub 6.6 Resumen 7. Herramientas de Git 7.1 Revisión por selección 7.2 Organización interactiva 7.3 Guardado rápido y Limpieza 7.4 Firmando tu trabajo 7.5 Buscando 7.6 Reescribiendo la Historia 7.7 Reiniciar Desmitificado 7.8 Fusión Avanzada 7.9 Rerere 7.10 Haciendo debug con Git 7.11 Submódulos 7.12 Agrupaciones 7.13 Replace 7.14 Almacenamiento de credenciales 7.15 Resumen 8. Personalización de Git 8.1 Configuración de Git 8.2 Git Attributes 8.3 Puntos de enganche en Git 8.4 Un ejemplo de implantación de una determinada política en Git 8.5 Recapitulación 9. Git y Otros Sistemas 9.1 Git como Cliente 9.2 Migración a Git 9.3 Resumen 10. Los entresijos internos de Git 10.1 Fontanería y porcelana 10.2 Los objetos Git 10.3 Referencias Git 10.4 Archivos empaquetadores 10.5 Las especificaciones para hacer referencia a…​ (refspec) 10.6 Protocolos de transferencia 10.7 Mantenimiento y recuperación de datos 10.8 Variables de entorno 10.9 Recapitulación A1. Apéndice A: Git en otros entornos A1.1 Interfaces gráficas A1.2 Git en Visual Studio A1.3 Git en Eclipse A1.4 Git con Bash A1.5 Git en Zsh A1.6 Git en Powershell A1.7 Resumen A2. Apéndice B: Integrando Git en tus Aplicaciones A2.1 Git mediante Línea de Comandos A2.2 Libgit2 A2.3 JGit A3. Apéndice C: Comandos de Git A3.1 Configuración A3.2 Obtener y Crear Proyectos A3.3 Seguimiento Básico A3.4 Ramificar y Fusionar A3.5 Compartir y Actualizar Proyectos A3.6 Inspección y Comparación A3.7 Depuración A3.8 Parcheo A3.9 Correo Electrónico A3.10 Sistemas Externos A3.11 Administración A3.12 Comandos de Fontanería 2nd Edition 3.6 Ramificaciones en Git - Reorganizar el Trabajo Realizado Reorganizar el Trabajo Realizado En Git tenemos dos formas de integrar cambios de una rama en otra: la fusión (merge) y la reorganización (rebase). En esta sección vas a aprender en qué consiste la reorganización, cómo utilizarla, por qué es una herramienta sorprendente y en qué casos no es conveniente utilizarla. Reorganización Básica Volviendo al ejemplo anterior, en la sección sobre fusiones Procedimientos Básicos de Fusión puedes ver que has separado tu trabajo y realizado confirmaciones (commit) en dos ramas diferentes. Figura 35. El registro de confirmaciones inicial La manera más sencilla de integrar ramas, tal y como hemos visto, es el comando git merge . Realiza una fusión a tres bandas entre las dos últimas instantáneas de cada rama (C3 y C4) y el ancestro común a ambas (C2); creando una nueva instantánea (snapshot) y la correspondiente confirmación (commit). Figura 36. Fusionar una rama para integrar el registro de trabajos divergentes Sin embargo, también hay otra forma de hacerlo: puedes capturar los cambios introducidos en C4 y reaplicarlos encima de C3. Esto es lo que en Git llamamos reorganizar ( rebasing , en inglés). Con el comando git rebase , puedes capturar todos los cambios confirmados en una rama y reaplicarlos sobre otra. Por ejemplo, puedes lanzar los comandos: $ git checkout experiment $ git rebase master First, rewinding head to replay your work on top of it... Applying: added staged command Haciendo que Git vaya al ancestro común de ambas ramas (donde estás actualmente y de donde quieres reorganizar), saque las diferencias introducidas por cada confirmación en la rama donde estás, guarde esas diferencias en archivos temporales, reinicie (reset) la rama actual hasta llevarla a la misma confirmación que la rama de donde quieres reorganizar, y finalmente, vuelva a aplicar ordenadamente los cambios. Figura 37. Reorganizando sobre C3 los cambios introducidos en C4 En este momento, puedes volver a la rama master y hacer una fusión con avance rápido (fast-forward merge). $ git checkout master $ git merge experiment Figura 38. Avance rápido de la rama master Así, la instantánea apuntada por C4' es exactamente la misma apuntada por C5 en el ejemplo de la fusión. No hay ninguna diferencia en el resultado final de la integración, pero el haberla hecho reorganizando nos deja un historial más claro. Si examinas el historial de una rama reorganizada, este aparece siempre como un historial lineal: como si todo el trabajo se hubiera realizado en series, aunque realmente se haya hecho en paralelo. Habitualmente, optarás por esta vía cuando quieras estar seguro de que tus confirmaciones de cambio (commits) se pueden aplicar limpiamente sobre una rama remota; posiblemente, en un proyecto donde estés intentando colaborar, pero no lleves tú el mantenimiento. En casos como esos, puedes trabajar sobre una rama y luego reorganizar lo realizado en la rama origin/master cuando lo tengas todo listo para enviarlo al proyecto principal. De esta forma, la persona que mantiene el proyecto no necesitará hacer ninguna integración con tu trabajo; le bastará con un avance rápido o una incorporación limpia. Cabe destacar que, la instantánea (snapshot) apuntada por la confirmación (commit) final, tanto si es producto de una reorganización (rebase) como si lo es de una fusión (merge), es exactamente la misma instantánea; lo único diferente es el historial. La reorganización vuelve a aplicar cambios de una rama de trabajo sobre otra rama, en el mismo orden en que fueron introducidos en la primera, mientras que la fusión combina entre sí los dos puntos finales de ambas ramas. Algunas Reorganizaciones Interesantes También puedes aplicar una reorganización (rebase) sobre otra cosa además de sobre la rama de reorganización. Por ejemplo, considera un historial como el de Un historial con una rama puntual sobre otra rama puntual . Has ramificado a una rama puntual ( server ) para añadir algunas funcionalidades al proyecto, y luego has confirmado los cambios. Después, vuelves a la rama original para hacer algunos cambios en la parte cliente (rama client ), y confirmas también esos cambios. Por último, vuelves sobre la rama server y haces algunos cambios más. Figura 39. Un historial con una rama puntual sobre otra rama puntual Imagina que decides incorporar tus cambios del lado cliente sobre el proyecto principal para hacer un lanzamiento de versión; pero no quieres lanzar aún los cambios del lado servidor porque no están aún suficientemente probados. Puedes coger los cambios del cliente que no están en server ( C8 y C9 ) y reaplicarlos sobre tu rama principal usando la opción --onto del comando git rebase : $ git rebase --onto master server client Esto viene a decir: “Activa la rama client , averigua los cambios desde el ancestro común entre las ramas client y server , y aplicalos en la rama master ”. Puede parecer un poco complicado, pero los resultados son realmente interesantes. Figura 40. Reorganizando una rama puntual fuera de otra rama puntual Y, tras esto, ya puedes avanzar la rama principal (ver Avance rápido de tu rama master , para incluir los cambios de la rama client ): $ git checkout master $ git merge client Figura 41. Avance rápido de tu rama master , para incluir los cambios de la rama client Ahora supongamos que decides traerlos (pull) también sobre tu rama server . Puedes reorganizar (rebase) la rama server sobre la rama master sin necesidad siquiera de comprobarlo previamente, usando el comando git rebase [rama-base] [rama-puntual] , el cual activa la rama puntual ( server en este caso) y la aplica sobre la rama base ( master en este caso): $ git rebase master server Esto vuelca el trabajo de server sobre el de master , tal y como se muestra en Reorganizando la rama server sobre la rama master . Figura 42. Reorganizando la rama server sobre la rama master Después, puedes avanzar rápidamente la rama base ( master ): $ git checkout master $ git merge server Y por último puedes eliminar las ramas client y server porque ya todo su contenido ha sido integrado y no las vas a necesitar más, dejando tu registro tras todo este proceso tal y como se muestra en Historial final de confirmaciones de cambio : $ git branch -d client $ git branch -d server Figura 43. Historial final de confirmaciones de cambio Los Peligros de Reorganizar Ahh…​, pero la dicha de la reorganización no la alcanzamos sin sus contrapartidas, las cuales pueden resumirse en una línea: Nunca reorganices confirmaciones de cambio (commits) que hayas enviado (push) a un repositorio público. Si sigues esta recomendación, no tendrás problemas. Pero si no lo haces, la gente te odiará y serás despreciado por tus familiares y amigos. Cuando reorganizas algo, estás abandonando las confirmaciones de cambio ya creadas y estás creando unas nuevas; que son similares, pero diferentes. Si envías (push) confirmaciones (commits) a alguna parte, y otros las recogen (pull) de allí; y después vas tú y las reescribes con git rebase y las vuelves a enviar (push); tus colaboradores tendrán que refusionar (re-merge) su trabajo y todo se volverá tremendamente complicado cuando intentes recoger (pull) su trabajo de vuelta sobre el tuyo. Veamos con un ejemplo como reorganizar trabajo que has hecho público puede causar problemas. Imagínate que haces un clon desde un servidor central, y luego trabajas sobre él. Tu historial de cambios puede ser algo como esto: Figura 44. Clonar un repositorio y trabajar sobre él Ahora, otra persona trabaja también sobre ello, realiza una fusión (merge) y lleva (push) su trabajo al servidor central. Tú te traes (fetch) sus trabajos y los fusionas (merge) sobre una nueva rama en tu trabajo, con lo que tu historial quedaría parecido a esto: Figura 45. Traer (fetch) algunas confirmaciones de cambio (commits) y fusionarlas (merge) sobre tu trabajo A continuación, la persona que había llevado cambios al servidor central decide retroceder y reorganizar su trabajo; haciendo un git push --force para sobrescribir el registro en el servidor. Tu te traes (fetch) esos nuevos cambios desde el servidor. Figura 46. Alguien envió (push) confirmaciones (commits) reorganizadas, abandonando las confirmaciones en las que tu habías basado tu trabajo Ahora los dos están en un aprieto. Si haces git pull crearás una fusión confirmada, la cual incluirá ambas líneas del historial, y tu repositorio lucirá así: Figura 47. Vuelves a fusionar el mismo trabajo en una nueva fusión confirmada Si ejecutas git log sobre un historial así, verás dos confirmaciones hechas por el mismo autor y con la misma fecha y mensaje, lo cual será confuso. Es más, si luego tu envías (push) ese registro de vuelta al servidor, vas a introducir todas esas confirmaciones reorganizadas en el servidor central. Lo que puede confundir aún más a la gente. Era más seguro asumir que el otro desarrollador no quería que C4 y C6 estuviesen en el historial; por ello había reorganizado su trabajo de esa manera. Reorganizar una Reorganización Si te encuentras en una situación como esta, Git tiene algunos trucos que pueden ayudarte. Si alguien de tu equipo sobreescribe cambios en los que se basaba tu trabajo, tu reto es descubrir qué han sobreescrito y qué te pertenece. Además de la suma de control SHA-1, Git calcula una suma de control basada en el parche que introduce una confirmación. A esta se le conoce como “patch-id”. Si te traes el trabajo que ha sido sobreescrito y lo reorganizas sobre las nuevas confirmaciones de tu compañero, es posible que Git pueda identificar qué parte correspondía específicamente a tu trabajo y aplicarla de vuelta en la rama nueva. Por ejemplo, en el caso anterior, si en vez de hacer una fusión cuando estábamos en Alguien envió (push) confirmaciones (commits) reorganizadas, abandonando las confirmaciones en las que tu habías basado tu trabajo ejecutamos git rebase teamone/master , Git hará lo siguiente: Determinar el trabajo que es específico de nuestra rama (C2, C3, C4, C6, C7) Determinar cuáles no son fusiones confirmadas (C2, C3, C4) Determinar cuáles no han sido sobreescritas en la rama destino (solo C2 y C3, pues C4 corresponde al mismo parche que C4') Aplicar dichas confirmaciones encima de teamone/master Así que en vez del resultado que vimos en Vuelves a fusionar el mismo trabajo en una nueva fusión confirmada , terminaremos con algo más parecido a Reorganizar encima de un trabajo sobreescrito reorganizado. . Figura 48. Reorganizar encima de un trabajo sobreescrito reorganizado. Esto solo funciona si C4 y el C4' de tu compañero son parches muy similares. De lo contrario, la reorganización no será capaz de identificar que se trata de un duplicado y agregará otro parche similar a C4 (lo cual probablemente no podrá aplicarse limpiamente, pues los cambios ya estarían allí en algún lugar). También puedes simplificar el proceso si ejecutas git pull --rebase en vez del tradicional git pull . O, en este caso, puedes hacerlo manualmente con un git fetch primero, seguido de un git rebase teamone/master . Si sueles utilizar git pull y quieres que la opción --rebase esté activada por defecto, puedes asignar el valor de configuración pull.rebase haciendo algo como esto git config --global pull.rebase true . Si consideras la reorganización como una manera de limpiar tu trabajo y tus confirmaciones antes de enviarlas (push), y si sólo reorganizas confirmaciones (commits) que nunca han estado disponibles públicamente, no tendrás problemas. Si reorganizas (rebase) confirmaciones (commits) que ya estaban disponibles públicamente y la gente había basado su trabajo en ellas, entonces prepárate para tener problemas, frustrar a tu equipo y ser despreciado por tus compañeros. Si tu compañero o tú ven que aun así es necesario hacerlo en algún momento, asegúrense que todos sepan que deben ejecutar git pull --rebase para intentar aliviar en lo posible la frustración. Reorganizar vs. Fusionar Ahora que has visto en acción la reorganización y la fusión, te preguntarás cuál es mejor. Antes de responder, repasemos un poco qué representa el historial. Para algunos, el historial de confirmaciones de tu repositorio es un registro de todo lo que ha pasado. Un documento histórico, valioso por sí mismo y que no debería ser alterado. Desde este punto de vista, cambiar el historial de confirmaciones es casi como blasfemar; estarías mintiendo sobre lo que en verdad ocurrió. ¿Y qué pasa si hay una serie desastrosa de fusiones confirmadas? Nada. Así fué como ocurrió y el repositorio debería tener un registro de esto para la posteridad. La otra forma de verlo puede ser que, el historial de confirmaciones es la historia de cómo se hizo tu proyecto. Tú no publicarías el primer borrador de tu novela, y el manual de cómo mantener tus programas también debe estar editado con mucho cuidado. Esta es el área que utiliza herramientas como rebase y filter-branch para contar la historia de la mejor manera para los futuros lectores. Ahora, sobre si es mejor fusionar o reorganizar: verás que la respuesta no es tan sencilla. Git es una herramienta poderosa que te permite hacer muchas cosas con tu historial, y cada equipo y cada proyecto es diferente. Ahora que conoces cómo trabajan ambas herramientas, será cosa tuya decidir cuál de las dos es mejor para tu situación en particular. Normalmente, la manera de sacar lo mejor de ambas es reorganizar tu trabajo local, que aún no has compartido, antes de enviarlo a algún lugar; pero nunca reorganizar nada que ya haya sido compartido. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
http://www.trello.com/customers
Trello Customer Case Studies: How Companies Use Trello | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Trello Customer Stories Trello is used by thousands of businesses every day. Click on the awesome companies below to see how they've adopted Trello as a vital part of their workflow. Palace Law Justice for working men and women. Thriving with less paper, more revenue, and happier clients. Instinct Dog behaviour & training - Taking a custom approach to top dog care nationwide. Scan2CAD Converting customer files and company workflows. Desk Plants Sprouting small business success with 30% increase in sales. SwagUp Evolving from innovative idea to multi-million dollar business. UNICEF United Nations Children’s Fund - Helping others when they need it most. BurgerFi Growing an on-brand franchise business–fast. DoSomething.org Enabling social change to happen at scale. McCorvey Industrial fabrication - Making history with a paperless manufacturing process. Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://sre.google/workbook/alerting-on-slos/
Google SRE - Prometheus Alerting: Turn SLOs into Alerts Chapter 5 - Alerting on SLOs Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Alerting on SLOs By Steven Thurgood with Jess Frame, Anthony Lenton, Carmela Quinito, Anton Tolchanov, and Nejc Trdin This chapter explains how to turn your SLOs into actionable alerts on significant events. Both our first SRE book and this book talk about implementing SLOs. We believe that having good SLOs that measure the reliability of your platform, as experienced by your customers, provides the highest-quality indication for when an on-call engineer should respond. Here we give specific guidance on how to turn those SLOs into alerting rules so that you can respond to problems before you consume too much of your error budget. Our examples present a series of increasingly complex implementations for alerting metrics and logic; we discuss the utility and shortcomings of each. While our examples use a simple request-driven service and Prometheus syntax , you can apply this approach in any alerting framework. Alerting Considerations In order to generate alerts from service level indicators (SLIs) and an error budget, you need a way to combine these two elements into a specific rule. Your goal is to be notified for a significant event : an event that consumes a large fraction of the error budget. Consider the following attributes when evaluating an alerting strategy: Precision The proportion of events detected that were significant. Precision is 100% if every alert corresponds to a significant event. Note that alerting can become particularly sensitive to nonsignificant events during low-traffic periods (discussed in Low-Traffic Services and Error Budget Alerting ). Recall The proportion of significant events detected. Recall is 100% if every significant event results in an alert. Detection time How long it takes to send notifications in various conditions. Long detection times can negatively impact the error budget. Reset time How long alerts fire after an issue is resolved. Long reset times can lead to confusion or to issues being ignored. Ways to Alert on Significant Events Constructing alert rules for your SLOs can become quite complex. Here we present six ways to configure alerting on significant events, in order of increasing fidelity, to arrive at an option that offers good control over the four parameters of precision, recall, detection time, and reset time simultaneously. Each of the following approaches addresses a different problem, and some eventually solve multiple problems at the same time. The first three nonviable attempts work toward the latter three viable alerting strategies, with approach 6 being the most viable and most highly recommended option. The first method is simple to implement but inadequate, while the optimal method provides a complete solution to defend an SLO over both the long and the short term. For the purposes of this discussion, “error budgets” and “error rates” apply to all SLIs, not just those with “error” in their name. In the section What to Measure: Using SLIs , we recommend using SLIs that capture the ratio of good events to total events. The error budget gives the number of allowed bad events, and the error rate is the ratio of bad events to total events. 1: Target Error Rate ≥ SLO Threshold For the most trivial solution, you can choose a small time window (for example, 10 minutes) and alert if the error rate over that window exceeds the SLO. For example, if the SLO is 99.9% over 30 days, alert if the error rate over the previous 10 minutes is ≥ 0.1%: - alert: HighErrorRate expr: job:slo_errors_per_request:ratio_rate10m{job="myjob"} >= 0.001 Note This 10-minute average is calculated in Prometheus with a Recording rule: record: job:slo_errors_per_request:ratio_rate10m expr: sum(rate(slo_errors[10m])) by (job) / sum(rate(slo_requests[10m])) by (job) If you don’t export slo_errors and slo_requests from your job, you can create the time series by renaming a metric: record: slo_errors expr: http_errors Alerting when the recent error rate is equal to the SLO means that the system detects a budget spend of: alerting window size reporting period " role="presentation" style="position: relative;"> alerting window size reporting period Figure 5-1 shows the relationship between detection time and error rate for an example service with an alert window of 10 minutes and a 99.9% SLO. Figure 5-1. Detection time for an example service with an alert window of 10 minutes and a 99.9% SLO Table 5-1 shows the benefits and disadvantages of alerting when the immediate error rate is too high. Table 5-1. Pros and cons of alerting when the immediate error rate is too high Pros Cons Detection time is good: 0.6 seconds for a total outage. This alert fires on any event that threatens the SLO, exhibiting good recall. Precision is low: The alert fires on many events that do not threaten the SLO. A 0.1% error rate for 10 minutes would alert, while consuming only 0.02% of the monthly error budget . Taking this example to an extreme, you could receive up to 144 alerts per day every day, not act upon any alerts, and still meet the SLO. 2: Increased Alert Window We can build upon the preceding example by changing the size of the alert window to improve precision. By increasing the window size, you spend a higher budget amount before triggering an alert. To keep the rate of alerts manageable , you decide to be notified only if an event consumes 5% of the 30-day error budget—a 36-hour window: - alert: HighErrorRate expr: job:slo_errors_per_request:ratio_rate36h{job="myjob"} > 0.001 Now, the detection time is: 1 − SLO error ratio × alerting window size " role="presentation" style="position: relative;"> 1 − SLO error ratio × alerting window size Table 5-2 shows the benefits and disadvantages of alerting when the error rate is too high over a larger window of time. Table 5-2. Pros and cons of alerting when the error rate is too high over a larger window of time Pros Cons Detection time is still good: 2 minutes and 10 seconds for a complete outage. Better precision than the previous example: by ensuring that the error rate is sustained for longer, an alert will likely represent a significant threat to the error budget. Very poor reset time: In the case of 100% outage, an alert will fire shortly after 2 minutes, and continue to fire for the next 36 hours. Calculating rates over longer windows can be expensive in terms of memory or I/O operations, due to the large number of data points. Figure 5-2 shows that while the error rate over a 36-hour period has fallen to a negligible level, the 36-hour average error rate remains above the threshold. Figure 5-2. Error rate over a 36-hour period 3: Incrementing Alert Duration Most monitoring systems allow you to add a duration parameter to the alert criteria so the alert won’t fire unless the value remains above the threshold for some time. You may be tempted to use this parameter as a relatively inexpensive way to add longer windows: - alert: HighErrorRate expr: job:slo_errors_per_request:ratio_rate1m{job="myjob"} > 0.001 for: 1h Table 5-3 shows the benefits and disadvantages of using a duration parameter for alerts. Table 5-3. Pros and cons of using a duration parameter for alerts Pros Cons Alerts can be higher precision. Requiring a sustained error rate before firing means that alerts are more likely to correspond to a significant event. Poor recall and poor detection time: Because the duration does not scale with the severity of the incident, a 100% outage alerts after one hour, the same detection time as a 0.2% outage. The 100% outage would consume 140% of the 30-day budget in that hour. If the metric even momentarily returns to a level within SLO, the duration timer resets. An SLI that fluctuates between missing SLO and passing SLO may never alert. For the reasons listed in Table 5-3 , we do not recommend using durations as part of your SLO-based alerting criteria. 1 Figure 5-3 shows the average error rate over a 5-minute window of a service with a 10-minute duration before the alert fires. A series of 100% error spikes lasting 5 minutes every 10 minutes never triggers an alert, despite consuming 35% of the error budget. Figure 5-3. A service with 100% error spikes every 10 minutes Each spike consumed almost 12% of the 30-day budget, yet the alert never triggered. 4: Alert on Burn Rate To improve upon the previous solution, you want to create an alert with good detection time and high precision. To this end, you can introduce a burn rate to reduce the size of the window while keeping the alert budget spend constant. Burn rate is how fast, relative to the SLO, the service consumes the error budget. Figure 5-4 shows the relationship between burn rates and error budgets. The example service uses a burn rate of 1, which means that it’s consuming error budget at a rate that leaves you with exactly 0 budget at the end of the SLO’s time window (see Chapter 4 in our first book). With an SLO of 99.9% over a time window of 30 days, a constant 0.1% error rate uses exactly all of the error budget: a burn rate of 1. Figure 5-4. Error budgets relative to burn rates Table 5-4 shows burn rates, their corresponding error rates, and the time it takes to exhaust the SLO budget. Table 5-4. Burn rates and time to complete budget exhaustion Burn rate Error rate for a 99.9% SLO Time to exhaustion 1 0.1% 30 days 2 0.2% 15 days 10 1% 3 days 1,000 100% 43 minutes By keeping the alert window fixed at one hour and deciding that a 5% error budget spend is significant enough to notify someone, you can derive the burn rate to use for the alert. For burn rate–based alerts, the time taken for an alert to fire is: 1 − SLO error ratio × alerting window size × burn rate " role="presentation" style="position: relative;"> 1 − SLO error ratio × alerting window size × burn rate The error budget consumed by the time the alert fires is: burn rate × alerting window size period " role="presentation" style="position: relative;"> burn rate × alerting window size period Five percent of a 30-day error budget spend over one hour requires a burn rate of 36. The alerting rule now becomes: - alert: HighErrorRate expr: job:slo_errors_per_request:ratio_rate1h{job="myjob"} > 36 * 0.001 Table 5-5 shows the benefits and limitations of alerting based on burn rate. Table 5-5. Pros and cons of alerting based on burn rate Pros Cons Good precision: This strategy chooses a significant portion of error budget spend upon which to alert. Shorter time window, which is cheaper to calculate. Good detection time. Better reset time: 58 minutes. Low recall: A 35x burn rate never alerts, but consumes all of the 30-day error budget in 20.5 hours. Reset time: 58 minutes is still too long. 5: Multiple Burn Rate Alerts Your alerting logic can use multiple burn rates and time windows, and fire alerts when burn rates surpass a specified threshold. This option retains the benefits of alerting on burn rates and ensures that you don’t overlook lower (but still significant) error rates. It’s also a good idea to set up ticket notifications for incidents that typically go unnoticed but can exhaust your error budget if left unchecked—for example, a 10% budget consumption in three days. This rate of errors catches significant events, but since the rate of budget consumption provides adequate time to address the event, you don’t need to page someone. We recommend 2% budget consumption in one hour and 5% budget consumption in six hours as reasonable starting numbers for paging, and 10% budget consumption in three days as a good baseline for ticket alerts. The appropriate numbers depend on the service and the baseline page load. For busier services, and depending on on-call responsibilities over weekends and holidays, you may want ticket alerts for the six-hour window. Table 5-6 shows the corresponding burn rates and time windows for percentages of SLO budget consumed. Table 5-6. Recommended time windows and burn rates for percentages of SLO budget consumed SLO budget consumption Time window Burn rate Notification 2% 1 hour 14.4 Page 5% 6 hours 6 Page 10% 3 days 1 Ticket The alerting configuration may look like something like: expr: ( job:slo_errors_per_request:ratio_rate1h{job="myjob"} > (14.4*0.001) or job:slo_errors_per_request:ratio_rate6h{job="myjob"} > (6*0.001) ) severity: page expr: job:slo_errors_per_request:ratio_rate3d{job="myjob"} > 0.001 severity: ticket Figure 5-5 shows the detection time and alert type according to the error rate. Figure 5-5. Error rate, detection time, and alert notification Multiple burn rates allow you to adjust the alert to give appropriate priority based on how quickly you have to respond. If an issue will exhaust the error budget within hours or a few days, sending an active notification is appropriate. Otherwise, a ticket-based notification to address the alert the next working day is more appropriate. 2 Table 5-7 lists the benefits and disadvantages of using multiple burn rates. Table 5-7. Pros and cons of using multiple burn rates Pros Cons Ability to adapt the monitoring configuration to many situations according to criticality: alert quickly if the error rate is high; alert eventually if the error rate is low but sustained. Good precision, as with all fixed-budget portion alert approaches. Good recall, because of the three-day window. Ability to choose the most appropriate alert type based upon how quickly someone has to react to defend the SLO. More numbers, window sizes, and thresholds to manage and reason about. An even longer reset time, as a result of the three-day window. To avoid multiple alerts from firing if all conditions are true, you need to implement alert suppression. For example: 10% budget spend in five minutes also means that 5% of the budget was spent in six hours, and 2% of the budget was spent in one hour. This scenario will trigger three notifications unless the monitoring system is smart enough to prevent it from doing so. 6: Multiwindow, Multi-Burn-Rate Alerts We can enhance the multi-burn-rate alerts in iteration 5 to notify us only when we’re still actively burning through the budget—thereby reducing the number of false positives. To do this, we need to add another parameter: a shorter window to check if the error budget is still being consumed as we trigger the alert. A good guideline is to make the short window 1/12 the duration of the long window, as shown in Figure 5-6 . The graph shows both alerting threshold. After experiencing 15% errors for 10 minutes, the short window average goes over the alerting threshold immediately, and the long window average goes over the threshold after 5 minutes, at which point the alert starts firing. The short window average drops below the threshold 5 minutes after the errors stop, at which point the alert stops firing. The long window average drops below the threshold 60 minutes after the errors stop. Figure 5-6. Short and long windows for alerting For example, you can send a page-level alert when you exceed the 14.4x burn rate over both the previous one hour and the previous five minutes. This alert fires only once you’ve consumed 2% of the budget, but exhibits a better reset time by ceasing to fire five minutes later, rather than one hour later: expr: ( job:slo_errors_per_request:ratio_rate1h{job="myjob"} > (14.4*0.001) and job:slo_errors_per_request:ratio_rate5m{job="myjob"} > (14.4*0.001) ) or ( job:slo_errors_per_request:ratio_rate6h{job="myjob"} > (6*0.001) and job:slo_errors_per_request:ratio_rate30m{job="myjob"} > (6*0.001) ) severity: page expr: ( job:slo_errors_per_request:ratio_rate24h{job="myjob"} > (3*0.001) and job:slo_errors_per_request:ratio_rate2h{job="myjob"} > (3*0.001) ) or ( job:slo_errors_per_request:ratio_rate3d{job="myjob"} > 0.001 and job:slo_errors_per_request:ratio_rate6h{job="myjob"} > 0.001 ) severity: ticket We recommend the parameters listed in Table 5-8 as the starting point for your SLO-based alerting configuration. Table 5-8. Recommended parameters for a 99.9% SLO alerting configuration Severity Long window Short window Burn rate Error budget consumed Page 1 hour 5 minutes 14.4 2% Page 6 hours 30 minutes 6 5% Ticket 3 days 6 hours 1 10% We have found that alerting based on multiple burn rates is a powerful way to implement SLO-based alerting. Table 5-9 shows the benefits and limitations of using multiple burn rates and window sizes. Table 5-9. Pros and cons of using multiple burn rates and window sizes Pros Cons A flexible alerting framework that allows you to control the type of alert according to the severity of the incident and the requirements of the organization. Good precision, as with all fixed-budget portion alert approaches. Good recall, because of the three-day window. Lots of parameters to specify, which can make alerting rules hard to manage. For more on managing alerting rules, see Alerting at Scale . Low-Traffic Services and Error Budget Alerting The multiwindow, multi-burn-rate approach just detailed works well when a sufficiently high rate of incoming requests provides a meaningful signal when an issue arises. However, these approaches can cause problems for systems that receive a low rate of requests. If a system has either a low number of users or natural low-traffic periods (such as nights and weekends), you may need to alter your approach. It’s harder to automatically distinguish unimportant events in low-traffic services. For example, if a system receives 10 requests per hour, then a single failed request results in an hourly error rate of 10%. For a 99.9% SLO, this request constitutes a 1,000x burn rate and would page immediately, as it consumed 13.9% of the 30-day error budget. This scenario allows for only seven failed requests in 30 days. Single requests can fail for a large number of ephemeral and uninteresting reasons that aren’t necessarily cost-effective to solve in the same way as large systematic outages. The best solution depends on the nature of the service: what is the impact 3 of a single failed request? A high-availability target may be appropriate if failed requests are one-off, high-value requests that aren’t retried. It may make sense from a business perspective to investigate every single failed request. However, in this case, the alerting system notifies you of an error too late. We recommend a few key options to handle a low-traffic service: Generate artificial traffic to compensate for the lack of signal from real users. Combine smaller services into a larger service for monitoring purposes. Modify the product so that either: It takes more requests to qualify a single incident as a failure. The impact of a single failure is lower. Generating Artificial Traffic A system can synthesize user activity to check for potential errors and high-latency requests. In the absence of real users, your monitoring system can detect synthetic errors and requests, so your on-call engineers can respond to issues before they impact too many actual users. Artificial traffic provides more signals to work with, and allows you to reuse your existing monitoring logic and SLO values. You may even already have most of the necessary traffic-generating components, such as black-box probers and integration tests. Generating artificial load does have some downsides. Most services that warrant SRE support are complex, and have a large system control surface. Ideally, the system should be designed and built for monitoring using artificial traffic. Even for a nontrivial service, you can synthesize only a small portion of the total number of user request types. For a stateful service, the greater number of states exacerbates this problem. Additionally, if an issue affects real users but doesn’t affect artificial traffic, the successful artificial requests hide the real user signal, so you aren’t notified that users see errors. Combining Services If multiple low-traffic services contribute to one overall function, combining their requests into a single higher-level group can detect significant events more precisely and with fewer false positives. For this approach to work, the services must be related in some way—you can combine microservices that form part of the same product, or multiple request types handled by the same binary. A downside to combining services is that a complete failure of an individual service may not count as a significant event. You can increase the likelihood that a failure will affect the group as a whole by choosing services with a shared failure domain, such as a common backend database. You can still use longer-period alerts that eventually catch these 100% failures for individual services. Making Service and Infrastructure Changes Alerting on significant events aims to provide sufficient notice to mitigate problems before they exhaust the entire error budget. If you can’t adjust the monitoring to be less sensitive to ephemeral events, and generating synthetic traffic is impractical, you might instead consider changing the service to reduce the user impact of a single failed request. For example, you might: Modify the client to retry, with exponential backoff and jitter. 4 Set up fallback paths that capture the request for eventual execution, which can take place on the server or on the client. These changes are useful for high-traffic systems, but even more so for low-traffic systems: they allow for more failed events in the error budget, more signal from monitoring, and more time to respond to an incident before it becomes significant. Lowering the SLO or Increasing the Window You might also want to reconsider if the impact of a single failure on the error budget accurately reflects its impact on users. If a small number of errors causes you to lose error budget, do you really need to page an engineer to fix the issue immediately? If not, users would be equally happy with a lower SLO. With a lower SLO, an engineer is notified only of a larger sustained outage. Once you have negotiated lowering the SLO with the service’s stakeholders (for example, lowering the SLO from 99.9% to 99%), implementing the change is very simple: if you already have systems in place for reporting, monitoring, and alerting based upon an SLO threshold, simply add the new SLO value to the relevant systems. Lowering the SLO does have a downside: it involves a product decision. Changing the SLO affects other aspects of the system, such as expectations around system behavior and when to enact the error budget policy. These other requirements may be more important to the product than avoiding some number of low-signal alerts. In a similar manner, increasing the time window used for the alerting logic ensures alerts that trigger pages are more significant and worthy of attention. In practice, we use some combination of the following methods to alert for low-traffic services: Generating fake traffic, when doing so is possible and can achieve good coverage Modifying clients so that ephemeral failures are less likely to cause user harm Aggregating smaller services that share some failure mode Setting SLO thresholds commensurate with the actual impact of a failed request Extreme Availability Goals Services with an extremely low or an extremely high availability goal may require special consideration. For example, consider a service that has a 90% availability target. Table 5-8 says to page when 2% of the error budget in a single hour is consumed. Because a 100% outage consumes only 1.4% of the budget in that hour, this alert could never fire. If your error budgets are set over long time periods, you may need to tune your alerting parameters. For services with an extremely high availability goal, the time to exhaustion for a 100% outage is extremely small. A 100% outage for a service with a target monthly availability of 99.999% would exhaust its budget in 26 seconds—which is smaller than the metric collection interval of many monitoring services, let alone the end-to-end time to generate an alert and pass it through notification systems like email and SMS. Even if the alert goes straight to an automated resolution system, the issue may entirely consume the error budget before you can mitigate it. Receiving notifications that you have only 26 seconds of budget left isn’t necessarily a bad strategy; it’s just not useful for defending the SLO. The only way to defend this level of reliability is to design the system so that the chance of a 100% outage is extremely low. That way, you can fix issues before consuming the budget. For example, if you initially roll out that change to only 1% of your users, and burn your error budget at the same rate of 1%, you now have 43 minutes before you exhaust your error budget. See Canarying Releases for tactics on designing such a system. Alerting at Scale When you scale your service, make sure that your alerting strategy is likewise scalable. You might be tempted to specify custom alerting parameters for individual services. If your service comprises 100 microservices (or equivalently, a single service with 100 different request types), this scenario very quickly accumulates toil and cognitive load that does not scale. In this case, we strongly advise against specifying the alert window and burn rate parameters independently for each service, because doing so quickly becomes overwhelming. 5 Once you decide on your alerting parameters, apply them to all your services. One technique for managing a large number of SLOs is to group request types into buckets of approximately similar availability requirements. For example, for a service with availability and latency SLOs, you can group its request types into the following buckets: CRITICAL For request types that are the most important, such as a request when a user logs in to the service. HIGH_FAST For requests with high availability and low latency requirements. These requests involve core interactive functionality, such as when a user clicks a button to see how much money their advertising inventory has made this month. HIGH_SLOW For important but less latency-sensitive functionality, such as when a user clicks a button to generate a report of all advertising campaigns over the past few years, and does not expect the data to return instantly. LOW For requests that must have some availability, but for which outages are mostly invisible to users—for example, polling handlers for account notifications that can fail for long periods of time with no user impact. NO_SLO For functionality that is completely invisible to the user—for example, dark launches or alpha functionality that is explicitly outside of any SLO. By grouping requests rather than placing unique availability and latency objectives on all request types, you can group requests into five buckets, like the example in Table 5-10 . Table 5-10. Request class buckets according to similar availability requirements and thresholds Request class Availability Latency @ 90% 6 Latency @ 99% CRITICAL 99.99% 100 ms 200 ms HIGH_FAST 99.9% 100 ms 200 ms HIGH_SLOW 99.9% 1,000 ms 5,000 ms LOW 99% None None NO_SLO None None None These buckets provide sufficient fidelity for protecting user happiness, but entail less toil than a system that is more complicated and expensive to manage that probably maps more precisely to the user experience. Conclusion If you set SLOs that are meaningful, understood, and represented in metrics, you can configure alerting to notify an on-caller only when there are actionable, specific threats to the error budget. The techniques for alerting on significant events range from alerting when your error rate goes above your SLO threshold to using multiple levels of burn rate and window sizes. In most cases, we believe that the multiwindow, multi-burn-rate alerting technique is the most appropriate approach to defending your application’s SLOs. We hope we have given you the context and tools required to make the right configuration decisions for your own application and organization. 1 Duration clauses can occasionally be useful when you are filtering out ephemeral noise over very short durations. However, you still need to be aware of the cons listed in this section. 2 As described in the introduction to Site Reliability Engineering , pages and tickets are the only valid ways to get a human to take action. 3 The section What to Measure: Using SLIs recommends a style of SLI that scales according to the impact on the user. 4 See “Overloads and Failure” in Site Reliability Engineering . 5 With the exception of temporary changes to alerting parameters, which are necessary when you’re fixing an ongoing outage and you don’t need to receive notifications during that period. 6 Ninety percent of requests are faster than this threshold . Previous Chapter 4 - Monitoring Next Chapter 6 - Eliminating Toil Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://scholar.google.com/citations?view_op=view_org&hl=ko&oe=ASCII&org=7484523900994334132
:root{--gm3-sys-color-on-surface-rgb:31,31,31}@media screen and (prefers-color-scheme:dark){:root{--gm3-sys-color-on-surface-rgb:227,227,227}}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.q4Wquf,.nfoC7c{display:block;height:25vh;position:relative}@media (min-width:600px){.q4Wquf,.nfoC7c{height:150px}}.q4Wquf.Irjbwb{height:auto}@media screen and (prefers-color-scheme:dark){.q4Wquf:not(.GtvzYd){display:none}}.nfoC7c{margin:0;overflow:hidden}.PwpMUe,.lVUmD{display:block;height:100%;margin:0 auto;width:100%}.St9mde{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.St9mde{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.PwpMUe,.lVUmD,.St9mde{-o-object-fit:contain;object-fit:contain}.wsArZ[data-ss-mode="1"] .q4Wquf,.wsArZ[data-ss-mode="1"] .St9mde{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .St9mde{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .q4Wquf,.NQ5OL .St9mde{height:auto;width:100%}.NQ5OL .St9mde{max-width:400px}}.q4Wquf.NWba7e,.q4Wquf.NWba7e .St9mde{height:auto}.q4Wquf.NWba7e .St9mde{height:auto;max-width:312px;width:100%}.q4Wquf.NWba7e.zpCp3 .St9mde{max-width:unset}.q4Wquf.IiQozc .St9mde{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.q4Wquf.Irjbwb .St9mde{height:auto;width:100%}.q4Wquf.EEeaqf .St9mde{max-height:144px;max-width:144px}.SnAaEd{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.SnAaEd{display:none}}.SnAaEd::after,.SnAaEd::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.SnAaEd::after,.SnAaEd::before{width:150px}}.SnAaEd::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.SnAaEd::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .SnAaEd~.St9mde{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .SnAaEd~.St9mde{width:auto}}.RHNWk .St9mde{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .RHNWk .St9mde{width:115px}}.cf660d .St9mde{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.tUhwwc .St9mde{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.Jkvqxd .St9mde{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.onc8Ic .St9mde{-webkit-transform:translate(var( --c-ps-s,24px ),0);-ms-transform:translate(var( --c-ps-s,24px ),0);transform:translate(var( --c-ps-s,24px ),0)}.WA89Yb .St9mde{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .XEN8Yb .St9mde{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .XEN8Yb .St9mde{max-width:115px}}@media (min-width:1240px) and (orientation:landscape),all and (min-width:1600px){.KyPKed .NQ5OL .XEN8Yb .St9mde{max-width:115px}}.IsSr6b .St9mde{max-width:300px}.mmskdd .St9mde{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@keyframes quantumWizBoxInkSpread{0%{-webkit-transform:translate(-50%,-50%) scale(.2);transform:translate(-50%,-50%) scale(.2)}100%{-webkit-transform:translate(-50%,-50%) scale(2.2);transform:translate(-50%,-50%) scale(2.2)}}@-webkit-keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@keyframes quantumWizIconFocusPulse{0%{-webkit-transform:translate(-50%,-50%) scale(1.5);transform:translate(-50%,-50%) scale(1.5);opacity:0}100%{-webkit-transform:translate(-50%,-50%) scale(2);transform:translate(-50%,-50%) scale(2);opacity:1}}@-webkit-keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkSpread{0%{-webkit-transform:scale(1.5);transform:scale(1.5);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@-webkit-keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}@keyframes quantumWizRadialInkFocusPulse{0%{-webkit-transform:scale(2);transform:scale(2);opacity:0}100%{-webkit-transform:scale(2.5);transform:scale(2.5);opacity:1}}:root{--wf-tfs:calc(var(--c-tfs,32)/16*1rem);--wf-tfs-bp2:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp3:calc(var(--c-tfs,36)/16*1rem);--wf-tfs-bp5:calc(var(--c-tfs,44)/16*1rem);--wf-stfs:calc(var(--c-stfs,16)/16*1rem);--wf-stfs-bp5:calc(var(--c-stfs,16)/16*1rem)}:root{--wf-harmonize-filter-light:none;--wf-harmonize-filter-dark:none}.Dzz9Db,.GpMPBe{display:block;height:25vh;position:relative}@media (min-width:600px){.Dzz9Db,.GpMPBe{height:150px}}@media screen and (prefers-color-scheme:dark){.Dzz9Db:not(.GtvzYd){display:none}}.Dzz9Db.Irjbwb{height:auto}.GpMPBe{margin:0;overflow:hidden}.UFQPDd,.JNOvdd{display:block;height:100%;margin:0 auto;-o-object-fit:contain;object-fit:contain;width:100%}.f4ZpM{display:block;-webkit-filter:var(--wf-harmonize-filter-light);filter:var(--wf-harmonize-filter-light);height:100%;max-width:100%;min-height:110px;position:relative;-webkit-transform:translate(-43%,-3%);-ms-transform:translate(-43%,-3%);transform:translate(-43%,-3%);width:auto;z-index:3}@media screen and (prefers-color-scheme:dark){.f4ZpM{-webkit-filter:var(--wf-harmonize-filter-dark);filter:var(--wf-harmonize-filter-dark)}}.wsArZ[data-ss-mode="1"] .Dzz9Db,.wsArZ[data-ss-mode="1"] .f4ZpM{height:auto;width:100%}.wsArZ[data-ss-mode="1"] .f4ZpM{max-width:400px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .Dzz9Db,.NQ5OL .f4ZpM{height:auto;width:100%}.NQ5OL .f4ZpM{max-width:400px}}.Dzz9Db.utFBGf,.Dzz9Db.utFBGf .f4ZpM{height:auto}.Dzz9Db.utFBGf .f4ZpM{height:auto;max-width:312px;width:100%}.Dzz9Db.utFBGf.zpCp3 .f4ZpM{max-width:unset}.Dzz9Db.IiQozc .f4ZpM{margin:0 auto;-webkit-transform:none;-ms-transform:none;transform:none}.Dzz9Db.Irjbwb .f4ZpM{height:auto;width:100%}.Dzz9Db.EEeaqf .f4ZpM{max-height:144px;max-width:144px}.nPt1pc{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(233,233,233,0)),color-stop(62.22%,rgba(233,233,233,0)),color-stop(40.22%,rgb(233,233,233)),to(rgba(233,233,233,0)));background-image:-webkit-linear-gradient(top,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);background-image:linear-gradient(to bottom,rgba(233,233,233,0) 0,rgba(233,233,233,0) 62.22%,rgb(233,233,233) 40.22%,rgba(233,233,233,0) 100%);height:100%;left:0;overflow:hidden;position:absolute;right:0;top:0;z-index:2}@media screen and (prefers-color-scheme:dark){.nPt1pc{display:none}}.nPt1pc::after,.nPt1pc::before{content:"";display:block;height:100%;min-width:110px;position:absolute;right:-10%;-webkit-transform:rotate(-104deg);-ms-transform:rotate(-104deg);transform:rotate(-104deg);width:25vh;z-index:2}@media (min-width:600px){.nPt1pc::after,.nPt1pc::before{width:150px}}.nPt1pc::before{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(243,243,243,0)),to(rgba(243,243,243,.9)));background-image:-webkit-linear-gradient(top,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);background-image:linear-gradient(to bottom,rgba(243,243,243,0) 0,rgba(243,243,243,.9) 100%);bottom:-10%}.nPt1pc::after{background-image:-webkit-gradient(linear,left top,left bottom,from(rgba(255,255,255,0)),to(rgba(255,255,255,.9)));background-image:-webkit-linear-gradient(top,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);background-image:linear-gradient(to bottom,rgba(255,255,255,0) 0,rgba(255,255,255,.9) 100%);bottom:-80%}.wsArZ[data-ss-mode="1"] .nPt1pc~.f4ZpM{width:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .nPt1pc~.f4ZpM{width:auto}}.ZS7CGc .f4ZpM{height:auto}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .ZS7CGc .f4ZpM{width:115px}}.qiRZ5e .f4ZpM{-webkit-transform:translate(-9%,-3%);-ms-transform:translate(-9%,-3%);transform:translate(-9%,-3%)}.vIv7Gf .f4ZpM{margin:auto;max-height:230px;right:0;top:-3%;-webkit-transform:none;-ms-transform:none;transform:none}.nvYXVd .f4ZpM{-webkit-transform:translate(9%,-3%);-ms-transform:translate(9%,-3%);transform:translate(9%,-3%)}.uOhnzd .f4ZpM{-webkit-transform:translate(24px,0);-ms-transform:translate(24px,0);transform:translate(24px,0)}.MsYMaf .f4ZpM{-webkit-transform:translate(0,0);-ms-transform:translate(0,0);transform:translate(0,0)}.wsArZ[data-ss-mode="1"] .YIi9qf .f4ZpM{max-width:115px}@media (min-width:600px) and (orientation:landscape),all and (min-width:1600px){.NQ5OL .YIi9qf .f4ZpM{max-width:115px}}.QG3Xbe .f4ZpM{max-width:300px}.F6gtje .f4ZpM{-webkit-transform:none;-ms-transform:none;transform:none}@-webkit-keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@keyframes mdc-ripple-fg-radius-in{from{-webkit-animation-timing-function:cubic-bezier(.4,0,.2,1);animation-timing-function:cubic-bezier(.4,0,.2,1);-webkit-transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1);transform:translate(var(--mdc-ripple-fg-translate-start,0)) scale(1)}to{-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}}@-webkit-keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@keyframes mdc-ripple-fg-opacity-in{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:0}to{opacity:var(--mdc-ripple-fg-opacity,0)}}@-webkit-keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}@keyframes mdc-ripple-fg-opacity-out{from{-webkit-animation-timing-function:linear;animation-timing-function:linear;opacity:var(--mdc-ripple-fg-opacity,0)}to{opacity:0}}.VfPpkd-ksKsZd-XxIAqe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity;position:relative;outline:none;overflow:hidden}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-ksKsZd-XxIAqe::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-ksKsZd-XxIAqe::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded],.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd{overflow:visible}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded]::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd::after{top:0;left:0;width:100%;height:100%}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::before,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe[data-mdc-ripple-is-unbounded].VfPpkd-ksKsZd-mWPk3d::after,.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd.VfPpkd-ksKsZd-mWPk3d::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-ksKsZd-XxIAqe::before,.VfPpkd-ksKsZd-XxIAqe::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-ksKsZd-XxIAqe:hover::before,.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe::before,.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):focus::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d)::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-ksKsZd-XxIAqe:not(.VfPpkd-ksKsZd-mWPk3d):active::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-ksKsZd-XxIAqe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe{font-size:24px;width:48px;height:48px;padding:12px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-Jh9lGc{width:40px;height:40px;margin-top:4px;margin-bottom:4px;margin-right:4px;margin-left:4px}.VfPpkd-Bz112c-LgbsSe.VfPpkd-Bz112c-LgbsSe-OWXEXe-e5LLRc-SxQuSe .VfPpkd-Bz112c-J1Ukfc-LhBDec{max-height:40px;max-width:40px}.VfPpkd-Bz112c-LgbsSe:disabled{color:rgba(0,0,0,.38);color:var(--mdc-theme-text-disabled-on-light,rgba(0,0,0,.38))}.VfPpkd-Bz112c-LgbsSe svg,.VfPpkd-Bz112c-LgbsSe img{width:24px;height:24px}.VfPpkd-Bz112c-LgbsSe{display:inline-block;position:relative;-webkit-box-sizing:border-box;box-sizing:border-box;border:none;outline:none;background-color:transparent;fill:currentColor;color:inherit;text-decoration:none;cursor:pointer;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;z-index:0;overflow:visible}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-RLmnJb{position:absolute;top:50%;height:48px;left:50%;width:48px;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-J1Ukfc-LhBDec,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-J1Ukfc-LhBDec{display:block}}.VfPpkd-Bz112c-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-Bz112c-LgbsSe[hidden]{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-KVuj8d-Q3DXx{-webkit-box-align:center;-webkit-align-items:center;align-items:center;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center}.VfPpkd-Bz112c-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:100%;width:100%;display:none}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-Bz112c-J1Ukfc-LhBDec::after{border-color:CanvasText}}.VfPpkd-Bz112c-kBDsod{display:inline-block}.VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd,.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod{display:none}.VfPpkd-Bz112c-LgbsSe-OWXEXe-IT5dJd .VfPpkd-Bz112c-kBDsod.VfPpkd-Bz112c-kBDsod-OWXEXe-IT5dJd{display:inline-block}.VfPpkd-Bz112c-mRLv6{height:100%;left:0;outline:none;position:absolute;top:0;width:100%}.VfPpkd-Bz112c-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1;z-index:var(--mdc-ripple-z-index,1)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{z-index:0;z-index:var(--mdc-ripple-z-index,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Bz112c-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{top:0;left:0;width:100%;height:100%}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0);width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Bz112c-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc::after{background-color:#000;background-color:var(--mdc-ripple-color,#000)}.VfPpkd-Bz112c-LgbsSe:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.VfPpkd-Bz112c-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.VfPpkd-Bz112c-LgbsSe:disabled:hover .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Bz112c-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Bz112c-Jh9lGc::before,.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Bz112c-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-Bz112c-LgbsSe:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Bz112c-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.VfPpkd-Bz112c-LgbsSe .VfPpkd-Bz112c-Jh9lGc{height:100%;left:0;pointer-events:none;position:absolute;top:0;width:100%;z-index:-1}.VfPpkd-dgl2Hf-ppHlrf-sM5MNb{display:inline}.VfPpkd-LgbsSe{position:relative;display:-webkit-inline-box;display:-webkit-inline-flex;display:inline-flex;-webkit-box-align:center;-webkit-align-items:center;align-items:center;-webkit-box-pack:center;-webkit-justify-content:center;justify-content:center;-webkit-box-sizing:border-box;box-sizing:border-box;min-width:64px;border:none;outline:none;line-height:inherit;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;-webkit-appearance:none;overflow:visible;vertical-align:middle;background:transparent}.VfPpkd-LgbsSe .VfPpkd-BFbNVe-bF1uUb{width:100%;height:100%;top:0;left:0}.VfPpkd-LgbsSe::-moz-focus-inner{padding:0;border:0}.VfPpkd-LgbsSe:active{outline:none}.VfPpkd-LgbsSe:hover{cursor:pointer}.VfPpkd-LgbsSe:disabled{cursor:default;pointer-events:none}.VfPpkd-LgbsSe[hidden]{display:none}.VfPpkd-LgbsSe .VfPpkd-kBDsod{margin-left:0;margin-right:8px;display:inline-block;position:relative;vertical-align:top}[dir=rtl] .VfPpkd-LgbsSe .VfPpkd-kBDsod,.VfPpkd-LgbsSe .VfPpkd-kBDsod[dir=rtl]{margin-left:8px;margin-right:0}.VfPpkd-LgbsSe .VfPpkd-UdE5de-uDEFge{font-size:0;position:absolute;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);top:50%;left:50%;line-height:normal}.VfPpkd-LgbsSe .VfPpkd-vQzf8d{position:relative}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{pointer-events:none;border:2px solid transparent;border-radius:6px;-webkit-box-sizing:content-box;box-sizing:content-box;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px);display:none}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-color:CanvasText}}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{content:"";border:2px solid transparent;border-radius:8px;display:block;position:absolute;top:50%;left:50%;-webkit-transform:translate(-50%,-50%);-ms-transform:translate(-50%,-50%);transform:translate(-50%,-50%);height:calc(100% + 4px);width:calc(100% + 4px)}@media screen and (forced-colors:active){.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-color:CanvasText}}@media screen and (forced-colors:active){.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-J1Ukfc-LhBDec,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-J1Ukfc-LhBDec{display:block}}.VfPpkd-LgbsSe .VfPpkd-RLmnJb{position:absolute;top:50%;height:48px;left:0;right:0;-webkit-transform:translateY(-50%);-ms-transform:translateY(-50%);transform:translateY(-50%)}.VfPpkd-vQzf8d+.VfPpkd-kBDsod{margin-left:8px;margin-right:0}[dir=rtl] .VfPpkd-vQzf8d+.VfPpkd-kBDsod,.VfPpkd-vQzf8d+.VfPpkd-kBDsod[dir=rtl]{margin-left:0;margin-right:8px}svg.VfPpkd-kBDsod{fill:currentColor}.VfPpkd-LgbsSe-OWXEXe-dgl2Hf{margin-top:6px;margin-bottom:6px}.VfPpkd-LgbsSe{-moz-osx-font-smoothing:grayscale;-webkit-font-smoothing:antialiased;text-decoration:none}.VfPpkd-LgbsSe{padding:0 8px 0 8px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{-webkit-transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1);transition:box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);padding:0 16px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 12px 0 16px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 16px 0 12px}.VfPpkd-LgbsSe-OWXEXe-INsAgc{border-style:solid;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-style:solid;border-color:transparent}.VfPpkd-LgbsSe{--mdc-ripple-fg-size:0;--mdc-ripple-left:0;--mdc-ripple-top:0;--mdc-ripple-fg-scale:1;--mdc-ripple-fg-translate-end:0;--mdc-ripple-fg-translate-start:0;-webkit-tap-highlight-color:rgba(0,0,0,0);will-change:transform,opacity}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{position:absolute;border-radius:50%;opacity:0;pointer-events:none;content:""}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{-webkit-transition:opacity 15ms linear,background-color 15ms linear;transition:opacity 15ms linear,background-color 15ms linear;z-index:1}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{z-index:0}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::before{-webkit-transform:scale(var(--mdc-ripple-fg-scale,1));-ms-transform:scale(var(--mdc-ripple-fg-scale,1));transform:scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{top:0;left:0;-webkit-transform:scale(0);-ms-transform:scale(0);transform:scale(0);-webkit-transform-origin:center center;-ms-transform-origin:center center;transform-origin:center center}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-ZNMTqd .VfPpkd-Jh9lGc::after{top:var(--mdc-ripple-top,0);left:var(--mdc-ripple-left,0)}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-lJfZMc .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards;animation:mdc-ripple-fg-radius-in 225ms forwards,mdc-ripple-fg-opacity-in 75ms forwards}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-Tv8l5d-OmS1vf .VfPpkd-Jh9lGc::after{-webkit-animation:mdc-ripple-fg-opacity-out .15s;animation:mdc-ripple-fg-opacity-out .15s;-webkit-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));-ms-transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1));transform:translate(var(--mdc-ripple-fg-translate-end,0)) scale(var(--mdc-ripple-fg-scale,1))}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{top:-50%;left:-50%;width:200%;height:200%}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d .VfPpkd-Jh9lGc::after{width:var(--mdc-ripple-fg-size,100%);height:var(--mdc-ripple-fg-size,100%)}.VfPpkd-Jh9lGc{position:absolute;-webkit-box-sizing:content-box;box-sizing:content-box;overflow:hidden;z-index:0;top:0;left:0;bottom:0;right:0}.VfPpkd-LgbsSe{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-text-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-k8QpJ:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-k8QpJ.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-filled-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-k8QpJ .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;-webkit-box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12);box-shadow:0 3px 1px -2px rgba(0,0,0,.2),0 2px 2px 0 rgba(0,0,0,.14),0 1px 5px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{background-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled){color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::before{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc::after{background-color:#fff}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.08}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-protected-button-pressed-state-layer-opacity,0.24)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-MV7yeb.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(.VfPpkd-ksKsZd-mWPk3d):focus{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:hover{-webkit-box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12);box-shadow:0 2px 4px -1px rgba(0,0,0,.2),0 4px 5px 0 rgba(0,0,0,.14),0 1px 10px 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:not(:disabled):active{-webkit-box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12);box-shadow:0 5px 5px -3px rgba(0,0,0,.2),0 8px 10px 1px rgba(0,0,0,.14),0 3px 14px 2px rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-MV7yeb:disabled{-webkit-box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12);box-shadow:0 0 0 0 rgba(0,0,0,.2),0 0 0 0 rgba(0,0,0,.14),0 0 0 0 rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc{font-family:Roboto,sans-serif;font-size:.875rem;letter-spacing:.0892857143em;font-weight:500;text-transform:uppercase;height:36px;border-radius:4px;padding:0 15px 0 15px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{color:rgba(0,0,0,.38)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-kBDsod{font-size:1.125rem;width:1.125rem;height:1.125rem}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::before{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc::after{background-color:#6200ee}.VfPpkd-LgbsSe-OWXEXe-INsAgc:hover .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-outlined-button-pressed-state-layer-opacity,0.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec{border-radius:2px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-J1Ukfc-LhBDec::after{border-radius:4px}.VfPpkd-LgbsSe-OWXEXe-INsAgc:not(:disabled){border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc:disabled{border-color:rgba(0,0,0,.12)}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-UbuQg{padding:0 11px 0 15px}.VfPpkd-LgbsSe-OWXEXe-INsAgc.VfPpkd-LgbsSe-OWXEXe-Bz112c-M1Soyc{padding:0 15px 0 11px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-Jh9lGc{top:-1px;left:-1px;bottom:-1px;right:-1px;border-width:1px}.VfPpkd-LgbsSe-OWXEXe-INsAgc .VfPpkd-RLmnJb{left:-1px;width:calc(100% + 2px)}.nCP5yc{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.nCP5yc .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.nCP5yc:not(:disabled){background-color:rgb(26,115,232);background-color:var(--gm-fillbutton-container-color,rgb(26,115,232))}.nCP5yc:not(:disabled){color:#fff;color:var(--gm-fillbutton-ink-color,#fff)}.nCP5yc:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-fillbutton-disabled-container-color,rgba(60,64,67,.12))}.nCP5yc:disabled{color:rgba(60,64,67,.38);color:var(--gm-fillbutton-disabled-ink-color,rgba(60,64,67,.38))}.nCP5yc .VfPpkd-Jh9lGc::before,.nCP5yc .VfPpkd-Jh9lGc::after{background-color:rgb(32,33,36);background-color:var(--gm-fillbutton-state-color,rgb(32,33,36))}.nCP5yc:hover .VfPpkd-Jh9lGc::before,.nCP5yc.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.16;opacity:var(--mdc-ripple-hover-opacity,.16)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.24;opacity:var(--mdc-ripple-focus-opacity,.24)}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.2;opacity:var(--mdc-ripple-press-opacity,.2)}.nCP5yc.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.2)}.nCP5yc .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:#fff}@media (-ms-high-contrast:active),screen and (forced-colors:active){.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.nCP5yc .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.nCP5yc:hover{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:hover .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:active{-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 2px 6px 2px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-fillbutton-keyshadow-color,rgba(60,64,67,.3)),0 2px 6px 2px var(--gm-fillbutton-ambientshadow-color,rgba(60,64,67,.15))}.nCP5yc:active .VfPpkd-BFbNVe-bF1uUb{opacity:0}.nCP5yc:disabled{-webkit-box-shadow:none;box-shadow:none}.nCP5yc:disabled:hover .VfPpkd-Jh9lGc::before,.nCP5yc:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.nCP5yc:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.nCP5yc:disabled .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);-webkit-box-shadow:none;box-shadow:none}.Rj2Mlf .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.Rj2Mlf:not(:disabled){color:rgb(26,115,232);color:var(--gm-hairlinebutton-ink-color,rgb(26,115,232))}.Rj2Mlf:not(:disabled){border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled):hover{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:not(:disabled).VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe,.Rj2Mlf:not(:disabled):not(.VfPpkd-ksKsZd-mWPk3d):focus{border-color:rgb(23,78,166);border-color:var(--gm-hairlinebutton-outline-color--stateful,rgb(23,78,166))}.Rj2Mlf:not(:disabled):active,.Rj2Mlf:not(:disabled):focus:active{border-color:rgb(218,220,224);border-color:var(--gm-hairlinebutton-outline-color,rgb(218,220,224))}.Rj2Mlf:disabled{color:rgba(60,64,67,.38);color:var(--gm-hairlinebutton-disabled-ink-color,rgba(60,64,67,.38))}.Rj2Mlf:disabled{border-color:rgba(60,64,67,.12);border-color:var(--gm-hairlinebutton-disabled-outline-color,rgba(60,64,67,.12))}.Rj2Mlf:hover:not(:disabled),.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.Rj2Mlf:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-hairlinebutton-ink-color--stateful,rgb(23,78,166))}.Rj2Mlf .VfPpkd-BFbNVe-bF1uUb{opacity:0}.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.Rj2Mlf .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:CanvasText}}.Rj2Mlf .VfPpkd-Jh9lGc::before,.Rj2Mlf .VfPpkd-Jh9lGc::after{background-color:rgb(26,115,232);background-color:var(--gm-hairlinebutton-state-color,rgb(26,115,232))}.Rj2Mlf:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:.04;opacity:var(--mdc-ripple-hover-opacity,.04)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-focus-opacity,.12)}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:.12;opacity:var(--mdc-ripple-press-opacity,.12)}.Rj2Mlf.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0.12)}.Rj2Mlf:disabled:hover .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled.VfPpkd-ksKsZd-XxIAqe-OWXEXe-ZmdkE .VfPpkd-Jh9lGc::before{opacity:0;opacity:var(--mdc-ripple-hover-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe .VfPpkd-Jh9lGc::before,.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):focus .VfPpkd-Jh9lGc::before{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-focus-opacity,0)}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d) .VfPpkd-Jh9lGc::after{-webkit-transition:opacity .15s linear;transition:opacity .15s linear}.Rj2Mlf:disabled:not(.VfPpkd-ksKsZd-mWPk3d):active .VfPpkd-Jh9lGc::after{-webkit-transition-duration:75ms;transition-duration:75ms;opacity:0;opacity:var(--mdc-ripple-press-opacity,0)}.Rj2Mlf:disabled.VfPpkd-ksKsZd-mWPk3d{--mdc-ripple-fg-opacity:var(--mdc-ripple-press-opacity,0)}.b9hyVd{font-family:"Google Sans",Roboto,Arial,sans-serif;font-size:.875rem;letter-spacing:.0107142857em;font-weight:500;text-transform:none;-webkit-transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1);transition:border .28s cubic-bezier(.4,0,.2,1),box-shadow .28s cubic-bezier(.4,0,.2,1),-webkit-box-shadow .28s cubic-bezier(.4,0,.2,1);border-width:0;-webkit-box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);box-shadow:0 1px 2px 0 rgba(60,64,67,.3),0 1px 3px 1px rgba(60,64,67,.15);-webkit-box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15));box-shadow:0 1px 2px 0 var(--gm-protectedbutton-keyshadow-color,rgba(60,64,67,.3)),0 1px 3px 1px var(--gm-protectedbutton-ambientshadow-color,rgba(60,64,67,.15))}.b9hyVd .VfPpkd-Jh9lGc{height:100%;position:absolute;overflow:hidden;width:100%;z-index:0}.b9hyVd:not(:disabled){background-color:#fff;background-color:var(--gm-protectedbutton-container-color,#fff)}.b9hyVd:not(:disabled){color:rgb(26,115,232);color:var(--gm-protectedbutton-ink-color,rgb(26,115,232))}.b9hyVd:disabled{background-color:rgba(60,64,67,.12);background-color:var(--gm-protectedbutton-disabled-container-color,rgba(60,64,67,.12))}.b9hyVd:disabled{color:rgba(60,64,67,.38);color:var(--gm-protectedbutton-disabled-ink-color,rgba(60,64,67,.38))}.b9hyVd:hover:not(:disabled),.b9hyVd.VfPpkd-ksKsZd-mWPk3d-OWXEXe-AHe6Kc-XpnDCe:not(:disabled),.b9hyVd:not(.VfPpkd-ksKsZd-mWPk3d):focus:not(:disabled),.b9hyVd:active:not(:disabled){color:rgb(23,78,166);color:var(--gm-protectedbutton-ink-color--stateful,rgb(23,78,166))}.b9hyVd .VfPpkd-BFbNVe-bF1uUb{opacity:0}.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-Bd00G{stroke:rgb(26,115,232)}@media (-ms-high-contrast:active),screen and (forced-colors:active){.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-uI4vCe-LkdAo,.b9hyVd .VfPpkd-UdE5de-uDEFge .VfPpkd-JGcpL-IdXvz-LkdAo-B
2026-01-13T09:29:19
https://huggingface.co/datasets/123olp/binance-futures-ohlcv-2018-2026
123olp/binance-futures-ohlcv-2018-2026 · Datasets at Hugging Face Hugging Face Models Datasets Spaces Community Docs Enterprise Pricing Log In Sign Up Datasets: 123olp / binance-futures-ohlcv-2018-2026 like 96 Tasks: Time Series Forecasting Modalities: Tabular Text Time-series Formats: csv Languages: English Chinese Size: 100M - 1B Tags: cryptocurrency bitcoin ethereum trading binance ohlcv + 3 Libraries: Datasets Dask Polars + 1 License: mit Dataset card Data Studio Files Files and versions xet Community 1 Dataset Viewer (First 5GB) Auto-converted to Parquet API Embed Duplicate Data Studio Subset (1) default · ~425M rows (showing the first 23M) default (~425M rows, showing the first 23M) Split (1) train · ~425M rows (showing the first 23M) train (~425M rows, showing the first 23M) SQL Console exchange string classes 1 value symbol string classes 47 values bucket_ts string lengths 22 22 open float64 0 69.2k high float64 0 69.2k low float64 0 68.9k close float64 0 69.2k volume float64 0 3.38B quote_volume float64 0 747M trade_count int64 0 346k is_closed string classes 1 value source string classes 2 values ingested_at string classes 47 values updated_at string classes 47 values taker_buy_volume float64 0 1.76B taker_buy_quote_volume float64 0 478M binance_futures_um BCHUSDT 2020-01-01 00:00:00+00 204.72 204.72 204.46 204.56 239.555 49,016.59418 37 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 89.258 18,257.06565 binance_futures_um BCHUSDT 2020-01-01 00:01:00+00 204.54 204.65 204.51 204.52 217.005 44,392.20173 21 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 150.82 30,855.12099 binance_futures_um BCHUSDT 2020-01-01 00:02:00+00 204.51 204.51 204.24 204.32 296.062 60,516.27492 37 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 31.025 6,338.9288 binance_futures_um BCHUSDT 2020-01-01 00:03:00+00 204.38 204.39 204.31 204.33 148.643 30,376.09949 26 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 107.002 21,867.89354 binance_futures_um BCHUSDT 2020-01-01 00:04:00+00 204.3 204.3 203.9 203.91 320.277 65,345.60006 39 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 22.35 4,560.61402 binance_futures_um BCHUSDT 2020-01-01 00:05:00+00 203.86 203.99 203.85 203.88 139.082 28,360.83888 19 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 69.717 14,217.26754 binance_futures_um BCHUSDT 2020-01-01 00:06:00+00 203.86 203.86 203.6 203.74 587.302 119,625.31372 87 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 407.215 82,940.16665 binance_futures_um BCHUSDT 2020-01-01 00:07:00+00 203.81 203.89 203.78 203.89 116.218 23,692.77598 16 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 96.933 19,761.68996 binance_futures_um BCHUSDT 2020-01-01 00:08:00+00 204.03 204.19 204.03 204.09 10.318 2,105.73306 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 5.774 1,178.4114 binance_futures_um BCHUSDT 2020-01-01 00:09:00+00 204.09 204.1 203.99 204.04 368.717 75,245.46317 32 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 12.742 2,599.87768 binance_futures_um BCHUSDT 2020-01-01 00:10:00+00 204.07 204.1 203.95 204.01 190.556 38,874.72339 27 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 50.585 10,322.86815 binance_futures_um BCHUSDT 2020-01-01 00:11:00+00 203.89 204.01 203.89 204 63.343 12,918.81176 16 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 49.149 10,024.6279 binance_futures_um BCHUSDT 2020-01-01 00:12:00+00 204 204.04 203.87 203.87 290.726 59,303.5175 27 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 187.498 38,246.9198 binance_futures_um BCHUSDT 2020-01-01 00:13:00+00 203.87 203.94 203.84 203.94 65.369 13,327.50079 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 48.122 9,811.44888 binance_futures_um BCHUSDT 2020-01-01 00:14:00+00 203.92 204.02 203.92 203.97 74.187 15,131.41293 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 43.657 8,903.18642 binance_futures_um BCHUSDT 2020-01-01 00:15:00+00 203.97 204.12 203.97 204.12 104.424 21,309.09531 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 98.304 20,060.66521 binance_futures_um BCHUSDT 2020-01-01 00:16:00+00 204.12 204.18 204.12 204.17 67.247 13,728.59728 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 12.796 2,612.62187 binance_futures_um BCHUSDT 2020-01-01 00:17:00+00 204.19 204.22 204.17 204.17 65.558 13,387.05842 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 19.677 4,018.43694 binance_futures_um BCHUSDT 2020-01-01 00:18:00+00 204.18 204.18 203.87 203.88 107.805 22,003.25262 21 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 13.189 2,692.30861 binance_futures_um BCHUSDT 2020-01-01 00:19:00+00 203.87 204 203.87 203.99 105.26 21,466.92982 22 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 94.085 19,188.40678 binance_futures_um BCHUSDT 2020-01-01 00:20:00+00 203.94 204 203.94 204 22.725 4,635.62356 8 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 18.867 3,848.75738 binance_futures_um BCHUSDT 2020-01-01 00:21:00+00 204 204.02 203.96 203.99 93.314 19,035.90975 16 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 90.021 18,364.19255 binance_futures_um BCHUSDT 2020-01-01 00:22:00+00 204.02 204.06 203.99 204 64.249 13,106.83714 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 16.175 3,300.05914 binance_futures_um BCHUSDT 2020-01-01 00:23:00+00 203.99 204.03 203.99 204.03 35.725 7,288.14368 10 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 11.536 2,353.69008 binance_futures_um BCHUSDT 2020-01-01 00:24:00+00 204 204 203.96 203.96 55.086 11,236.82044 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 1.03 210.0994 binance_futures_um BCHUSDT 2020-01-01 00:25:00+00 203.96 204.16 203.96 204.13 304.107 62,047.7813 30 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 275.496 56,208.07551 binance_futures_um BCHUSDT 2020-01-01 00:26:00+00 204.13 204.27 204.13 204.13 18.218 3,720.32924 9 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 10.635 2,172.41145 binance_futures_um BCHUSDT 2020-01-01 00:27:00+00 204.13 204.13 203.97 204.05 544.207 111,086.88311 16 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 8.494 1,733.19808 binance_futures_um BCHUSDT 2020-01-01 00:28:00+00 204.05 204.23 203.97 203.97 64.767 13,218.51295 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 56.663 11,564.98694 binance_futures_um BCHUSDT 2020-01-01 00:29:00+00 203.97 204.15 203.97 204.08 10.322 2,106.43137 10 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 3.533 721.18919 binance_futures_um BCHUSDT 2020-01-01 00:30:00+00 204.02 204.49 204.02 204.49 38.125 7,786.5164 14 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 31.338 6,400.35339 binance_futures_um BCHUSDT 2020-01-01 00:31:00+00 204.49 204.55 204.43 204.43 208.886 42,715.98115 31 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 4.938 1,010.0679 binance_futures_um BCHUSDT 2020-01-01 00:32:00+00 204.42 204.55 204.42 204.47 195.294 39,933.58953 27 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 183.314 37,484.22003 binance_futures_um BCHUSDT 2020-01-01 00:33:00+00 204.47 204.47 204.47 204.47 10.644 2,176.37868 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 0 0 binance_futures_um BCHUSDT 2020-01-01 00:34:00+00 204.47 204.51 204.47 204.47 19.548 3,997.19698 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 5.481 1,120.91749 binance_futures_um BCHUSDT 2020-01-01 00:35:00+00 204.47 204.51 204.44 204.46 68.939 14,095.17869 14 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 18.036 3,688.02451 binance_futures_um BCHUSDT 2020-01-01 00:36:00+00 204.48 204.58 204.45 204.47 99.246 20,295.57063 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 87.426 17,878.93523 binance_futures_um BCHUSDT 2020-01-01 00:37:00+00 204.52 204.67 204.52 204.67 27.256 5,576.44382 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 20.939 4,284.16825 binance_futures_um BCHUSDT 2020-01-01 00:38:00+00 204.65 204.7 204.63 204.68 20.151 4,123.88491 12 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 7.079 1,448.80865 binance_futures_um BCHUSDT 2020-01-01 00:39:00+00 204.68 204.68 204.58 204.58 155.018 31,720.23144 22 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 10.091 2,065.23495 binance_futures_um BCHUSDT 2020-01-01 00:40:00+00 204.58 204.7 204.58 204.64 205.712 42,102.47703 28 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 134.334 27,494.22196 binance_futures_um BCHUSDT 2020-01-01 00:41:00+00 204.72 204.72 204.68 204.68 30.73 6,290.71132 10 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 26.951 5,517.2256 binance_futures_um BCHUSDT 2020-01-01 00:42:00+00 204.71 204.76 204.61 204.66 224.947 46,040.36672 27 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 136.69 27,974.0838 binance_futures_um BCHUSDT 2020-01-01 00:43:00+00 204.63 204.78 204.63 204.64 105.479 21,590.13948 18 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 44.463 9,102.01386 binance_futures_um BCHUSDT 2020-01-01 00:44:00+00 204.64 204.65 204.64 204.64 16.378 3,351.62446 10 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 0 0 binance_futures_um BCHUSDT 2020-01-01 00:45:00+00 204.64 204.64 204.55 204.62 168.428 34,463.06674 21 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 6.477 1,325.38128 binance_futures_um BCHUSDT 2020-01-01 00:46:00+00 204.63 204.63 204.54 204.54 38.57 7,891.65916 13 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 17.815 3,645.06984 binance_futures_um BCHUSDT 2020-01-01 00:47:00+00 204.53 204.56 204.51 204.56 49.875 10,200.79449 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 17.114 3,500.5605 binance_futures_um BCHUSDT 2020-01-01 00:48:00+00 204.54 204.56 204.54 204.55 72.549 14,840.20889 14 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 69.08 14,130.65963 binance_futures_um BCHUSDT 2020-01-01 00:49:00+00 204.55 204.56 204.54 204.55 22.682 4,639.59661 10 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 18.26 3,735.10537 binance_futures_um BCHUSDT 2020-01-01 00:50:00+00 204.57 204.62 204.55 204.6 32.759 6,702.21318 9 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 17.105 3,499.59752 binance_futures_um BCHUSDT 2020-01-01 00:51:00+00 204.6 204.61 204.53 204.53 82.167 16,809.42675 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 31.9 6,526.79837 binance_futures_um BCHUSDT 2020-01-01 00:52:00+00 204.51 204.51 204.36 204.38 149.249 30,508.70857 23 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 103.358 21,124.66155 binance_futures_um BCHUSDT 2020-01-01 00:53:00+00 204.37 204.45 204.33 204.38 596.781 121,963.0557 36 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 466.798 95,396.77617 binance_futures_um BCHUSDT 2020-01-01 00:54:00+00 204.41 204.42 204.34 204.34 141.255 28,871.9842 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 86.032 17,584.77262 binance_futures_um BCHUSDT 2020-01-01 00:55:00+00 204.34 204.4 204.34 204.4 70.541 14,416.13086 12 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 58.724 12,001.35358 binance_futures_um BCHUSDT 2020-01-01 00:56:00+00 204.42 204.51 204.37 204.37 114.337 23,370.5141 24 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 7.09 1,449.64236 binance_futures_um BCHUSDT 2020-01-01 00:57:00+00 204.37 204.4 204.27 204.4 138.771 28,355.00535 14 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 129.827 26,527.21827 binance_futures_um BCHUSDT 2020-01-01 00:58:00+00 204.4 204.66 204.4 204.49 23.101 4,724.82414 12 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 6.868 1,404.88297 binance_futures_um BCHUSDT 2020-01-01 00:59:00+00 204.5 204.61 204.49 204.61 14.381 2,941.05151 11 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 1.789 366.04729 binance_futures_um BCHUSDT 2020-01-01 01:00:00+00 204.51 204.51 204.51 204.51 9.1 1,861.041 9 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 0 0 binance_futures_um BCHUSDT 2020-01-01 01:01:00+00 204.51 205.14 204.47 204.94 449.636 92,108.4888 57 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 342.034 70,058.87896 binance_futures_um BCHUSDT 2020-01-01 01:02:00+00 205.01 205.62 205.01 205.5 177.57 36,439.55562 47 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 135.444 27,799.05292 binance_futures_um BCHUSDT 2020-01-01 01:03:00+00 205.59 205.59 205.18 205.52 246.126 50,556.12912 40 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 38.641 7,933.63812 binance_futures_um BCHUSDT 2020-01-01 01:04:00+00 205.43 205.65 205.43 205.65 199.948 41,089.17166 24 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 169.37 34,805.30198 binance_futures_um BCHUSDT 2020-01-01 01:05:00+00 205.65 205.7 205.31 205.31 496.922 102,111.0587 59 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 22.443 4,615.4092 binance_futures_um BCHUSDT 2020-01-01 01:06:00+00 205.3 205.55 205.07 205.53 337.032 69,201.37635 78 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 292.058 59,966.4546 binance_futures_um BCHUSDT 2020-01-01 01:07:00+00 205.5 205.79 205.5 205.67 181.003 37,227.9314 30 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 88.921 18,291.05242 binance_futures_um BCHUSDT 2020-01-01 01:08:00+00 205.61 205.61 205.48 205.53 332.045 68,253.43997 45 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 134.475 27,639.45266 binance_futures_um BCHUSDT 2020-01-01 01:09:00+00 205.52 205.87 205.52 205.58 433.664 89,190.00456 62 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 275.104 56,581.63812 binance_futures_um BCHUSDT 2020-01-01 01:10:00+00 205.55 205.7 205.55 205.67 172.911 35,559.67602 26 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 83.868 17,248.90248 binance_futures_um BCHUSDT 2020-01-01 01:11:00+00 205.69 205.8 205.66 205.73 235.027 48,344.70273 24 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 226.538 46,598.37234 binance_futures_um BCHUSDT 2020-01-01 01:12:00+00 205.73 205.77 205.7 205.71 146.621 30,164.20961 19 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 22.96 4,723.41708 binance_futures_um BCHUSDT 2020-01-01 01:13:00+00 205.7 205.74 205.7 205.71 13.566 2,790.84333 8 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 8.383 1,724.67509 binance_futures_um BCHUSDT 2020-01-01 01:14:00+00 205.71 205.73 205.63 205.63 129.23 26,577.88482 22 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 7.168 1,474.37214 binance_futures_um BCHUSDT 2020-01-01 01:15:00+00 205.62 205.73 205.49 205.72 453.527 93,232.98858 71 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 335.122 68,892.44661 binance_futures_um BCHUSDT 2020-01-01 01:16:00+00 205.73 205.73 205.69 205.69 144.458 29,717.96832 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 102.842 21,156.74499 binance_futures_um BCHUSDT 2020-01-01 01:17:00+00 205.73 205.82 205.73 205.82 65.722 13,523.66473 15 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 18.939 3,897.02342 binance_futures_um BCHUSDT 2020-01-01 01:18:00+00 205.82 205.82 205.76 205.76 102.989 21,193.07252 19 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 5.717 1,176.56633 binance_futures_um BCHUSDT 2020-01-01 01:19:00+00 205.76 206.02 205.76 206.02 289.151 59,519.8541 31 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 232.051 47,763.98364 binance_futures_um BCHUSDT 2020-01-01 01:20:00+00 206 206.22 206 206.09 569.969 117,492.11857 49 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 312.737 64,466.36341 binance_futures_um BCHUSDT 2020-01-01 01:21:00+00 206.14 206.14 205.92 205.96 363.572 74,900.7002 36 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 41.002 8,445.23939 binance_futures_um BCHUSDT 2020-01-01 01:22:00+00 205.96 206.22 205.71 205.79 707.253 145,717.47629 77 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 436.995 90,044.79249 binance_futures_um BCHUSDT 2020-01-01 01:23:00+00 205.71 205.86 205.52 205.78 153.778 31,636.67422 33 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 62.397 12,842.66081 binance_futures_um BCHUSDT 2020-01-01 01:24:00+00 205.78 205.79 205.65 205.71 267.743 55,071.8762 24 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 247.547 50,916.08066 binance_futures_um BCHUSDT 2020-01-01 01:25:00+00 205.71 206.68 205.71 206.59 980.99 202,338.25987 141 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 712.653 146,988.83153 binance_futures_um BCHUSDT 2020-01-01 01:26:00+00 206.58 206.58 206.2 206.2 229.953 47,456.78952 52 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 74.223 15,319.34667 binance_futures_um BCHUSDT 2020-01-01 01:27:00+00 206.32 207.29 206.21 207.09 785.006 162,306.60687 106 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 602.672 124,619.98571 binance_futures_um BCHUSDT 2020-01-01 01:28:00+00 207.02 207.28 206.73 206.88 700.282 144,935.58407 125 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 400.322 82,848.74776 binance_futures_um BCHUSDT 2020-01-01 01:29:00+00 206.86 207.16 206.62 206.62 572.781 118,498.205 97 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 327.419 67,738.30283 binance_futures_um BCHUSDT 2020-01-01 01:30:00+00 206.71 207.26 206.68 207.24 329.615 68,208.89178 58 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 263.988 54,638.11474 binance_futures_um BCHUSDT 2020-01-01 01:31:00+00 207.18 207.18 206.72 206.72 420.66 87,084.81333 79 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 131.973 27,326.10057 binance_futures_um BCHUSDT 2020-01-01 01:32:00+00 206.82 206.93 206.26 206.43 527.637 108,935.94421 69 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 250.829 51,770.98892 binance_futures_um BCHUSDT 2020-01-01 01:33:00+00 206.54 206.6 206.09 206.28 383.564 79,124.47195 55 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 260.182 53,676.81898 binance_futures_um BCHUSDT 2020-01-01 01:34:00+00 206.3 206.3 206.18 206.2 181.579 37,447.1228 30 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 80.2 16,541.40171 binance_futures_um BCHUSDT 2020-01-01 01:35:00+00 206.2 206.29 206.06 206.06 293.648 60,559.9187 44 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 122.286 25,222.24973 binance_futures_um BCHUSDT 2020-01-01 01:36:00+00 206.01 206.34 206.01 206.28 188.771 38,926.29793 47 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 175.486 36,187.97785 binance_futures_um BCHUSDT 2020-01-01 01:37:00+00 206.29 206.29 206.09 206.12 163.473 33,707.57901 25 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 4.38 902.95415 binance_futures_um BCHUSDT 2020-01-01 01:38:00+00 206.11 206.33 206.09 206.32 230.702 47,560.23683 34 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 205.938 42,451.33629 binance_futures_um BCHUSDT 2020-01-01 01:39:00+00 206.4 206.44 206.21 206.21 198.062 40,872.33116 26 t binance_zip 2025-11-13 19:27:00.723115+00 2025-11-13 19:27:00.723115+00 17.787 3,671.50048 End of preview. Expand in Data Studio Previous 1 2 3 ... 230,200 Next 📊 数据概览 🎯 包含币种 📁 数据字段 📈 candles_1m (1分钟K线) 📊 futures_metrics_5m (期货指标) 🚀 快速使用 数据格式说明 恢复到 TimescaleDB 或使用恢复脚本 查询示例 📈 应用场景 🔄 数据来源 📦 完整版数据集 ⚠️ 免责声明 🔗 相关链接 📜 许可证 ❤️ 支持项目 🐱 币安期货 Main4 数据集 (BTC/ETH/BNB/SOL) 精简版币安期货历史数据 - 只包含 4 个主要币种,适合快速下载和研究使用。 📊 数据概览 文件 记录数 压缩大小 时间范围 说明 candles_1m_main4_*.bin.zst 998 万 366 MB 2020-01 ~ 2026-01 1分钟K线 (Binary) futures_metrics_main4_*.bin.zst 152 万 49 MB 2021-12 ~ 2026-01 期货指标 (Binary) schema_*.sql.zst - 6.3 KB - TimescaleDB Schema 总计 : 1150 万条记录,压缩后约 415MB 🎯 包含币种 币种 K线记录数 K线时间范围 期货指标记录数 完整度 BTCUSDT 317 万 2020-01-01 ~ 2026-01-11 43 万 100% ETHUSDT 317 万 2020-01-01 ~ 2026-01-11 24 万 100% BNBUSDT 311 万 2020-02-10 ~ 2026-01-11 43 万 100% SOLUSDT 53 万 2025-01-01 ~ 2026-01-04 43 万 99.99% 📁 数据字段 📈 candles_1m (1分钟K线) 字段 类型 说明 示例 exchange string 交易所 binance_futures_um symbol string 交易对 BTCUSDT bucket_ts timestamp K线时间 (UTC) 2024-01-01 00:00:00 open decimal 开盘价 42150.50 high decimal 最高价 42180.00 low decimal 最低价 42100.00 close decimal 收盘价 42165.30 volume decimal 成交量 (币) 125.5 quote_volume decimal 成交额 (USDT) 5289150.25 trade_count int 成交笔数 3521 taker_buy_volume decimal 主动买入量 68.2 taker_buy_quote_volume decimal 主动买入额 2875420.10 📊 futures_metrics_5m (期货指标) 字段 类型 说明 用途 create_time timestamp 时间戳 (UTC) 数据时间 symbol string 交易对 币种标识 sum_open_interest decimal 持仓量 (张) 市场参与度 sum_open_interest_value decimal 持仓价值 (USDT) 资金规模 sum_toptrader_long_short_ratio decimal 大户多空比 主力方向 sum_taker_long_short_vol_ratio decimal 主动买卖比 即时情绪 count_long_short_ratio decimal 散户多空比 散户情绪 🚀 快速使用 数据格式说明 本数据集使用 PostgreSQL COPY Binary + Zstandard 格式压缩,需要导入 TimescaleDB 使用。 恢复到 TimescaleDB # 1. 解压并恢复 schema zstd -d schema_*.sql.zst -c | psql -h localhost -p 5433 -U postgres -d market_data # 2. 恢复 K线数据 zstd -d candles_1m_main4_*.bin.zst -c | psql -h localhost -p 5433 -U postgres -d market_data \ -c "COPY market_data.candles_1m FROM STDIN WITH (FORMAT binary)" # 3. 恢复期货指标 zstd -d futures_metrics_main4_*.bin.zst -c | psql -h localhost -p 5433 -U postgres -d market_data \ -c "COPY market_data.binance_futures_metrics_5m FROM STDIN WITH (FORMAT binary)" 或使用恢复脚本 chmod +x restore_main4_*.sh ./restore_main4_*.sh 查询示例 -- 查看 BTC 最近数据 SELECT * FROM market_data.candles_1m WHERE symbol = 'BTCUSDT' ORDER BY bucket_ts DESC LIMIT 10 ; -- 统计各币种数据量 SELECT symbol, COUNT ( * ), MIN (bucket_ts), MAX (bucket_ts) FROM market_data.candles_1m GROUP BY symbol; -- BTC 日K聚合 SELECT time_bucket( '1 day' , bucket_ts) AS day , FIRST ( open , bucket_ts) AS open , MAX (high) AS high, MIN (low) AS low, LAST ( close , bucket_ts) AS close , SUM (volume) AS volume FROM market_data.candles_1m WHERE symbol = 'BTCUSDT' GROUP BY day ORDER BY day DESC ; 📈 应用场景 场景 说明 🤖 量化回测 策略历史验证,分钟级精度 🧠 机器学习 价格预测、趋势分类、异常检测 📊 技术分析 计算各类技术指标 (MA/RSI/MACD等) 💹 市场研究 波动率分析、相关性分析 📉 情绪分析 多空比、持仓量变化分析 🔄 数据来源 交易所 : Binance Futures (USDT-M 永续合约) 采集方式 : WebSocket 实时推送 + REST API 历史回填 数据延迟 : < 5 秒 导出时间 : 2026-01-11 📦 完整版数据集 如需全部 615+ 币种的完整数据,请查看: Kaggle : binance-crypto-ohlcv-futures ⚠️ 免责声明 本数据集仅供教育和研究目的 不构成任何投资建议 加密货币交易存在重大风险 使用前请自行验证数据准确性 🔗 相关链接 📦 项目源码 : TradeCat 📢 Telegram 频道 : @tradecat_ai_channel 💬 交流群 : @glue_coding 🐦 Twitter : @123olp 📜 许可证 MIT License - 可自由用于商业和研究用途 ❤️ 支持项目 如果这个数据集对你有帮助,欢迎: ⭐ 给项目点 Star 📢 分享给更多人 网络 地址 币安 UID 572155580 Tron (TRC20) TQtBXCSTwLFHjBqTS4rNUp7ufiGx51BRey Solana HjYhozVf9AQmfv7yv79xSNs6uaEU5oUk2USasYQfUYau Ethereum (ERC20) 0xa396923a71ee7D9480b346a17dDeEb2c0C287BBC Downloads last month 478 Use this dataset Size of the auto-converted Parquet files (First 5GB): 1.43 GB Number of rows (First 5GB): 23,020,000 Estimated number of rows: 424,919,005 System theme Company TOS Privacy About Careers Website Models Datasets Spaces Pricing Docs
2026-01-13T09:29:19
https://git-scm.com/book/az/v2/GitHub-GitHub-Skriptl%c9%99m%c9%99
Git - GitHub Skriptləmə About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Başlanğıc 1.1 Versiyaya Nəzarət Haqqında 1.2 Git’in Qısa Hekayəsi 1.3 Git Nədir? 1.4 Əmr Sətiri 1.5 Git’i Quraşdırmaq 1.6 İlk Dəfə Git Quraşdırması 1.7 Kömək Almaq 1.8 Qısa Məzmun 2. Git’in Əsasları 2.1 Git Deposunun Əldə Edilməsi 2.2 Depoda Dəyişikliklərin Qeyd Edilməsi 2.3 Commit Tarixçəsinə Baxış 2.4 Ləğv Edilən İşlər (Geri qaytarılan) 2.5 Uzaqdan İşləmək 2.6 Etiketləmə 2.7 Git Alias’lar 2.8 Qısa Məzmun 3. Git’də Branch 3.1 Nutshell’də Branch’lar 3.2 Sadə Branching və Birləşdirmə 3.3 Branch İdarəedilməsi 3.4 Branching İş Axınları 3.5 Uzaq Branch’lar 3.6 Rebasing 3.7 Qısa Məzmun 4. Server’də Git 4.1 Protokollar 4.2 Serverdə Git Əldə Etmək 4.3 Sizin öz SSH Public Key’nizi yaratmaq 4.4 Server qurmaq 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Üçüncü Tərəf Seçimləri 4.10 Qısa Məzmun 5. Paylanmış Git 5.1 Distribyutorluq İş Axınları 5.2 Layihəyə Töhfə vermək 5.3 Layihənin Saxlanılması 5.4 Qısa Məzmun 6. GitHub 6.1 Hesab Qurma və Konfiqurasiya 6.2 Bir Layihəyə Töhfə Vermək 6.3 Bir Layihənin Saxlanılması 6.4 Bir Təşkilatı Idarə Etmək 6.5 GitHub Skriptləmə 6.6 Qısa Məzmun 7. Git Alətləri 7.1 Reviziya Seçimi 7.2 Interaktiv Səhnələşdirmə 7.3 Stashing və Təmizləmə 7.4 İşinizin İmzalanması 7.5 Axtarış 7.6 Tarixi Yenidən Yazmaq 7.7 Reset Demystified 7.8 İnkişaf etmiş Birləşmə 7.9 Rerere 7.10 Git ilə Debugging 7.11 Alt Modullar 7.12 Bundling 7.13 Dəyişdirmək 7.14 Etibarlı Yaddaş 7.15 Qısa Məzmun 8. Git’i Fərdiləşdirmək 8.1 Git Konfiqurasiyası 8.2 Git Atributları 8.3 Git Hook’ları 8.4 Git-Enforced Siyasət Nümunəsi 8.5 Qısa Məzmun 9. Git və Digər Sistemlər 9.1 Git Müştəri kimi 9.2 Git’ə Miqrasiya 9.3 Qısa Məzmun 10. Git’in Daxili İşləri 10.1 Plumbing və Porcelain 10.2 Git Obyektləri 10.3 Git Referansları 10.4 Packfile’lar 10.5 Refspec 10.6 Transfer Protokolları 10.7 Maintenance və Məlumatların Bərpası 10.8 Mühit Dəyişənləri 10.9 Qısa Məzmun A1. Appendix A: Digər Mühitlərdə Git A1.1 Qrafik interfeyslər A1.2 Visual Studio’da Git A1.3 Visual Studio Code’da Git A1.4 Eclipse’də Git A1.5 Sublime Text’də Git A1.6 Bash’da Git A1.7 Zsh’də Git A1.8 PowerShell’də Git A1.9 Qısa Məzmun A2. Appendix B: Proqramlara Git Daxil Etmək A2.1 Əmr-sətri Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Əmrləri A3.1 Quraşdırma və Konfiqurasiya A3.2 Layihələrin Alınması və Yaradılması A3.3 Sadə Snapshotting A3.4 Branching və Birləşmə A3.5 Layihələrin Paylaşılması və Yenilənməsi A3.6 Yoxlama və Müqayisə A3.7 Debugging A3.8 Patching A3.9 E-poçt A3.10 Xarici Sistemlər A3.11 İdarəetmə A3.12 Plumbing Əmrləri 2nd Edition 6.5 GitHub - GitHub Skriptləmə GitHub Skriptləmə Beləliklə, indi GitHub’un əsas xüsusiyyətləri və iş axınlarının hamısını əhatə etdik, ancaq hər hansı bir böyük qrup və ya layihənin istədikləri və ya inteqrasiya etmək istədikləri xarici xidmətlər ola bilər. Bizim üçün xoşbəxtlikdən, GitHub bir çox cəhətdən həqiqətən hack-lənə bilər. Bu hissədə GitHub-un istədiyi şəkildə işləməsi üçün GitHub hooks sistemindən və onun API-dən necə istifadə edəcəyimizi əhatə edəcəyik. Servislər and Hook-lar GitHub depo idarəsinin Servislər and Hook-larr bölməsi GitHub’un xarici sistemlərlə qarşılıqlı əlaqəsinin ən asan yoludur. Servislər Əvvəlcə Servislərə nəzər salacağıq. Həm Hook-lar, həm də Servislər inteqrasiyasını əvvəlcədən Həmkarlar əlavə etmək və layihənizin standart bölməsini dəyişdirməyə baxdığımız deponun Ayarlar bölməsində tapa bilərsiniz. “Webhooks and Services” bölməsində Servislər and Hook-lar konfiqurasiya bölməsi kimi bir şey görəcəksiniz. Figure 130. Servislər and Hook-lar konfiqurasiya bölməsi Seçdiyiniz onlarla servis var, əksəriyyəti digər kommersiya və açıq mənbə sistemlərinə inteqrasiya edir. Onların əksəriyyəti Davamlı İnteqrasiya xidmətləri, səhv və problem izləyiciləri, söhbət otağı sistemləri və sənədləşdirmə sistemləri üçündür. E-poçt fork-unu quraşdırmaqla çox sadə bir şəkildə keçəcəyik. “Add Service” açılan menyudan “email” seçsəniz, E-poçt xidmətinin konfiqurasiyası kimi bir konfiqurasiya ekranı alacaqsınız. Figure 131. E-poçt xidmətinin konfiqurasiyası Bu vəziyyətdə, “Add service” düyməsini vursaq, göstərilən e-poçt ünvanı, hər kimsə depoya push edərkən hər dəfə bir e-poçt alacaq. Servislər çox sayda müxtəlif hadisəni dinləyə bilər, ancaq çoxu push etmə hadisələrini dinləyir və sonra bu məlumatlarla bir şey edə bilər. GitHub ilə inteqrasiya etmək istədiyiniz bir sistem varsa, mövcud bir servis inteqrasiyasının olub olmadığını görmək üçün buranı yoxlamalısınız. Məsələn, kod bazasında testlər aparmaq üçün Jenkins-dan istifadə edirsinizsə, Jenkins-inin servis inteqrasiyasını kimsə depolarına push etdikdə hər dəfə bir test işə salmağa imkan verə bilərsiniz. Hook-lar Daha spesifik bir şeyə ehtiyacınız varsa və ya bu siyahıya daxil olmayan bir servis və ya saytla inteqrasiya etmək istəsəniz, bunun əvəzinə daha ümumi hook-lar sistemindən istifadə edə bilərsiniz. GitHub depo hook-ları olduqca sadədir. Bir URL göstərmisinizsə və GitHub istədiyiniz hər hansı bir hadisədə HTTP yüklənməsini həmin URL-ə göndərəcəkdir. Ümumiyyətlə bu iş üsulu, GitHub fork yükünü dinləmək və qəbul edildikdə məlumatla bir şey etmək üçün kiçik bir veb xidməti qura bilərsiniz. Bir fork aktivləşdirmək üçün Servislər and Hook-lar konfiqurasiya bölməsi -dakı “Add webhook” düyməsini klikləyin. Bu sizi Web hook konfiqurasiyası kimi görünən bir səhifəyə gətirəcəkdir. Figure 132. Web hook konfiqurasiyası Bir web hook üçün konfiqurasiya olduqca sadədir. Əksər hallarda sadəcə bir URL və gizli bir açar daxil edir və “Add webhook” düyməsini vurursunuz. GitHub-ın yükləməsini istədiyiniz hadisələrin bir neçə variantı var - varsayılan kimsə deponuzun hər hansı bir branch-ına yeni kodu push hadisəsi üçün bir yük almaqdır. Bir web hook-u idarə etmək üçün qura biləcəyiniz bir veb servisin kiçik bir nümunəsini baxaq. Ruby web framework Sinatra’dan istifadə edəcəyik, çünki olduqca qısadır və etdiyimiz işləri asanlıqla görə bilərsiniz. Müəyyən bir şəxs müəyyən bir sənəd dəyişdirərək layihəmizin müəyyən bir branch-na push etsə, e-poçt almaq istədiyimizi deyək. Bunu kifayət qədər asanlıqla bu kimi kodla edə bilərik: require 'sinatra' require 'json' require 'mail' post '/payload' do push = JSON.parse(request.body.read) # parse the JSON # gather the data we're looking for pusher = push["pusher"]["name"] branch = push["ref"] # get a list of all the files touched files = push["commits"].map do |commit| commit['added'] + commit['modified'] + commit['removed'] end files = files.flatten.uniq # check for our criteria if pusher == 'schacon' && branch == 'ref/heads/special-branch' && files.include?('special-file.txt') Mail.deliver do from 'tchacon@example.com' to 'tchacon@example.com' subject 'Scott Changed the File' body "ALARM" end end end Burada GitHub-un bizə verdiyi JSON yükünü götürürük və onu kimin push etdiyini, hansı branch-a push etdiyini və push etdikləri bütün sənədlərdə hansı sənədlərə toxunduğunu axtarırıq. Sonra meyarlarımıza uyğun olduğunu yoxlayırıq və uyğun olduqda bir e-poçt göndəririk. Bu kimi bir şeyi inkişaf etdirmək və sınamaq üçün, fork hazırladığınız eyni ekranda gözəl bir inkişaf etdirici konsolunuz var. GitHub’un webhook üçün etdiyi cəhdləri görə bilərsiniz. Hər bir fork üçün təhvil verildiyi anda istəyə və cavaba cavab verən body və başlıqları incəliyə bilərsiniz. Bu, hook-ları test və debug etməyi olduqca asanlaşdırır. Figure 133. Web hook debugging informasiyası Bunun digər böyük xüsusiyyəti, xidmətinizi asanlıqla sınamaq üçün hər hansı bir payload-ı geri qaytara biləcəyinizdir. Webhook-ların necə yazılacağı və dinləyə biləcəyiniz bütün fərqli hadisə növləri haqqında daha çox məlumat üçün https://developer.github.com/webhooks/ saytından GitHub Developer sənədlərinə baxın. GitHub API Servislər və hook-lar, depolarınızda baş verən hadisələr barədə push bildirişlərini almaq üçün bir yol verir, ancaq bu hadisələr haqqında daha çox məlumat lazımdırsa, nə etməlisiniz? Əməkdaşlarınızı və ya etiketləmə məsələlərini əlavə etmək kimi bir şeyi avtomatlaşdırmaq lazımdırsa, nə etməli? GitHub API lazımlı olduğu yerdir. GitHub-da veb saytında avtomatlaşdırılmış bir şəkildə edə biləcəyiniz hər şeyi etmək üçün çox sayda API nöqtəsi var. Bu bölmədə doğrulamağı və API-yə necə qoşulacağımızı, bir məsələyə necə şərh verəcəyinizi və API vasitəsilə Pull Request-in vəziyyətini necə dəyişdirəcəyinizi öyrənəcəyik. Əsas İstifadə Siz edə biləcəyiniz ən əsas şey doğrulama tələb etməyən endpoint-də sadə bir GET request-dir. Bu istifadəçi və ya açıq mənbəli bir layihədə yalnız oxu məlumatı ola bilər. Məsələn, “schacon” adlı bir istifadəçi haqqında daha çox bilmək istəyiriksə, belə bir şey işlədə bilərik: $ curl https://api.github.com/users/schacon { "login": "schacon", "id": 70, "avatar_url": "https://avatars.githubusercontent.com/u/70", # … "name": "Scott Chacon", "company": "GitHub", "following": 19, "created_at": "2008-01-27T17:19:28Z", "updated_at": "2014-06-10T02:37:23Z" } Təşkilatlar, layihələr, məsələlər, commit-lər haqqında məlumat almaq üçün bu kimi endpoint-lər var - GitHub-da hər şey haqqında açıq şəkildə görə biləcəksiniz . Siz hətta ixtiyari Markdown göstərmək üçün API-dən istifadə edə və ya .gitignore şablonunu tapa bilərsiniz. $ curl https://api.github.com/gitignore/templates/Java { "name": "Java", "source": "*.class # Mobile Tools for Java (J2ME) .mtj.tmp/ # Package Files # *.jar *.war *.ear # virtual machine crash logs, see https://www.java.com/en/download/help/error_hotspot.xml hs_err_pid* " } Bir Məsələyə Münasibət Bildirmək Bununla birlikdə, bir Issue or Pull Request şərhini şərh etmək kimi bir veb saytında bir hərəkət etmək istəsəniz və ya şəxsi məzmuna baxmaq və ya qarşılıqlı əlaqə qurmaq istəyirsinizsə, identifikasiya lazımdır. İdentifikasiyanın bir neçə yolu var. Əsas identifikasiyanı yalnız istifadəçi adınızı və şifrənizi istifadə edə bilərsiniz, ancaq ümumiyyətlə şəxsi giriş tokenindən istifadə etmək daha yaxşı bir fikirdir. Bunu parametrlər səhifənizdəki “Applications” bölməsindən əldə edə bilərsiniz. Figure 134. Parametrlər səhifənizdəki “Applications” bölməsindən giriş tokeninizi yaradın Bu işarə və təsvir üçün hansı sahələri istədiyinizi soruşacaqdır. Skriptiniz və ya tətbiqiniz artıq istifadə edilmədiyi zaman token-i çıxartmağı rahat hiss etdiyiniz üçün yaxşı bir təsviri istifadə etdiyinizə əmin olun. GitHub sizə bir token-i yalnız bir dəfə göstərəcəkdir, buna görə də onu kopyalayın. İndi istifadə edərək istifadəçi adı və şifrə istifadə etmək əvəzinə skriptinizdə identifikasiya etmək üçün istifadə edə bilərsiniz. Bu, çox xoşdur, çünki nə etmək istədiyinizi məhdudlaşdıra bilərsiniz və token-in ləğvi mümkündür. Bu da dərəcəniz həddini artırmağın əlavə üstünlüyünə malikdir. Doğrulama olmadan, saatda 60 sorğu ilə məhdudlaşacaqsınız. Doğruladığınız təqdirdə saatda 5000 müraciət edə bilərsiniz. Beləliklə, məsələlərimizdən birinə şərh etmək üçün istifadə edək. Deyək ki, 6-cı sayda müəyyən bir məsələyə münasibət bildirmək istəyirik. Bunu etmək üçün, yalnız Avtorizasiya başlığı olaraq hazırladığımız işarə ilə repos/<user>/<repo>/issues/<num>/comments -a bir HTTP POST tələb etməliyik. $ curl -H "Content-Type: application/json" \ -H "Authorization: token TOKEN" \ --data '{"body":"A new comment, :+1:"}' \ https://api.github.com/repos/schacon/blink/issues/6/comments { "id": 58322100, "html_url": "https://github.com/schacon/blink/issues/6#issuecomment-58322100", ... "user": { "login": "tonychacon", "id": 7874698, "avatar_url": "https://avatars.githubusercontent.com/u/7874698?v=2", "type": "User", }, "created_at": "2014-10-08T07:48:19Z", "updated_at": "2014-10-08T07:48:19Z", "body": "A new comment, :+1:" } İndi bu məsələyə keçsəniz, yalnız GitHub API-dən bir şərh göndərildi -da uğurla yayımladığımızı görə bilərsiniz. Figure 135. GitHub API-dən bir şərh göndərildi Veb saytından edə biləcəyiniz hər şeyi etmək üçün API-dən istifadə edə bilərsiniz - mərhələləri yaratmaq və təyin etmək, İnsanları Issues və Pull Requests-a təyin etmək, etiketlər yaratmaq və dəyişdirmək, commit məlumatlarına daxil olmaq, yeni commit-lər və branch-lar yaratmaq, açmaq, bağlamaq və ya etmək Pull Requests-in birləşməsi, qruplar yaratmaq və redaktə etmək, Pull Request-dəki kod sətirlərinə şərh vermək, saytı axtarmaq və s. Pull Request Statusunun Dəyişdirilməsi Pull Request-ləri ilə işləyirsinizsə, həqiqətən faydalı olduğundan baxacağımız bir son nümunə var. Hər bir commit-in onunla əlaqəli bir və ya daha çox statusu ola bilər və bu statusu əlavə etmək və soruşmaq üçün bir API var. Davamlı İnteqrasiya və sınaq servislərinin əksəriyyəti bu API-dən istifadə edilən kodu sınamaqla push etməərə reaksiya vermək üçün istifadə edir və sonra bu commit-in bütün sınaqlardan keçdiyini geri bildirir. Əgər təqdimatçı bütün töhfələr qaydalarına əməl edirsə, commit düzgün imzalanıbsa, göndərmə mesajının düzgün formatlandığını yoxlamaq üçün istifadə edə bilərsiniz. Deyək ki, əmanət mesajınızdakı Signed-off-by sətirini yoxlayan kiçik bir veb xidmətə xitabən depo qurduq. require 'httparty' require 'sinatra' require 'json' post '/payload' do push = JSON.parse(request.body.read) # parse the JSON repo_name = push['repository']['full_name'] # look through each commit message push["commits"].each do |commit| # look for a Signed-off-by string if /Signed-off-by/.match commit['message'] state = 'success' description = 'Successfully signed off!' else state = 'failure' description = 'No signoff found.' end # post status to GitHub sha = commit["id"] status_url = "https://api.github.com/repos/#{repo_name}/statuses/#{sha}" status = { "state" => state, "description" => description, "target_url" => "http://example.com/how-to-signoff", "context" => "validate/signoff" } HTTParty.post(status_url, :body => status.to_json, :headers => { 'Content-Type' => 'application/json', 'User-Agent' => 'tonychacon/signoff', 'Authorization' => "token #{ENV['TOKEN']}" } ) end end Ümid edirik, bunu izləmək olduqca sadədir. Bu web hook işlədicisinə, sadəcə push etdikləri hər bir commit-ə baxırıq, commit mesajında Signed-off-by sətirini axtarırıq və nəhayət HTTP vasitəsilə /repos/<user>/<repo>/statuses/<commit_sha> vəziyyəti olan API endpointi tapırıq. Bu halda bir vəziyyət ( success , failure , error ), baş verənlərin təsviri, istifadəçinin əlavə məlumat üçün gedə biləcəyi hədəf URL və bir commit üçün çox statuslu “context” göndərə bilərsiniz.. Məsələn, bir test servisi bir status təqdim edə bilər və bu kimi bir yoxlama servisi də bir status təqdim edə bilər - “context” sahəsi necə fərqləndiyini göstərir. Kimsə GitHub-da yeni Pull Request açırsa və bu fork qurulubsa, API vasitəsilə Commit status kimi bir şey görə bilərsiniz. Figure 136. API vasitəsilə Commit status İndi mesajdakı “Signed-off-by” sətri olan və yazarın imzalamağı unutduğu yerdən qırmızı xaç olan işarənin yanında bir az yaşıl bir işarə görə bilərsiniz. Ayrıca Pull Request branch-dakı son commitin vəziyyətini aldığını və uğursuz olduqda xəbərdarlıq etdiyini görə bilərsiniz. Bu API-ni test nəticələri üçün istifadə edirsinizsə, sonuncu commit-in uğursuz olduğu testləri təsadüfən birləşdirməməyiniz üçün çox faydalıdır. Octokit Bu nümunələrdə curl və sadə HTTP sorğuları ilə demək olar ki, hər şeyi etdiyimizə baxmayaraq bu API-ni daha idiomatik şəkildə təqdim edən bir neçə açıq mənbəli kitabxana mövcuddur. Bu yazı zamanı dəstəklənən dillərə Go, Objective-C, Ruby və .NET daxildir. Bunlar haqqında daha çox məlumat üçün https://github.com/octokit səhifəsinə baxın, çünki sizin üçün HTTP-nin çox hissəsi idarə olunur. Ümid edirik, bu vasitələr xüsusi iş axınlarınız üçün daha yaxşı işləmək üçün GitHub-u düzəltməyə və dəyişdirməyə kömək edə bilər. Bütün API sənədləri və ümumi tapşırıqlar üçün təlimatlar üçün https://developer.github.com səhifəsinə baxın. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
http://www.trello.com/teams/product
Trello For Product Management Teams | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog < Go back to Team Solutions Trello For Product Management Teams Bring high-quality products to market faster with Trello. Discover the most effective ways product management teams can track product roadmaps, simplify sprints, and launch new updates with ease. Trello’s boards, lists, and cards enable teams to go from ideas to action in seconds. Visual and easy-to-use, Trello helps teams bring projects to life and keep them moving forward. Join over 2,000,000 teams worldwide who are using Trello to get more done. Your Product Management Team’s Workspace Set your product management team on the right track to achieve successful product launches by harnessing the power of Trello templates. Explore all product management templates PRODUCT ROADMAP Easily track product development, feature requests, and collaborate on development processes. AGILE SPRINT Keep code, specs, and plans organized and handy, collaborate with engineers, product managers, and scrum masters, and easily share progress and future plans. FEATURE REQUESTS Track customer and client feedback in a central and visible space, enabling your sales, support, and product team to better collaborate on bug fixes, upgrades, and updates. PRODUCT ROADMAP Easily track product development, feature requests, and collaborate on development processes. AGILE SPRINT Keep code, specs, and plans organized and handy, collaborate with engineers, product managers, and scrum masters, and easily share progress and future plans. FEATURE REQUESTS Track customer and client feedback in a central and visible space, enabling your sales, support, and product team to better collaborate on bug fixes, upgrades, and updates. Clear the way for more seamless product launches. Empower your Product Management team to move projects across the finish line faster and more easily with Timeline View. Learn more about Trello views Power-Up Your Launches Spend more time launching products and less time doing administrative heavy-lifting. With Power-Ups, you can connect your favorite digital tools like Google Drive and Slack with Trello, and features like Calendar allow your teams to get a bird’s eye view of product roadmaps. Explore 150+ Power-Ups Move Work Forward, Auto-magically Trello’s built-in automation makes it easy to automate the repetitive, everyday tasks that keep your team from focusing on the work that matters most. Let the robots do the work Resources To Back Better Products How To Plan And Prioritize Your Product Roadmap In Trello [Blog Post] Get ahead of your competition by prioritizing and planning your product roadmap in Trello. Doing so will ensure your team builds the features that matter most, and ultimately enable you to achieve product excellence. READ MORE Trello Agile Series: Retrospectives and Roadmaps [Webinar] How often is your team reflecting? Discover how your product management team can use the power of Trello to effectively manage retrospectives and product roadmaps in order to make better decisions, together. Watch Now Trello Agile Series: The Sprint Board [Webinar] Discover how agile and scrum techniques can be applied to your workflows in Trello and more specifically, your team sprints. Full steam ahead! Watch Now NOW AVAILABLE A 14 Day Free Trial of Premium! Get unlimited boards, Trello views, and limitless automation, plus a ton more. Try it today SwagUp: Scale Any Business With Trello BC “Not only did [Trello] unify our process and help everyone understand their role, but we were able to automate essential steps of the process so we could move a lot faster and grow to a multi-million dollar company with a team of ten.” –Founder, SwagUp Learn More Instinct Dog Training “As you scale, you need visibility, accountability, and organization. Trello has provided that for us.” –Brian Burton, Founder, Instinct Dog Training Learn More Desk Plants “We chose Trello because it is well-designed—it’s beautifully designed—intuitive, and really hit the nail on the head with what we needed to solve.” –Lawrence Hanley, founder of Desk Plants Learn More Scan2Cad “We use Trello because our data becomes alive. A bullet point list turns into real tasks that are assigned to real people with due dates and connections to our other apps. The power in that is fantastic.” –Luke Kennedy, CEO, Scan2Cad Learn More Palace Law “People have generally been happier since we started using Trello. It has made our lives a lot easier. People are less stressed.” -Jordan Couch, attorney at Palace Law Learn More Trello Your Way Ready to get your team up and running? Try the business tool so easy to use, your team can’t help but collaborate. Learn more about our pricing and plans Free $ 0 USD Free for up to 10 collaborators per Workspace Capture your to-dos, get organized, and get sh*t done. Standard $ 5 USD Per user/month if billed annually ($6 billed monthly) Get more stuff done with unlimited boards, card mirroring, and more automation. Learn more about Standard Premium $ 10 USD Per user/month if billed annually ($12.50 billed monthly) Add AI to your boards and admin controls to your toolkit. Plus, get more perspective with views. Learn more about Premium Enterprise $ 17.50 USD Per user/month - billed annually ($210.00 annual price per user) Add enterprise-grade security and controls. This plan includes Atlassian Guard Standard and 24/7 Enterprise Admin support. Est. cost for users Learn more about Enterprise Well, what are you waiting for? Head over to your boards and start getting things done! Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://git-scm.com/book/fa/v2/%d9%85%d9%82%d8%af%d9%85%d8%a7%d8%aa-%da%af%db%8c%d8%aa-git-basics-chapter-%da%a9%d8%a7%d8%b1-%da%a9%d8%b1%d8%af%d9%86-%d8%a8%d8%a7-%d8%b1%db%8c%d9%85%d9%88%d8%aa-%d9%87%d8%a7-Working-with-Remotes
Git - کار کردن با ریموت ها (Working with Remotes) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. شروع به کار (getting started) 1.1 درباره ورژن کنترل (About Version Control) 1.2 تاریخچه کوتاهی از گیت (A Short History of Git) 1.3 گیت چیست؟ (What is Git) 1.4 نصب گیت (Installing Git) 1.5 ستاپ اولیه گیت (First-Time Git Setup) 1.6 دریافت کمک (Getting Help) 1.7 خلاصه (summary) 2. مقدمات گیت (git basics chapter) 2.1 گرفتن یک مخزن گیت (Getting a Git Repository) 2.2 ثبت تغییرات در مخزن (Recording Changes to the Repository) 2.3 مشاهده تاریخچه کامیت‌ها (Viewing the Commit History) 2.4 بازگرداندن تغییرات (Undoing Things) 2.5 کار کردن با ریموت ها (Working with Remotes) 2.6 تگ کردن (Tagging) 2.7 نام مستعار گیت (Git Aliases) 2.8 خلاصه (summary) 3. انشعاب‌گیری در گیت (Git Branching) 3.1 شاخه‌ها در یک نگاه (Branches in a Nutshell) 3.2 شاخه‌بندی و ادغام پایه‌ای (Basic Branching and Merging) 3.3 مدیریت شاخه‌ها (Branch Management) 3.4 روندهای کاری شاخه‌ها (Branching Workflows) 3.5 شاخه‌های راه دور (Remote Branches) 3.6 بازپایه‌گذاری (Rebasing) 3.7 خلاصه (Summary) 4. گیت روی سرور (Git on the server) 4.1 پروتکل‌ها (The Protocols) 4.2 راه‌اندازی گیت روی یک سرور (Getting Git on a Server) 4.3 ایجاد کلید عمومی SSH شما (Generating Your SSH Public Key) 4.4 نصب و راه‌اندازی سرور (Setting up server) 4.5 سرویس‌دهنده گیت (Git Daemon) 4.6 HTTP هوشمند (Smart HTTP) 4.7 گیت‌وب (GitWeb) 4.8 گیت‌لب (GitLab) 4.9 گزینه‌های میزبانی شخص ثالث (Third Party Hosted Options) 4.10 خلاصه (Summary) 5. گیت توزیع‌شده (Distributed git) 5.1 جریان‌های کاری توزیع‌شده (Distributed Workflows) 5.2 مشارکت در یک پروژه (Contributing to a Project) 5.3 نگهداری یک پروژه (Maintaining a Project) 5.4 خلاصه (Summary) 6. گیت هاب (GitHub) 6.1 راه‌اندازی و پیکربندی حساب کاربری (Account Setup and Configuration) 6.2 مشارکت در یک پروژه (Contributing to a Project) 6.3 نگهداری یک پروژه (Maintaining a Project) 6.4 مدیریت یک سازمان (Managing an organization) 6.5 اسکریپتنویسی در گیتهاب (Scripting GitHub) 6.6 خلاصه (Summary) 7. ابزارهای گیت (Git Tools) 7.1 انتخاب بازبینی (Revision Selection) 7.2 مرحله‌بندی تعاملی (Interactive Staging) 7.3 ذخیره موقت و پاک‌سازی (Stashing and Cleaning) 7.4 امضای کارهای شما (Signing Your Work) 7.5 جستجو (Searching) 7.6 بازنویسی تاریخچه (Rewriting History) 7.7 بازنشانی به زبان ساده (Reset Demystified) 7.8 ادغام پیشرفته (Advanced Merging) 7.9 بازاستفاده خودکار از حل تضادها (Rerere) 7.10 اشکال‌زدایی با گیت (Debugging with Git) 7.11 سابماژول ها (Submodules) 7.12 بسته‌بندی (Bundling) 7.13 جایگزینی (Replace) 7.14 ذخیره‌سازی اطلاعات ورود (Credential Storage) 7.15 خلاصه (Summary) 8. سفارشی‌سازی Git (Customizing Git) 8.1 پیکربندی گیت (Git Configuration) 8.2 ویژگی‌های گیت (Git Attributes) 8.3 هوک‌های گیت (Git Hooks) 8.4 یک نمونه سیاست اعمال شده توسط گیت (An Example Git-Enforced Policy) 8.5 خلاصه (Summary) 9. گیت و سیستم‌های دیگر (Git and Other Systems) 9.1 گیت به‌عنوان کلاینت (Git as a Client) 9.2 مهاجرت به گیت (Migrating to Git) 9.3 خلاصه (Summary) 10. مباحث درونی گیت (Git Internals) 10.1 ابزارها و دستورات سطح پایین (Plumbing and Porcelain) 10.2 اشیا گیت (Git Objects) 10.3 مراجع گیت (Git References) 10.4 فایل‌های بسته (Packfiles) 10.5 نگاشت (The Refspec) 10.6 پروتکل‌های انتقال (Transfer Protocols) 10.7 نگهداری و بازیابی داده‌ها (Maintenance and Data Recovery) 10.8 متغیرهای محیطی (Environment Variables) 10.9 خلاصه (Summary) A1. پیوست A: گیت در محیط‌های دیگر (Git in Other Environments) A1.1 رابط های گرافیکی (Graphical Interfaces) A1.2 گیت در ویژوال استودیو (Git in Visual Studio) A1.3 گیت در Visual Studio Code (Git in Visual Studio Code) A1.4 گیت در IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine (Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine) A1.5 گیت در Sublime Text (Git in Sublime Text) A1.6 گیت در بش (Git in Bash) A1.7 گیت در Zsh (Git in Zsh) A1.8 گیت در PowerShell (Git in PowerShell) A1.9 خلاصه (Summary) A2. پیوست B: گنجاندن گیت در برنامه‌های شما (Embedding Git in your Applications) A2.1 خط فرمان گیت (Command-line Git) A2.2 کتابخانهٔ گیت به زبان سی (Libgit2) A2.3 کتابخانه گیت برای زبان جاوا (JGit) A2.4 کتابخانه گیت برای زبان گو (go-git) A2.5 کتابخانه گیت پایتون (Dulwich) A3. پیوست C: دستورات گیت (Git Commands) A3.1 تنظیم و پیکربندی (Setup and Config) A3.2 گرفتن و ایجاد پروژه‌ها (Getting and Creating Projects) A3.3 نمونه‌برداری پایه‌ای (Basic Snapshotting) A3.4 انشعاب‌گیری و ادغام (Branching and Merging) A3.5 به‌اشتراک‌گذاری و به‌روزرسانی پروژه‌ها (Sharing and Updating Projects) A3.6 بازرسی و مقایسه (Inspection and Comparison) A3.7 عیب‌یابی (Debugging) A3.8 اعمال تغییرات به صورت پچ (Patching) A3.9 ایمیل (Email) A3.10 سیستم‌های خارجی (External Systems) A3.11 مدیریت (Administration) A3.12 دستورات سطح پایین گیت (Plumbing Commands) 2nd Edition 2.5 مقدمات گیت (git basics chapter) - کار کردن با ریموت ها (Working with Remotes) کار کردن با ریموت ها (Working with Remotes) برای اینکه بتوانید در هر پروژه گیت همکاری کنید، باید بدانید که چگونه مخازن از راه دور خود را مدیریت کنید. مخازن از راه دور نسخه هایی از پروژه شما هستند که در اینترنت یا شبکه جایی میزبانی می شوند. شما می توانید چند تا از آنها را داشته باشید، که هر یک از آنها به طور کلی فقط برای خواندن یا خواندن / نوشتن برای شما است. همکاری با دیگران شامل مدیریت این مخازن از راه دور و فشار دادن و کشیدن داده ها به آنها و از آنها زمانی که شما نیاز به به اشتراک گذاشتن کار. مدیریت مخازن از راه دور شامل دانستن چگونگی اضافه کردن مخازن از راه دور، حذف کنترل های از راه دور که دیگر معتبر نیستند، مدیریت شاخه های از راه دور مختلف و تعریف آنها به عنوان ردیابی یا نه و بیشتر است. در این بخش، ما برخی از این مهارت های مدیریت از راه دور را پوشش خواهیم داد. یادداشت Remote repositories can be on your local machine. کاملا ممکن است که شما با یک مخزن " `remote ` " کار کنید که در واقع در همان میزبان شما است. کلمه “remote” لزوماً به این معنی نیست که مخزن در جایی دیگر در شبکه یا اینترنت است ، فقط این است که در جای دیگری است. کار با چنین مخزن از راه دور هنوز هم شامل تمام استاندارد فشار، کشیدن و گرفتن عملیات به عنوان با هر از راه دور دیگر. نمایش ریموت ها (Showing Your Remotes) برای دیدن اینکه کدام سرورهای از راه دور را پیکربندی کرده اید، می توانید دستور git remote را اجرا کنید. نام های کوتاه هر دستگیره ای که مشخص کرده اید را لیست می کند. اگر مخزن خود را کلان کرده اید، حداقل باید origin را ببینید — این نام پیش فرضی است که Git به سرور که از آن کلان کرده اید می دهد: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin شما همچنین می توانید -v را مشخص کنید، که به شما URL هایی را نشان می دهد که Git برای نام کوتاه ذخیره کرده است که هنگام خواندن و نوشتن به آن از راه دور استفاده می شود: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) اگر شما بیش از یک ریموت دارید، دستور همه آنها را لیست می کند. به عنوان مثال، یک مخزن با چندین کنترل از راه دور برای کار با چندین همکار ممکن است چیزی شبیه به این باشد. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) این به این معنی است که ما می توانیم به راحتی از هر یک از این کاربران کمک بگیریم. ما ممکن است علاوه بر این اجازه داشته باشیم که به یکی یا چند مورد از این ها فشار بیاریم، اگرچه نمی توانیم این را اینجا بگوییم. توجه داشته باشید که این کنترل های از راه دور از پروتکل های مختلفی استفاده می کنند؛ ما بیشتر در مورد این موضوع در راه‌اندازی گیت روی یک سرور (Getting Git on a Server) صحبت خواهیم کرد. افزودن مخازن راه‌دور (Adding Remote Repositories) ما اشاره کرده ایم و برخی از تظاهرات را نشان داده ایم که چگونه دستور `git clone ` به طور ضمنی ریموت `origin ` را برای شما اضافه می کند. در اینجا نحوه اضافه کردن یک ریموت جدید به طور صریح است. برای اضافه کردن یک مخزن Git از راه دور به عنوان یک نام کوتاه که به راحتی می توانید به آن مراجعه کنید، `git remote add <shortname> <url> ` را اجرا کنید: $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) حالا شما می توانید از رشته pb در خط فرمان به جای کل URL استفاده کنید. به عنوان مثال، اگر می خواهید تمام اطلاعاتی را که Paul دارد اما هنوز در مخزن خود ندارید، بدست آورید، می توانید git fetch pb را اجرا کنید: $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit شعبه master پاول در حال حاضر به صورت محلی به عنوان pb/master قابل دسترسی است — شما می توانید آن را به یکی از شعبه های خود ادغام کنید، یا اگر می خواهید آن را بازرسی کنید می توانید یک شعبه محلی را در آن نقطه بررسی کنید. ما در انشعاب‌گیری در گیت (Git Branching) درباره شاخه ها و نحوه استفاده از آنها به طور مفصل تر صحبت خواهیم کرد. فچ و پول کردن از ریموت های شما (Fetching and Pulling from Your Remotes) همانطور که دیدید، برای دریافت داده ها از پروژه های از راه دور، می توانید دستورات را اجرا کنید. $ git fetch <remote> این فرمان به سمت آن پروژه ی از راه دور می رود و تمام داده هایی را که شما هنوز ندارید از آن پروژه ی از راه دور می گیرد. بعد از انجام این کار، شما باید مرجع تمام شاخه ها را از آن ریموت داشته باشید، که می توانید در هر زمان ادغام یا بازرسی کنید. اگر یک مخزن را کلان کنید، دستور به طور خودکار آن مخزن از راه دور را با نام “origin” اضافه می کند. بنابراین، git fetch origin هر کار جدیدی را که از زمانی که کلونش کردید (یا آخرین بار از آن گرفته شده اید) به آن سرور ارسال شده است، می آورد. مهم است که توجه داشته باشید که دستور git fetch فقط داده ها را به مخزن محلی شما دانلود می کند — آن را به طور خودکار با هر یک از کارهای شما ادغام نمی کند یا آنچه در حال حاضر روی آن کار می کنید را اصلاح نمی کند. وقتی آماده شدی باید دست به دست با کارت ادغامش کنی. اگر شاخه فعلی شما برای ردیابی یک شاخه از راه دور تنظیم شده است (برای اطلاعات بیشتر به بخش بعدی و انشعاب‌گیری در گیت (Git Branching) مراجعه کنید) ، می توانید از دستور git pull برای گرفتن خودکار و سپس ادغام آن شاخه از راه دور در شاخه فعلی خود استفاده کنید. این ممکن است یک جریان کار ساده تر یا راحت تر برای شما باشد؛ و به طور پیش فرض، دستور git clone به طور خودکار شاخه محلی master شما را برای ردیابی شاخه master از راه دور (یا هر نام دیگری که شاخه پیش فرض نامیده می شود) در سرور که از آن کلون شده است، تنظیم می کند. اجرای git pull به طور کلی داده ها را از سرور که در اصل از آن کلان کرده اید می گیرد و به طور خودکار سعی می کند آن را در کد که در حال حاضر روی آن کار می کنید ادغام کند. یادداشت از نسخه 2.27 Git به بعد، اگر متغیر pull.rebase تنظیم نشده باشد، git pull هشدار می دهد. گیت به شما هشدار می دهد تا زمانی که متغیر را تنظیم کنید. اگر می خواهید رفتار پیش فرض گیت را داشته باشید (اگر امکان دارد سریع پیش بروید، در غیر این صورت یک commit ادغام ایجاد کنید): git config --global pull.rebase "false" اگه ميخواي موقع کشيدن بازي کني: git config --global pull.rebase "true" Pushing to Your Remotes (پوش کردن به ریموت های شما) وقتی پروژه تان را در نقطه ای قرار می دهید که می خواهید به اشتراک بگذارید، باید آن را به سمت بالا بکشید. دستور برای این کار ساده است: git push <remote> <branch> `. اگر می خواهید شاخه `master خود را به سرور origin خود فشار دهید (باز هم، کلون کردن به طور معمول هر دو نام را برای شما به طور خودکار تنظیم می کند) ، پس می توانید این کار را انجام دهید تا هر کاری که انجام داده اید را به سرور برگردانید: $ git push origin master این دستور فقط در صورتی کار می کند که شما از یک سرور که به آن دسترسی نوشتن دارید و اگر هیچ کس در این میان فشار نداده است، کلان کنید. اگر شما و شخص دیگری در همان زمان شبیه سازی کنید و آنها به سمت بالا فشار بیاورند و سپس شما به سمت بالا فشار بیاورید، فشار شما به درستی رد خواهد شد. بايد اول کار اونا رو بگيري و اونو به کار خودت اضافه کني تا بتوني فشار بياري برای اطلاعات دقیق تر در مورد نحوه ارسال به سرورهای از راه دور، به انشعاب‌گیری در گیت (Git Branching) مراجعه کنید. بررسی مخزن راه‌دور (Inspecting a Remote) اگر می خواهید اطلاعات بیشتری در مورد یک ریموت خاص مشاهده کنید، می توانید از دستور git remote show <remote> استفاده کنید. اگر شما این دستور را با یک نام کوتاه خاص اجرا کنید، مانند origin ، چیزی شبیه به این به دست می آورید: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) این لیست URL برای مخزن از راه دور و همچنین اطلاعات شاخه ردیابی را لیست می کند. دستور به شما کمک می کند که اگر شما در شاخه master هستید و git pull را اجرا کنید، آن را به طور خودکار شاخه master از راه دور به محلی پس از آن آورده شده است. همچنین تمام ارجاعات از راه دور را که برداشته است لیست می کند. این یک مثال ساده است که احتمالا با آن روبرو خواهید شد. با این حال، هنگامی که شما از گیت به شدت استفاده می کنید، ممکن است اطلاعات بیشتری از `git از راه دور نمایش ` مشاهده کنید: $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) این دستور نشان می دهد که چه شاخه ای به طور خودکار فشار داده می شود زمانی که شما git push را در حالی که در شاخه های خاص اجرا می کنید. همچنین به شما نشان می دهد که چه شاخه های از راه دور در سرور شما هنوز وجود ندارد، چه شاخه های از راه دور شما از سرور حذف شده است، و چندین شاخه محلی که قادر به ادغام به طور خودکار با شاخه های ردیابی از راه دور خود را هنگامی که شما اجرا git pull . تغییر نام یا حذف ریموت (Renaming and Removing Remotes) شما می توانید git remote rename را اجرا کنید تا نام کوتاه یک ریموت را تغییر دهید. به عنوان مثال، اگر می خواهید نام pb را به paul تغییر دهید، می توانید این کار را با git remote rename انجام دهید: $ git remote rename pb paul $ git remote origin paul لازم به ذکر است که این تغییر نام تمام شاخه های ردیابی از راه دور شما را نیز تغییر می دهد. آنچه که قبلاً در pb/master به آن اشاره می شد، اکنون در paul/master است. اگر می خواهید از راه دور برای برخی از دلایل حذف کنید - شما سرور را جابجا کرده اید یا دیگر از یک آینه خاص استفاده نمی کنید، یا شاید یک مشارکت کننده دیگر مشارکت نمی کند - شما می توانید از `git remote remove ` یا `git remote rm ` استفاده کنید: $ git remote remove paul $ git remote origin هنگامی که شما ارجاع به یک ریموت را از این طریق حذف می کنید، تمام شاخه های ردیابی از راه دور و تنظیمات پیکربندی مرتبط با آن ریموت نیز حذف می شوند. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/a10networks-a10-thunder-tps-ddos-defense-solutions/?trk=products_seo_search
A10 Defend - Intelligent & Automated DDoS Protection | LinkedIn Skip to main content LinkedIn A10 Networks, Inc in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in A10 Defend - Intelligent & Automated DDoS Protection DDoS Protection Software by A10 Networks, Inc See who's skilled in this Add as skill Learn more Report this product About A10 Defend provides a holistic DDoS protection solution that is scalable, economical, precise, and intelligent to help customers ensure optimal user and subscriber experiences. Media Products media viewer No more previous content A10 Defend Threat Control The A10 Defend suite, composed of A10 Defend Mitigator, A10 Defend Detector, A10 Defend Threat Control, and A10 Defend Orchestrator, provides a holistic solution that is scalable, economical, precise, and intelligent to help customers ensure optimal user and subscriber experiences. A10 Defend Suite Used by the top service providers and online gaming companies, the A10 Defend suite consists of several components. A10 Defend Detector efficiently identifies abnormal traffic, A10 Defend Mitigator (previously Thunder TPS) automatically and intelligently mitigates the identified inbound DDoS attack, A10 Defend Threat Control proactively provides standalone layered defense and actionable insights, and A10 Defend Orchestrator (previously aGalaxy) provides seamless DDoS defense execution. Demo: A10 Defend Threat Control A10 Defend Threat Control provides a robust first layer of DDoS defense. By leveraging proprietary ML/AL-enhanced data processing techniques, Threat Control proactively monitors attackers and understands key DDoS attack methods, with or without dedicated DDoS prevention solutions. No more next content Featured customers of A10 Defend - Intelligent & Automated DDoS Protection​ World Wide Technology IT Services and IT Consulting 806,684 followers Fastly Software Development 64,600 followers Imperium Dynamics IT Services and IT Consulting 18,464 followers Dell Technologies Computer Hardware Manufacturing 5,431,605 followers Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
http://www.trello.com/enterprise
Trello Enterprise | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Trello Enterprise Add enterprise-grade security and control. This plan includes Atlassian Guard Standard and 24/7 Enterprise Admin support. Contact Sales Trello is where productivity meets enterprise at scale From Product to Marketing to Support to… you get the idea. All organized in one central Admin Dashboard. Enhanced Productivity Focus on what matters Summarize scattered to-dos across your notes, emails, Slack, and Teams using the AI-powered Quick Capture feature to save them as Trello cards in your Trello Inbox. Better organize your saved cards from Inbox onto your boards and block out focus time on Planner to avoid distractions and get more work done. Integrate with the tools you care about Access hundreds of Power-Ups to integrate Trello Enterprise with your other must-have tools like Slack, Confluence, Google Drive, and more! Goodbye manual tasks, hello automation Enjoy Trello’s automation to automatically update cards, assign members, and help keep the team on-track with due dates. Admins Manage users across the org Leverage one central Admin Dashboard to manage permission settings, security, and so much more Membership & organization controls Manage your org’s access at the individual, board-specific, or company-wide levels. Ensure your teams can collaborate easily, while still keeping sensitive work private. Single sign-on made easy with Atlassian Guard Seamlessly set up SAML SSO for your teams with Atlassian Guard , available across all Atlassian cloud products. Powering productivity for millions of users worldwide, at companies like Top-tier integrations. Top-rated by users. Power your work with 200+ Power-Ups. Connect Trello to your chat tools, marketing tools, support tools, and anything else your business runs on. Explore our Power-Ups Enterprise-Grade Security You Can Trust Manage users and control permissions with just a few clicks. With 99.99% uptime and industry-leading compliance certifications, we’ll never leave you hanging. Learn more Curated content for businesses and teams Read to discover the latest in productivity hacks, leadership insight, and tips for successful remote work. Explore the blog Pricing Calculator How many users? Your price: $ 17.50 USD Per user per month $ 10,500 USD Total per year* *$10k minimum spend threshold when purchasing Trello Enterprise. Need help? Contact us. Contact sales Ready to learn more? Get in touch with one of our sales representatives to discuss Trello Enterprise options that might be right for your business needs. Trello Enterprise is for businesses with 50 seats or more, so if you are a smaller team you may be a fit for Trello Premium! Check out Premium Customer support Already a customer? For specific account or billing questions, security concerns, or answers not available on our help site, we’re happy to help. Contact customer support Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://huggingface.co/datasets/m-a-p/COIG-CQIA
m-a-p/COIG-CQIA · Datasets at Hugging Face Hugging Face Models Datasets Spaces Community Docs Enterprise Pricing Log In Sign Up Datasets: m-a-p / COIG-CQIA like 691 Follow Multimodal Art Projection 976 Tasks: Question Answering Text Classification Text Generation Modalities: Text Formats: json Languages: Chinese Size: 10K - 100K ArXiv: arxiv: 2403.18058 arxiv: 2304.07987 arxiv: 2307.09705 Libraries: Datasets Dask Croissant + 1 Dataset card Data Studio Files Files and versions xet Community 31 Dataset Viewer Auto-converted to Parquet API Embed Duplicate Data Studio Subset (13) chinese_traditional · 1.11k rows chinese_traditional (1.11k rows) coig_pc (3k rows) douban (3.09k rows) exam (4.86k rows) finance (11.3k rows) human_value (1.01k rows) logi_qa (421 rows) ruozhiba (240 rows) segmentfault (458 rows) wiki (10.6k rows) wikihow (1.49k rows) xhs (1.51k rows) zhihu (5.63k rows) Split (1) train · 1.11k rows train (1.11k rows) SQL Console instruction string lengths 9 438 input string classes 1 value output string lengths 13 578 task_type dict domain sequence lengths 2 2 metadata string classes 1 value answer_from string classes 1 value human_verified bool 1 class copyright string classes 1 value 暗度陈仓 成语释义: 指正面迷惑敌人,而从侧翼进行突然袭击。亦比喻暗中进行活动。陈仓,古县名,在今陕西省宝鸡市东,为通向汉中的交通孔道。 成语出处:《史记·高祖本纪》……项王使卒三万人从,楚与诸侯之慕从者数万人,从杜南入蚀中。去辄烧绝栈道,以备诸侯盗兵袭之,亦示项羽无东意……八月,汉王用韩信之计,从故道还,袭雍王章邯—迎击汉陈仓,雍兵败,… { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 家有敝帚,享之千金 这句话的意思是什么? 敝帚破扫帚;享供奉。自家的破扫帚被认为价值千金。比喻自己的东西即使不好也倍觉珍贵。有时用于自谦。 成语出处:汉·班固《东观汉记·光武帝纪》帝闻之,下诏让吴汉副将刘禹曰‘城降,婴儿老母,口以万数,一旦放兵纵火,闻之可谓酸鼻。家有敝帚,享之千金。禹宗室子孙,故尝更职,何忍行此!’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?别风淮雨 答案: 这是列风淫雨的误写,因别和列、淮与淫字殂相似。后称书籍中因错别字而以讹传讹为别风淮雨。 成语出处:南朝梁·刘勰《文心雕龙·练字》《尚书大传》有‘别风淮雨’,《帝王世纪》云‘列风淫雨’。‘别’‘列’、‘淮’‘淫’字似潜移。‘淫’‘列’义当而不奇,‘淮’‘别’理乖而新异。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?画龙点睛 答案: 原形容梁代画家张僧繇作画的神妙。后多比喻写文章或讲话时,在关键处用几句话点明实质,使内容生动有力。 成语出处:唐·张彦远《历代名画记·张僧繇》金陵安乐寺四白龙不点眼睛,每云‘点睛即飞去。’人以为妄诞,固请点之。须臾,雷电破壁,两龙乘云腾去上天,二龙未点眼者见在。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:独有千古 具有流传久远的价值;具有独特的专长或优点。同独有千秋。 成语出处:清·翁方纲《石洲诗话》卷七唐之李义山(李商隐)、宋之黄涪翁(黄庭坚),皆杜法也。先生撮在此一首中,真得其精微矣。方翁(陆游)、道园(虞集)皆未尝有此等议论,即使不读遗山诗集,已自可以独有千古矣。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 螳螂捕蝉,黄雀在后是什么意思? 螳螂正要捉蝉,不知黄雀在它后面正要吃它。比喻目光短浅,只想到算计别人,没想到别人在算计他。 成语出处:《庄子·山木》睹一蝉,方得美荫而忘其身,螳螂执翳而搏之,见得而忘其形;异鹊从而利之,见利而忘其真。汉·韩婴《韩诗外传》螳螂方欲食蝉,而不知黄雀在后,举其颈欲啄而食之也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 鹅湖之会 这个成语的意思是什么? 比喻具有开创性的辩论会。 成语出处:南宋淳熙二年(1175年)在信州(今江西上饶)鹅湖寺举行的一次著名的哲学辩论会。由吕祖谦邀集,意图调和朱熹和陆九渊两派争执。实质上是朱的客观唯心主义和陆的主观唯心主义的一场争论。它是中国哲学史上一次堪称典范的学术讨论会,首开书院会讲之先河。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 望风而靡的含义是什么? ①见对方的威势就服服帖帖。形容畏惧之状。②望见对方就为之折服倾倒。形容钦敬之状。③形容军无斗志。同望风披靡。 成语出处:《汉书·杜周传》天下莫不望风而靡,自尚书近臣皆结舌杜口,骨肉亲属莫不股栗。唐·陈子昂《堂弟孜墓志铭》是以乡里长幼,望风而靡;邦国贤豪,闻名而悦服。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:纸上谈兵 在纸面上谈论打仗。比喻空谈理论,不能解决实际问题。也比喻空谈不能成为现实。 成语出处:《史记·廉颇蔺相如列传》记载战国时赵国名将赵奢之子赵括,年轻时学兵法,谈起兵事来父亲也难不倒他。后来他接替廉颇为赵将,在长平之战中。只知道根据兵书办,不知道变通,结果被秦军大败。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 塞翁失马的意思是什么? 塞边界险要之处;翁老头。比喻一时虽然受到损失,也许反而因此能得到好处。也指坏事在一定条件下可变为好事。 成语出处:《淮南子·人间训》近塞上之人有善术者,马无故亡而入胡。人皆吊之。其父曰‘此何遽不为铬?’居数月,其马将胡骏马而归。人皆贺之。…故福之为祸,祸之为福,化不可极,深不可测也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 盲翁扪籥 这个成语的意思是什么? 比喻只凭片面了解或局部经验就对事物妄加判断。 成语出处:宋·苏轼《日喻》生而眇者不识日,问之有目者。或告之曰‘日之状如铜槃。’扣槃而得其声。他日闻锺以为日也。或告之曰‘日之光如烛。’扪烛而得其形,他日揣籥以为日也。日之与锺籥亦远矣,而眇者不知其异,以其未尝见而求之人也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:不以为意 不把它放在心上。表示对人、对事抱轻视态度。 成语出处:《三国志·吴书·陆凯传》定大恨凯,思中伤之,凯终不以为意,乃心公家,义形于色,表疏皆指事不饰,忠恳内发。北魏·杨衒之《洛阳伽蓝记·秦太上君寺》临淄官徒有在京邑,闻怀砖慕势,咸共耻之,唯崔孝忠一人不以为意。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 杳如黄鹤的意思是什么? 杳无影无声;黄鹤传说中仙人所乘的鹤。原指传说中仙人骑着黄鹤飞去,从此不再回来。现比喻无影无踪或下落不明。 成语出处:南朝梁·任昿《述异记》荀瓌憩江夏黄鹤楼上,望西南有物飘然降自云汉,乃驾鹤之宾也。宾主欢对辞去,跨鹤腾空,眇然烟灭。唐·崔灏《黄鹤楼》诗黄鹤一去不复返,白云千载空悠悠。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 众盲摸象 这个词语是什么意思? 许多瞎子摸象,摸到象腿的说象一根柱子,摸到象身子的说象一堵墙,摸到象尾的说象一条蛇,互相争论不休。比喻看问题以偏概全。 成语出处:《大般涅槃经》三二其触牙者即言象形如芦菔根,其触耳者言象如箕,其触头者言象如石,其触鼻者言象如杵,其触脚者言象如木臼,其触脊者言象如床,其触腹者言象如畒,其触尾者言象如绳。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 禁中颇牧 这句话的意思是什么? 比喻宫廷侍从官中文才武略兼备者。 成语出处:廉颇、李牧皆战国时赵国守边御敌之良将。唐宣宗时,党项扰河西,翰林学士毕謕上破羌条陈甚悉,帝大悦,曰吾将择能帅者,孰谓颇牧在吾禁署,卿为朕行乎!于是拜謕为邠宁节度使、河西供军安抚使。謕于任内多所建树。事见《新唐书·毕謕传》。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:丰姿绰约 丰姿风姿,风度姿态。绰约柔美的样子。形容女子体态柔美,神采飘逸。同丰神绰约。 成语出处:清·曾朴《孽海花》第七回 一回头时,却见那轿子里坐着个十四五岁的不长不短、不肥不瘦的女郎,面如瓜子,脸若桃花,两条欲蹙不蹙的蛾眉,一双似开非开的凤眼,似曾相识,莫道无情,正是说不尽的体态风流,丰姿绰约。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 摇尾涂中的含义是什么? 比喻自由自在地生活。曳尾,犹摇尾。 成语出处:战国时,楚王派大夫去请庄子做官,庄子对楚大夫说听说楚国有神龟,死去已三千年,现在把它用匣子装起来藏在庙堂之上。你看此龟是留下骨头让人珍藏好呢,还是活着曳尾于泥涂中好?大夫说当然是活着曳尾于涂中好。庄子说那末我将曳尾于涂中。见 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?礼先壹饭 答案: 指在礼节上自己年岁稍长。壹饭,犹言一顿饭,喻指极短的时间。也指在礼节上先有恩惠与人。 成语出处:《国语·越语上》句践对曰‘昔天以越予吴,而吴不受命,今天以吴予越,越可以无听天之命而听君之令乎?吾请达王甬句东,吾与君为二君乎!’夫差对曰‘寡人礼先壹饭矣。’亦作礼先一饭。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 东观续史是什么意思? 东观汉代官家藏书的地方。原指汉代女史学家班昭奉诏就东观续成其兄班固没有完成的《汉书》。后用以指女子才学高深。 成语出处:《后汉书·曹世叔妻传》扶风曹世叔妻者,同郡班彪之女也,名昭,字惠班,……兄固著《汉书》,其八表及天文志未及竟而卒,和帝诏昭,就东观藏书阁踵而成之。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 文君司马 这个词语是什么意思? 后指相爱的情人或夫妻。 成语出处:《史记·司马相如列传》载汉代辞赋家司马相如,在宴席上以琴音挑动临邛富商卓王孙寡居的女儿卓文君;文君夜奔相如,与之结为夫妇。卓王孙不认这门亲事,司马相如与卓文君开了个酒馆,文君当炉卖酒,相如穿着犊鼻裤与奴仆一道洗器皿,卓王孙感到耻 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?巾国英雄 答案: 巾国即巾帼,古代妇女配戴的头巾和发饰,后借指妇女。指女子中的英雄。亦作巾帼英雄。 成语出处:范克明《张学良传·郭松龄》原来,是由于有超人胆识,被誉为‘巾国英雄’的韩淑秀‘不顾风险,冒死拦截刑车,陈述郭松龄是他的未婚夫,归奉完婚,根本没有参加革命党的事,因此郭才未遭杀害’并应聘当了张作霖的教训队的教官。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:醉吐相茵 后以这一典故喻指宽以待人必然会有好的回报。 成语出处:典出《汉书·丙吉传》吉驭吏耆酒,数逋荡,尝从吉出, 醉欧丞相车上。西曹主吏白欲斥之,吉曰‘以醉饱之失去士,使此人将复何所容?西曹地忍之,此不过汙丞相车茵耳。’遂不去也。此驭吏为边郡人,熟悉边事,后来为防务工作提出切实有用的建议。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 梅妻鹤子 这个成语的意思是什么? 以梅为妻,以鹤为子。比喻清高或隐居。宋·林逋隐居西湖孤山,植梅养鹤,终生不娶,人谓梅妻鹤子。 成语出处:宋·沈括《梦溪笔谈·人事二》林逋隐居杭州孤山,常畜两鹤,纵之则飞入云霄,盘旋久之,復入笼中。逋常泛小艇,游西湖诸寺。有客至逋所居,则一童子出应门,延客坐,为开笼纵鹤。良久,逋必棹小船而归。盖尝以鹤飞为验也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:白衣公卿 古时指进士。唐代人极看重进士,宰相多由进士出身,故推重进士为白衣卿相,是说虽是白衣之士,但享有卿相的资望。 成语出处:五代·王定保《唐摭言·卷一·散序进士》缙绅虽位极人臣,不由进士者,终不为美,以至岁贡常不减八九百人,其推重谓之‘白衣公卿’,又曰‘一品白衫’。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 一人善射,百夫决拾 成语释义: 古谚语,意思是为将者善战,其士卒亦必勇敢无前。亦比喻凡事为首者倡导于前,则其众必起而效之。 成语出处:《国语·吴语》夫申胥、华登简服吴国志士于甲兵,而未尝有所挫也。夫一人善射,百夫决拾,胜未可成也。韦昭注决,勾弦。拾,捍。言申胥、华登善用兵众必化之;犹一人善射,而百夫竞著决拾而效之。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 捉襟露肘 成语释义: 指整一整衣襟就露出肘子。形容衣衫褴褛。引申为顾此失彼,处境困难。亦形容书法生动而有气势。同捉衿见肘。 成语出处:清·薛雪《一瓢诗话》分题拈韵,诗家之厄也……一遭窍,未免捉襟露肘。清·西周生《醒世姻缘传》第七十一回童七道‘咱实得百十两银接接手才好哩;要不,也就捉襟露肘了。’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 狗血淋头 这个成语的意思是什么? 旧时迷信说法,谓狗血淋在妖人头上,就可使其妖法失灵。后形容骂得很凶,使被骂者如淋了狗血的妖人一样,无言以对,无计可施。 成语出处:明·施耐庵《水浒全传》第五十三回马知府道‘必然是个妖人!’教去取些法物来。牢子、节级将李逵捆翻,驱下厅前草地里,一个虞候掇一盆狗血没头一淋。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 日程月课的含义是什么? 每日每月按一定的程序课试。形容因循守旧,无所创新。 成语出处:清·莫友芝《〈巢经堂诗钞〉序》[子尹]不肯以诗人自居。当其兴到,顷刻千言;无所感触,或经时不作一字,又脱稿不自收拾,子弟钞存十之三四而已。而其盘盘之气,熊熊之光,浏漓顿挫,不主故常,以视近世日程月课,楦酿篇牍,自张风雅者,其贵贱何如也? { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:尺二秀才 旧时用以讥讽写俗字的书生。尺二即指当时眒字的俗体尽字。 成语出处:宋·孙奕《履斋示儿编·文说·声画押韵贵乎审》诚斋先生杨公考玄南漕试,……先生见卷子上书‘眒’字作‘尽’,必欲摈斥。考官乃上庠人,力争不可。先生云‘明日揭榜,有喧传以为场屋取得个尺二秀才,则吾辈将胡颜?’竟黜之。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 酒酸不售的含义是什么? 酒已经变酸了,依然卖不出去。原比喻奸臣阻拦了有学问、有贤德的人为国家效力,使国君受到蒙蔽。后比喻经营无方或办事用人不当。 成语出处:《韩非子·外储说右上》宋人有酤酒者,……著然不售,酒酸,怪其故,问其所知,问长者杨倩,……曰‘狗猛则酒何故而不售?’曰‘人畏焉。……,而狗迎龁之,此酒所以酸而不售也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 保残守缺 这句话的意思是什么? ①指汉代今文经学派儒生墨守残缺不全的今文经典而拘执一家之言。后常用以比喻泥古守旧,不知改进。②指保藏残缺的古籍文献。 成语出处:《汉书·刘歆传》信口说而背传记,是末师而非往古……犹欲保残守缺,挟恐见破之私意,而无从善服义之公心,或怀妒嫉,不考情实,雷同相从,随声是非,抑此三学。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?飘茵落溷 答案: 随风而落,有的飘在茵席上,有的落在粪坑里。比喻由于偶然的机缘而有富贵贫贱的不同命运。也指女子堕落风尘。 成语出处:《梁书·儒林传·范缜》人之生譬如一树花,同发一枝,俱开一蒂,随风而堕,自有拂帘幌坠于茵席之上,自有关篱墙落于粪溷之侧。坠茵席者,殿下是也;落粪溷者,下官是也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?白衣卿相 答案: 古时指进士。唐代人极看重进士,宰相多由进士出身,故推重进士为白衣卿相,是说虽是白衣之士,但享有卿相的资望。 成语出处:五代·王定保《唐摭言·卷一·散序进士》缙绅虽位极人臣,不由进士者,终不为美,以至岁贡常不减八九百人,其推重谓之‘白衣公卿’,又曰‘一品白衫’。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 伏龙凤雏 这句话的意思是什么? 伏龙(卧龙)诸葛孔明。凤雏庞士元。两人都是汉末三国时期著名的谋略家,军事家。后指隐而未现的有较高学问和能耐的人。 成语出处:《三国志·蜀志·诸葛亮传》裴松之注引《襄阳记》刘备访世事于司马德操。德操曰‘儒生俗士,岂识时务?识时务者在乎俊杰。此间自有伏龙、凤雏。’备问为谁,曰‘诸葛孔明、庞士元也。’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这个词语是什么意思?日渐月染 答案: 濡浸润。染熏染。天长日久地渐渐熏染。指受外界事物的影响而发生逐渐的变化。同日濡月染。 成语出处:宋·程珌《丙子轮对札子》招之得其地矣,又当各分其屯,无杂官军,盖一与之染,则日渐月染,尽成弃甲之人,不幸有警,则彼此相持,莫肯先进;一有微功,则彼此交集,反戈自戕,岂暇向敌哉。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 班荆道旧 这个成语的意思是什么? 指朋友相遇于途,铺荆坐地,共叙情怀。后泛指朋友相遇,共叙离情。亦作班荆道故。 成语出处:典出《左传·襄公二十六年》楚伍参与蔡太师子朝友,其子伍举与声子相善……伍举奔郑,将遂奔晋。声子将如晋,遇之於郑郊,班荆相与食,而言复故。杜预注班,布也。布荆坐地,共议归楚,事朋友世亲。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:画荻和丸 含义: 用以称赞母亲教子有方。同画荻丸熊。 成语出处:宋·欧阳修幼时,母郑氏以荻画地教子读书。唐·柳仲郢幼嗜学,母韩氏用熊胆和制丸子,使郢夜咀咽以提神醒脑。郭沫若《虎符》附录《写作缘起》但要写母爱,在儿女小的时候容易表现,如推干就湿、画荻和丸之类,都是儿女小时的事。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 退食从容 这个词语是什么意思? 指官吏品行节俭正直,仪容从容自得,可为楷模。 成语出处:语出《诗·召南·羔羊》退食自公,委蛇委蛇。郑玄笺退食,谓减膳也。自,从也。从于公,谓正直顺于事也。委蛇,委曲自得之貌。节俭而顺心志定,故可自得也。朱熹集传南国化文王之政,在位皆节俭正直,故诗人美其衣服有常,而从容自 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:入室操戈 操拿;戈古代象矛的武器。到他的屋里去,拿起他的武器攻击他。比喻引用对方的论点反驳对方。 成语出处:《后汉书·郑玄传》任城何休好《公羊》学,遂著《公羊墨守》、左氏膏肓》、《谷梁废疾》。玄乃‘发《墨守》、针《膏肓》、起《废疾》’。休见而叹曰‘康成入吾室,操吾戈以伐我乎?’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 楚囚相对是什么意思? 形容人们遭遇国难或其他变故,相对无策,徒然悲伤。 成语出处:南朝·宋刘·义庆《世说新语·言语》过江诸人,每至美日,辄相邀新亭,藉卉饮宴。周侯(周顗)中坐而叹曰‘风景不殊,正自有山河之异!’皆相视流泪。唯王丞相(王导)愀然变色曰‘当共戮力王室,克复神州,何至作楚囚相对!’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:八方呼应 含义: 呼应彼此声气相通。泛指周围、各地。形容各方面互通声气,互相配合。 成语出处:余秋雨《寂寞天柱山》三我认为,天柱山之所以能给古人一种居家感,一个比较现实的原因是它地处江淮平原,四相钩连,八方呼应,水陆交通畅达,虽幽深而无登高之苦,虽奇丽而无柴米之匮,总而言之,既宁静又方便。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:青史标名 青史史书。标记下。在史书上记下姓名。比喻在历史上留下好的名声。亦作青史留名、青史传名、青史名留、青史流芳。 成语出处:明·方汝浩《禅真逸史》第三十六回众将军年虽弱冠,各负雄才,文武兼通,正堪为朝廷之股肱,庙廊之梁栋。今能顺天知命,解甲而降,准拟青史标名,流芳千古。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 意兴索然 成语释义: 索然全无,空尽。兴致全无。形容一点兴致也没有。亦作兴致索然。 成语出处:《东周列国志》第七十一回景公意兴索然。左右问曰‘将回宫乎?’景公曰‘可移於梁邱大夫之家。’清·蒲松龄《聊斋志异·仙人岛》绿云顾父曰圣人无字门人者,‘孝哉……’一句,即是人言。王闻之,意兴索然。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 天女散花的意思是什么? 原为佛教故事天女散花以试菩萨和声闻弟子的道行,花至菩萨身上即落去,至弟子身上便不落。后多形容抛洒东西或大雪纷飞的样子。 成语出处:《维摩经·观众生品》时维摩诘室有一天女,见诸大人闻所说说法,便现其身,即以天华散诸菩萨、大弟子上,华至诸菩萨即皆堕落,至大弟子便著不堕。一切弟子神力去华,不能令去。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 托骥之蝇 这个词语是什么意思? 喻指追随贤能之后而得以显名的人。 成语出处:《后汉书·隗器传》帝报以手书曰‘慕乐德义,思相结纳……数蒙伯乐一顾之价,而苍蝇之飞,不过数步,即托骥尾,得以绝群。’李贤注张敞书曰‘苍蝇之飞,不过十步,自托骐骥之尾,乃腾千里之路。然无损于骐骥,得使苍蝇绝群也,’见敞传。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:今日有酒今日醉 比喻过一天算一天。也形容人只顾眼前,没有长远打算。同今朝有酒今朝醉。 成语出处:茅盾《狂欢的解剖》他们这种‘自信’,这种‘有前途’的自觉,就使得他们的要求快乐跟罗马帝国衰落时代的有钱人的纵乐完全不同,那时罗马的有钱人感得大难将到而又无可挽救,于是‘今日有酒今日醉’了。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:离经畔道 含义: ①指思想、言行背离儒家经典和规范。②指背离占统治地位的思想和行为规范。 成语出处:元·费唐臣《贬黄州》第一折且本官志大言浮,离经畔道,见新法之行,往往行诸吟咏。清·李百川《绿野仙踪》第七回子真不待教而诛之人也!吾房中师弟授受,绍闻知之统,继精一之传,岂可以容离经畔道之人哉! { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 曲突徙薪 这句话的意思是什么? 曲弯;突烟囱;徒迁移;薪柴草。把烟囱改建成弯的,把灶旁的柴草搬走。比喻事先采取措施,才能防止灾祸。 成语出处:《汉书·霍光传》臣闻客有过主人者,见其灶直突,傍有积薪。客谓主人,更为曲突,远徙其薪,不者且有火患,主人嘿然不应。俄而家果失火,邻里共救之,幸而得息。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 金字招牌 这句话的意思是什么? 旧时店铺为显示资金雄厚而用金箔贴字的招牌。现比喻高人一等可以炫耀的名义或称号。也比喻名誉好。 成语出处:清·曾朴《孽海花》第二十五回珏斋郻只出使了一次朝鲜,办结了甲申金玉均一案,又曾同威毅伯和日本伊滕博文定了出兵朝鲜彼此知会的条约,总算一帆风顺,文武全才的金字招牌,还高高挂着。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:荒淫无道 荒淫淫乱无度,贪恋酒色。无道不讲或不行道义。多指君主生活糜烂,重用奸佞,残害忠良,奴役百姓。 成语出处:明·罗贯中《三国演义》第一百九回今主上荒淫无道,亵近娼优,听信谗言,闭塞贤路其罪甚于汉之昌邑,不能主天下。吾谨按伊尹、霍光之法,别立新君,以保社稷,以安天下,如何? { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 家有弊帚,享之千金的含义是什么? 弊帚破扫帚;享供奉。自家的破扫帚被认为价值千金。比喻自己的东西即使不好也倍觉珍贵。有时用于自谦。 成语出处:汉·班固《东观汉记·光武帝纪》帝闻之,下诏让吴汉副将刘禹曰‘城降,婴儿老母,口以万数,一旦放兵纵火,闻之可谓酸鼻。家有弊帚,享之千金。禹宗室子孙,故尝更职,何忍行此!’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:飘茵堕溷 比喻由于偶然的机缘而有富贵贫贱的不同命运。亦多指女子堕落风尘。 成语出处:《梁书·儒林传·范缜》子良(竟陵王萧子良)问曰‘君不信因果,世间何得有富贵,何得有贱贫?’缜答曰‘人之生譬如一树花,同发一枝,俱开一蒂,随风而堕,自有拂帘幌坠于茵席之上,自有关篱墙落于粪混之侧。坠茵席者,殿下是也;落粪溷者 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 大题小做是什么意思? 把大题目作成小文章。比喻把重大的问题当做小事情来处理。 成语出处:周作人《鬼的生长》但是千百年来已非一日,载籍浩如烟海,门外摸索,不得象尾,而且鬼界的问题似乎也多得很,尽够研究院里先生们一生的检讨,我这里只提出一个题目,即上面所说的鬼之生长,姑且大题小做,略陈管见,仁候明教。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 犁生髐角 成语释义: 指杂色牛生纯赤色、角周正的小牛。比喻劣父生贤明的儿女。 成语出处:语出《论语·雍也》子谓仲弓,曰‘犁牛之子髐且角,虽欲勿用,山川其舍诸?’邢昺疏杂文曰犁。髐,纯赤色也。角者,角周正也。舍,弃也。诸,之也。仲弓父贱人而行不善,故孔子称谓仲弓,……言仲弓父虽不善,不害于子之美也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:随珠和璧 含义: 随侯珠与和氏璧的并称。泛指珍宝或珍宝中的上品。事见《韩非子·和氏》、《淮南子·览冥训》。 成语出处:传说古代随国姬姓诸侯见一大蛇伤断,以药敷之而愈;后蛇于江中衔明月珠以报德,因曰随侯珠,又称灵蛇珠。楚人卞和于荆山得一璞玉,先后献给武王、文王,均以为石,和以欺君罪被砍断两足;成王登位,使人剖璞,果得夜光宝玉,因命之曰和氏璧。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 颠颠倒倒的含义是什么? ①指神思迷糊错乱。②指事情不顺或言行无条理,不可置信。 成语出处:《朱子全书》卷六向时有一截学者贪多务得,要读《周礼》、诸史、本朝典故,一向尽要理会,得许多没紧要底工夫,少刻,自己都恁自地颠颠倒倒,没顿放处。《二刻拍案惊奇》卷十七彼此夙缘,颠颠倒倒,皆非偶然也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 董狐直笔的意思是什么? 董狐春秋时晋国的史官。直笔根据事实,如实记载。指敢于秉笔直书,尊重史实,不阿权贵的正直史家。 成语出处:《左传·宣公二年》载赵穿杀晋灵公,身为正卿的赵盾没有管,董狐认为赵盾应负责任,便在史策上记载说赵盾弑其君。为赵盾所杀。后孔子称赞说董狐,古之良史也,书法不隐。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 下面词语是什么意思:十八层地狱 层重。地狱佛教、基督教等指死后灵魂受苦的地方。迷信认为人在生时为非作恶,死后进入十八层地狱,不得翻身。比喻悲惨的报应。 成语出处:明·凌濛初《初刻拍案惊奇》卷三十五我赖了你的,我堕十八层地狱。明·冯梦龙《喻世明言》第九卷分明是十八层地狱的苦鬼,直升到三十三天去了。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 过河卒子的含义是什么? 象棋规则中卒子只能向前,不能后退,过了河之后可以横着走,威力更大。比喻只能前进,不能后退。 成语出处:刘绍棠《狼烟》二十四花言巧语蒙哄不了二位老人家,甜言蜜语也迷惑不了二位老人家,于是心慌意乱,闭上眼睛,手捧着怦怦乱跳的胸口,失悔自己的冒险进城,然而已经骑虎难下,只有做一名过河卒子了。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 这句话的含义是什么:蜂窠蚁穴 比喻占据的地方极为窄小,借以对偏安一隅的地方势力的蔑称。 成语出处:宋·曾憊《类说·见闻录·胡讷》开宝八年,王师围金陵。朝廷殿试《桥梁渡长江赋》、《习水战诗》;江南亦试《王德惟亲赋》、《谈笑却秦诗》。太祖笑曰‘江南畜文臣武将,迨同飞走,岂不知中原有真主耶?’赵普曰‘蜂窠蚁穴不足挂圣虑。’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 封胡羯末 这句话的意思是什么? 均为兄弟的小名封指谢韶,胡指谢朗,羯指谢玄,末指谢川。后用以称美兄弟子侄之辞。 成语出处:《晋书·列女传·王凝之妻谢氏》(谢道韫)初適凝之,还,甚不乐。安曰‘王郎,逸少子,不恶,汝何恨也?’答曰‘一门叔父有阿大(谢尚)、中郎(谢据);群从兄弟复有封胡羯末,不意天壤之中乃有王郎!’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 为蛇画足 这个词语是什么意思? 比喻做事节外生枝,不但无益,反而害事。 成语出处:《战国策·齐策二》楚有祠者,赐其舍人卮酒。舍人相谓曰‘数人饮之不足,一人饮之有余;请画地为蛇,先成者饮酒。’一人蛇先成,引酒且饮之,乃左手持卮,右手画蛇,曰‘吾能为之足。’未成,一人之蛇成,夺其卮曰‘蛇固无足,子安能为之 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 断杼择邻 这个成语的意思是什么? 孟母为了教育孟子不要中途荒废学业,用被割断的纱不成布来做比喻;孟母三迁居处,选择良好的环境,来影响孟子的成长。 成语出处:汉·刘向《列女传·母仪传》孟子之少也,既学而归,孟母方绩,问曰‘学何所至矣?’孟子曰‘自若也。’孟母以刀断其织。孟子惧而问其故,孟母曰‘子之学,若吾断斯织也。’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 又红又专 这个词语是什么意思? 具有无产阶级的世界观,又掌握专业知识和专门技术。 成语出处:邓小平《在全国科学大会开幕式上的讲话》绝大多数科学技术人员热爱党、热爱社会主义,努力同工农兵相结合,满腔热情地对待自己从事的科学技术工作,作出了成绩……就整个说来,不愧是我们工人阶级自己的又红又专的科学技术队伍。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 高悬秦镜 成语释义: 秦镜秦始皇时的能照见人心善恶的镱子。高挂能照见人心善恶的镜子。后用以比喻官吏断案公正,执法严明。 成语出处:晋·葛洪《西京杂记》第三卷(咸阳宫)有方镜,广四尺,高五尺九寸……人有疾病在内则掩心而照之,则知病之所在。又女子有邪心,是胆张心动。秦始皇常以照宫人,胆张心动者则杀之。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 盲翁扪钥 这个词语是什么意思? 比喻只凭片面了解或局部经验就对事物妄加判断。亦作盲翁扪籥。 成语出处:宋·苏轼《日喻》生而眇者不识日,问之有目者。或告之曰‘日之状如铜盘。’扣盘而得其声。他日闻锺以为日也。或告之曰‘日之光如烛。’扪烛而得其形,他日揣钥以为日也。日之与锺钥亦远矣,而眇者不知其异,以其未尝见而求之人也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 改梁换柱 成语释义: 比喻玩弄手法,暗中改变事物的内容或事情的性质。 成语出处:郭沫若《文艺论集读·梁任公〈墨子新社会之组织法〉》便是胡適和梁任也都很知道他这种宗教的循环论证不足以满足我们近代人的要求,所以极力在用改梁换柱的方法,要把他的根本观念改移到另一个较为好看一点的节目上去。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:望帝啼鹃 含义: 相传战国时蜀王杜宇称帝,号望帝,为蜀治水有功,后禅位臣子,退隐西山,死后化为杜鹃鸟,啼声凄切。后常指悲哀凄惨的啼哭。 成语出处:元·关汉卿《窦娥冤》若没些儿灵圣与世人传,也不见得湛湛清天。我不要半星血红尘洒,都只在八尺旗枪素练悬。等他四下里皆瞧见,这就是咱苌弘化碧,望帝啼鹃。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 断织之诫 成语释义: 孟子的母亲用割断织布机上的纱,使机上的纱不能成布的损失来告诫中途放弃学业的儿子。后用这个故事告诫中途辍学的人。 成语出处:汉·刘向《列女传·母仪传》孟子之少也,既学而归,孟母方绩,问曰‘学何所至矣?’孟子曰‘自若也。’孟母以刀断其织。孟子惧而问其故,孟母曰‘子之学,若吾断斯织也。’ { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 御沟红叶 成语释义: 御沟流经宫苑的河道。指红叶题诗的故事,后用以比喻男女奇缘。亦作御沟流叶、红叶之题。 成语出处:唐·孟棨《本事诗》记载,顾况在洛阳游苑中,流水上得大梧叶,上有宫女题诗,顾况次日也于上游题诗叶上,泛于波中,以此传情。又一说,题诗宫女名韩翠苹,诗为于祐所得,于又题诗为韩所得,韩、于最终成为夫妻。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 跑龙套 成语释义: 原指戏曲中拿着旗子做兵卒的角色,后比喻在人手下做无关紧要的事。 成语出处:沈从文《跑龙套》跑龙套在戏台上象是个无固定任务角色,姓名通常不上海报,虽然每一出戏文中大将或寨主出场,他都得前台露面打几个转,而且要严肃认真,不言不笑,凡事照规矩行动,随后才必恭必敬的分站两旁。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 扪参历井的意思是什么? 参、井,皆星宿名,分别为蜀秦分野。指自秦入蜀途中,山势高峻,可以摸到参、井两星宿。形容山势高峻,道路险阻。亦形容世路艰难。 成语出处:唐·李白《蜀道难》诗扪参历井仰胁息,以手抚膺坐长叹。宋·王铚《王公四六话》邓温伯知成都谢上表云‘扪参历井,敢辞蜀道之难;就日望云,愈觉长安之远。’自后凡官两川者,谢表相承用此一联。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 伈伈睍睍 这个成语的意思是什么? 伈伈小心恐惧的样子。睍睍也作伣伣,眼睛不敢睁大的样子。小心害怕或低声下气的样子。 成语出处:唐·韩愈《祭鳄鱼文》刺史虽驽弱,亦安肯为鳄鱼低首下心,伈伈睍睍,为民吏羞,以偷活于此耶?《明史·邹智传》及与议事,又唯诺惟谨,伈伈伣伣,若有所不敢,反不如一二俗吏足以任事。此陛下所为疑也,臣窃以为过矣。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 逍遥事外 这个词语是什么意思? ①指犯法者没有受到法律制裁。亦泛指做坏事或与坏事有牵连的人不受追查。②指置身事外,毫不关心。 成语出处:清·包世臣《致广东按察姚中丞书》首祸正凶,逍遥事外。鲁迅《花边文学·女人未必多说谎》关于杨妃,禄山之乱以后的文人就都撒着大谎,玄宗逍遥事外,倒说是许多坏事情都由她。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 岁在龙蛇的含义是什么? 岁,岁星;龙,指辰;蛇,指巳。后指命数当终。 成语出处:《后汉书·郑玄传》五年春,梦孔子告之曰‘起,起,今年岁在辰,来年岁在巳。’既寤,以谶合之,知当命终,有顷寝疾。李贤注北齐刘昼《高才不遇传》论玄曰‘辰为龙,巳为蛇,岁至龙蛇。贤人嗟,玄以谶合之’,盖谓此也。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:见素抱朴 含义: 老子提出的治国的三项具体措施之一。是说要推举圣人,实行法治,即用无为之治取代有为之治。对应于绝圣弃智。 成语出处:《老子》曰见素抱朴。老子《道德经》新解见现,呈现,推出。素没有染色的生丝。这里比喻品质纯洁、高尚的圣人。朴没有加工的原木。这里比喻合乎自然法则的社会法律。。见素抱朴、绝学无忧、少私寡欲是老子提出的治国的三项具体措施。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 成语释义:直上直下 含义: ①上下一贯。②从上到下,从头到脚。③形容陡直。 成语出处:《朱子全书》卷四凡事都分做两边,是底放一边,非底放一边。是底是天理,非底是人欲。是即守而勿失,非即去而不留此治一身之法也。治一家则分别一家之是非,治一邑则分别一邑之邪正,推而一州一路以至天下,莫不皆然此~之道。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 箕山之节 这个词语是什么意思? 箕山古代传说唐尧时的隐士许由、巢父隐居的地方。节名节,节操。指归隐以保全节操。旧时用以称誉不愿在乱世做官的人。亦作箕山之志、箕山之操。 成语出处:《吕氏春秋·求人》昔尧朝许由于沛泽之中,曰‘……请属天下于夫子。’许由辞曰‘为天下之不治与?而既已治矣。自为与?啁噍巢于林,不过一枝;偃鼠饮于河,不过满腹。归已君乎!恶用天下?’遂之箕山之下,颖水之阳,耕而食,终身无经天下之色。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 天保九如 成语释义: 天保《诗经·小雅》中的篇名;九如该诗中连用了九个如字,有棕福寿延绵不绝之意。旧时祝寿的话,棕福寿绵长。 成语出处:《诗经·小雅·天保》天保定尔,以莫不兴。如山如皋,如冈如陵,如川之方至,以莫不增……如月之恒,如日之升,如南山之寿,不骞不崩,如松柏之茂,无不尔或承。 { "major": [ "文本生成" ], "minor": [ "成语释义" ] } [ "中国传统文化", "成语" ] 暂无元数据信息 human true 暂无版权及作者信息 孤鸾照镜的意思是什么? 比喻无偶或失偶者对命运的伤悼。 成语出处:南朝宋范泰《鸾鸟诗》序羪宾王结羋峻祁之山,获一鸾鸟。王甚爱之,欲其鸣而不能致也。乃饰以金樊,飨以珍羞,对之愈戚,三年不鸣。其夫人曰‘尝闻鸟见其类而后鸣,何不悬镜以映之?’王从其言。鸾睹形感契,慨然悲鸣��
2026-01-13T09:29:19
https://docs.softwareheritage.org/genindex.html
Index — Software Heritage documentation Skip to main content Back to top Ctrl + K Development API reference Usage Resources Infrastructure More About GitLab PyPI System Status Software Heritage Homepage Development API reference Usage Resources Infrastructure About GitLab PyPI System Status Software Heritage Homepage Index Symbols | A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z Symbols --after swh-scheduler-task-list command line option --all swh-db-init command line option , [1] --all-origins swh-graph-find-context command line option --api-url swh-scanner-scan command line option --archive swh-deposit command line option --archive-deposit swh-deposit command line option --athena-prefix swh-datasets-luigi command line option --author swh-deposit command line option --backoff-factor swh-scheduler-task_type-add command line option --base-directory swh-datasets-luigi command line option --base-rust-executable-dir swh-datasets-luigi command line option --base-sensitive-directory swh-datasets-luigi command line option --batch-size swh-datasets-luigi command line option swh-indexer-journal-client command line option swh-scheduler-task-schedule_origins command line option --before swh-scheduler-task-list command line option swh-scheduler-task-list-pending command line option --broker swh-indexer-journal-client command line option --bundle-type swh-vault-cook command line option --check-config swh-storage command line option --check-dst swh-objstorage-replay command line option , [1] --check-flavor swh-graph-compress command line option --check-hashes swh-scrubber-check-init command line option --check-references swh-scrubber-check-init command line option --check-src-hashes swh-objstorage-replay command line option , [1] --clear swh-storage-masking-patching-set command line option --client-id swh-auth command line option --collection swh-deposit command line option swh-deposit-admin-user-create command line option --columns swh-scheduler-task-schedule command line option --concise swh-indexer-mapping-list-terms command line option --concurrency swh-objstorage-replay command line option , [1] --config-file swh-auth command line option swh-datasets command line option swh-deposit-admin command line option swh-fs command line option swh-graph command line option swh-indexer command line option swh-lister command line option swh-loader command line option swh-objstorage command line option swh-scheduler command line option swh-scrubber command line option swh-search command line option swh-storage command line option swh-vault-cook command line option swh-vault-rpc-serve command line option --config-id swh-scrubber-check-run command line option swh-scrubber-check-running command line option swh-scrubber-check-stalled command line option swh-scrubber-check-stats command line option --content-swhid swh-graph-find-context command line option --create-origin swh-deposit command line option --daemon swh-fs-mount command line option --database swh-scheduler command line option --dataset-name swh-datasets-luigi command line option --db-name swh-db-init command line option , [1] --dbname swh-db-init command line option , [1] --debug swh-indexer-rpc-serve command line option swh-objstorage-rpc-serve command line option swh-scheduler-rpc-serve command line option swh-search-rpc-serve command line option swh-storage-rpc-serve command line option swh-vault-rpc-serve command line option --debug-http swh-scanner-scan command line option --default-interval swh-scheduler-task_type-add command line option --delimiter swh-scheduler-task-schedule command line option --deposit-id swh-deposit command line option , [1] swh-deposit-admin-deposit-reschedule command line option --dereference swh-identify command line option --disable-global-patterns swh-scanner-scan command line option --disable-vcs-patterns swh-scanner-scan command line option --domain swh-deposit-admin-user-create command line option --dry-run swh-scheduler-task-schedule_origins command line option swh-storage-backfill command line option --ef swh-graph-reindex command line option --email swh-deposit-admin-user-create command line option --end-object swh-scrubber-fix command line option swh-scrubber-locate command line option swh-storage-backfill command line option --exclude swh-identify command line option swh-scanner-scan command line option --exclude-cleared-requests swh-storage-blocking-list-requests command line option swh-storage-masking-list-requests command line option --exclude-mapping swh-indexer-mapping-list-terms command line option --exclude-sha1-file swh-objstorage-replay command line option , [1] --exclude-template swh-scanner-scan command line option --export-base-directory swh-datasets-luigi command line option --export-name swh-datasets-luigi command line option --file swh-storage-blocking-update-objects command line option swh-storage-masking-update-objects command line option --filename swh-graph-find-context command line option swh-identify command line option --firstname swh-deposit-admin-user-create command line option --flavor swh-db-init command line option , [1] --for swh-scrubber-check-stalled command line option --force swh-graph-reindex command line option swh-storage-remove-old-object-reference-partitions command line option --foreground swh-fs-mount command line option --format swh-deposit command line option , [1] , [2] swh-scheduler-celery-monitor-list-running command line option --fqswhid swh-graph-find-context command line option --graph swh-graph-grpc-serve command line option swh-graph-rpc-serve command line option --graph-base-directory swh-datasets-luigi command line option --graph-grpc-server swh-graph-find-context command line option --graph-name swh-graph-compress command line option --group-id swh-indexer-journal-client command line option --grpc-api swh-datasets-luigi command line option --host swh-graph-rpc-serve command line option swh-indexer-rpc-serve command line option swh-objstorage-rpc-serve command line option swh-scheduler-rpc-serve command line option swh-search-rpc-serve command line option swh-storage-rpc-serve command line option swh-vault-rpc-serve command line option --include-cleared-requests swh-storage-blocking-list-requests command line option swh-storage-masking-list-requests command line option --input-dataset swh-graph-compress command line option --json swh-scrubber-check-stats command line option --keep swh-storage-masking-patching-set command line option --known-mismatched-hashes swh-storage-replay command line option --lastname swh-deposit-admin-user-create command line option --limit swh-scheduler-task-list command line option swh-scheduler-task-list-pending command line option swh-scheduler-task-schedule_origins command line option swh-scrubber-check-run command line option --list-runs swh-scheduler-task-list command line option --lister swh-lister-run command line option --log-config swh command line option --log-level swh command line option --luigi-config swh-datasets-luigi command line option --manage-rw-images swh-objstorage-winery-rbd command line option --max-interval swh-scheduler-task_type-add command line option --max-ram swh-datasets-luigi command line option --message swh-storage-blocking-clear-request command line option swh-storage-blocking-new-request command line option swh-storage-blocking-update-objects command line option swh-storage-masking-clear-request command line option swh-storage-masking-new-request command line option swh-storage-masking-update-objects command line option --metadata swh-deposit command line option , [1] --metadata-deposit swh-deposit command line option --metadata-provenance-url swh-deposit command line option --migration swh-storage-cassandra-mark-upgraded command line option swh-storage-cassandra-upgrade command line option --min-interval swh-scheduler-task_type-add command line option --min-mapped-hosts swh-objstorage-winery-rw-shard-cleaner command line option --module-config-key swh-db-init command line option , [1] --module-is-path swh-db-init command line option , [1] --name swh-datasets-download-export command line option swh-datasets-download-graph command line option swh-deposit command line option swh-deposit-admin-collection-create command line option swh-graph-download command line option swh-scrubber-check-init command line option --nb-partitions swh-scrubber-check-init command line option --next-run swh-scheduler-task-add command line option swh-scheduler-task-respawn command line option --no-all-origins swh-graph-find-context command line option --no-archive-deposit swh-deposit command line option --no-check-dst swh-objstorage-replay command line option , [1] --no-check-hashes swh-scrubber-check-init command line option --no-check-references swh-scrubber-check-init command line option --no-debug swh-objstorage-rpc-serve command line option swh-search-rpc-serve command line option swh-storage-rpc-serve command line option swh-vault-rpc-serve command line option --no-dereference swh-identify command line option --no-dry-run swh-scheduler-task-schedule_origins command line option --no-filename swh-identify command line option --no-fqswhid swh-graph-find-context command line option --no-metadata-deposit swh-deposit command line option --no-partial swh-deposit command line option --no-random-origin swh-graph-find-context command line option --no-replace swh-deposit command line option --no-stdout swh-scheduler command line option --no-trace swh-graph-find-context command line option --no-verbose swh-deposit command line option --no-web-ui swh-scanner-scan command line option --nodebug swh-indexer-rpc-serve command line option swh-scheduler-rpc-serve command line option --object-type swh-scrubber-check-init command line option swh-search-journal-client-objects command line option --oidc-server-url swh-auth command line option --only-prefix swh-objstorage-winery-rbd command line option --origin-url swh-graph-find-context command line option --output-directory swh-graph-compress command line option --output-format swh-scanner-scan command line option --page-token swh-scheduler-task-schedule_origins command line option --parallelism swh-datasets-download-export command line option swh-datasets-download-graph command line option swh-graph-download command line option --parent-dataset-name swh-datasets-luigi command line option --parent-export-name swh-datasets-luigi command line option --partial swh-deposit command line option --password swh-auth-generate-token command line option swh-deposit command line option , [1] , [2] swh-deposit-admin-user-create command line option --pattern swh-scheduler-celery-monitor command line option --period swh-scheduler-runner command line option --platform swh-deposit-admin command line option --plugins swh-scheduler-task_type-register command line option --policy swh-scheduler-task-add command line option swh-scheduler-task-list command line option --port swh-graph-grpc-serve command line option swh-graph-rpc-serve command line option swh-indexer-rpc-serve command line option swh-objstorage-rpc-serve command line option swh-scheduler-rpc-serve command line option swh-search-rpc-serve command line option swh-storage-rpc-serve command line option swh-vault-rpc-serve command line option --prefix swh-indexer-journal-client command line option swh-search-journal-client-objects command line option --previous-dataset-name swh-datasets-luigi command line option --priority swh-scheduler-task-add command line option swh-scheduler-task-list command line option --profile swh-graph command line option --project-config-file swh-scanner-scan command line option --provenance swh-scanner-scan command line option --provenance-batch-size swh-scanner-scan command line option --provenance-concurrency swh-scanner-scan command line option --provider-url swh-deposit-admin-user-create command line option --random-origin swh-graph-find-context command line option --realm-name swh-auth command line option --recursive swh-identify command line option --replace swh-deposit command line option --reset swh-scrubber-check-stalled command line option --retry-luigi-delay swh-datasets-luigi command line option --s3-athena-output-location swh-datasets-luigi command line option --s3-bucket swh-datasets-list command line option swh-graph-list-datasets command line option --s3-prefix swh-datasets-download-export command line option swh-datasets-download-graph command line option swh-datasets-luigi command line option swh-graph-download command line option --s3-url swh-graph-download command line option --sensitive-input-dataset swh-graph-compress command line option --sensitive-output-directory swh-graph-compress command line option --sentry-dsn swh command line option --size-limit swh-objstorage-replay command line option , [1] --slug swh-deposit command line option --start-object swh-scrubber-fix command line option swh-scrubber-locate command line option swh-storage-backfill command line option --status swh-scheduler-task-list command line option --steps swh-graph-compress command line option --stop-after-objects swh-indexer-journal-client command line option swh-objstorage-replay command line option , [1] swh-search-journal-client-objects command line option swh-storage-replay command line option --stop-after-shards swh-objstorage-winery-packer command line option swh-objstorage-winery-rw-shard-cleaner command line option --stop-instead-of-waiting swh-objstorage-winery-rbd command line option swh-objstorage-winery-rw-shard-cleaner command line option --storage-url swh-scheduler-task-schedule_origins command line option --swhid swh-deposit command line option --task-id swh-scheduler-task-list command line option --task-type swh-scheduler-runner command line option swh-scheduler-task-list command line option --task-type-pattern swh-scheduler-runner command line option --task_name swh-scheduler-task_type-list command line option --task_type swh-scheduler-task_type-list command line option --timeout swh-scheduler-celery-monitor command line option --token swh-auth-config command line option --trace swh-graph-find-context command line option --type swh-identify command line option swh-storage-replay command line option --url swh-deposit command line option , [1] , [2] swh-scheduler command line option --use-journal swh-scrubber-check-run command line option --username swh-auth-config command line option swh-deposit command line option , [1] , [2] swh-deposit-admin-user-create command line option --verbose swh-deposit command line option swh-graph-link command line option swh-scheduler-task_type-list command line option --verify swh-identify command line option --web-ui swh-scanner-scan command line option --with-priority swh-scheduler-runner command line option --without-priority swh-scheduler-runner command line option -a swh-db-init command line option , [1] swh-scheduler-task-list command line option swh-storage-blocking-list-requests command line option swh-storage-masking-list-requests command line option -b swh-indexer-journal-client command line option swh-scheduler-task-list command line option swh-scheduler-task-list-pending command line option swh-scheduler-task-schedule_origins command line option -C swh-auth command line option swh-datasets command line option swh-deposit-admin command line option swh-fs command line option swh-graph command line option swh-indexer command line option swh-lister command line option swh-loader command line option swh-objstorage command line option swh-scheduler command line option swh-scrubber command line option swh-search command line option swh-storage command line option swh-vault-cook command line option swh-vault-rpc-serve command line option -c swh-auth command line option swh-graph-find-context command line option swh-scanner-scan command line option swh-scheduler-task-schedule command line option -d swh-db-init command line option , [1] swh-fs-mount command line option swh-scheduler command line option swh-scheduler-task-schedule command line option -f swh-deposit command line option , [1] , [2] swh-fs-mount command line option swh-graph-find-context command line option swh-scanner-scan command line option swh-scheduler-task_type-add command line option swh-storage-blocking-update-objects command line option swh-storage-masking-update-objects command line option -g swh-graph-compress command line option swh-graph-find-context command line option swh-graph-grpc-serve command line option swh-graph-rpc-serve command line option swh-scheduler-task-schedule_origins command line option -h swh-graph-rpc-serve command line option -i swh-graph-compress command line option swh-scheduler-task-list command line option swh-scheduler-task_type-add command line option , [1] -j swh-datasets-download-export command line option swh-datasets-download-graph command line option swh-graph-download command line option swh-scrubber-check-stats command line option -l swh command line option swh-lister-run command line option swh-scheduler-task-list command line option swh-scheduler-task-list-pending command line option -m swh-indexer-journal-client command line option swh-search-journal-client-objects command line option swh-storage-blocking-clear-request command line option swh-storage-blocking-new-request command line option swh-storage-blocking-update-objects command line option swh-storage-masking-clear-request command line option swh-storage-masking-new-request command line option swh-storage-masking-update-objects command line option -n swh-objstorage-replay command line option , [1] swh-scheduler-task-add command line option swh-scheduler-task-respawn command line option swh-scheduler-task_type-list command line option swh-storage-replay command line option -o swh-graph-compress command line option swh-graph-find-context command line option swh-search-journal-client-objects command line option -P swh-scheduler-task-add command line option swh-scheduler-task-list command line option -p swh-auth-generate-token command line option swh-db-init command line option , [1] swh-graph-grpc-serve command line option swh-graph-rpc-serve command line option swh-objstorage-rpc-serve command line option swh-scheduler-runner command line option swh-scheduler-task-add command line option swh-scheduler-task-list command line option swh-scheduler-task_type-register command line option swh-search-journal-client-objects command line option -r swh-auth command line option swh-identify command line option swh-scheduler-task-list command line option -s swh-graph-compress command line option swh-scheduler-task-list command line option -t swh-identify command line option swh-scanner-scan command line option swh-scheduler-task-list command line option swh-scheduler-task_type-list command line option swh-storage-replay command line option -u swh-auth command line option swh-scanner-scan command line option swh-scheduler command line option -v swh-graph-link command line option swh-identify command line option swh-scheduler-task_type-list command line option -X swh-storage-replay command line option -x swh-identify command line option swh-scanner-scan command line option [SWHID]... swh-fs-mount command line option A abort() (swh.loader.svn.replay.Editor method) absolute_uri_validator() (in module swh.deposit.loader.checks) absolute_url() (in module swh.web.alter.templatetags.alter_extras) abstract (swh.coarnotify.server.models.Notification.Meta attribute) (swh.web.alter.models.BaseModel.Meta attribute) AbstractCache (class in swh.fuse.cache) accept_all_directories() (in module swh.model.from_disk) accept_all_paths() (in module swh.model.from_disk) accept_mimetypes (swh.core.api.negotiation.Negotiator property) ACCEPTED (swh.coarnotify.server.models.Statuses attribute) (swh.web.add_forge_now.models.RequestStatus attribute) access_token (swh.auth.django.models.OIDCUser attribute) active (swh.web.alter.utils.Step attribute) Actor (class in swh.coarnotify.server.models) actor (swh.web.add_forge_now.models.RequestHistory attribute) Actor.DoesNotExist Actor.MultipleObjectsReturned actor_role (swh.web.add_forge_now.models.RequestHistory attribute) actor_set (swh.coarnotify.server.models.Organization attribute) ActorAdmin (class in swh.coarnotify.server.admin) ActorChangeForm (class in swh.coarnotify.server.forms) ActorChangeForm.Meta (class in swh.coarnotify.server.forms) ActorCreationForm (class in swh.coarnotify.server.forms) ActorCreationForm.Meta (class in swh.coarnotify.server.forms) ActorManager (class in swh.coarnotify.server.models) add() (swh.core.collections.SortedList method) (swh.counters.api.client.RemoteCounters method) (swh.counters.in_memory.InMemory method) (swh.counters.interface.CountersInterface method) (swh.counters.redis.Redis method) (swh.export.test.test_edges.FakeDiskSet method) (swh.export.utils.LevelDBSet method) (swh.export.utils.SQLiteSet method) (swh.indexer.storage.in_memory.SubStorage method) (swh.objstorage.api.client.RemoteObjStorage method) (swh.objstorage.backends.azure.AzureCloudObjStorage method) (swh.objstorage.backends.http.HTTPReadOnlyObjStorage method) (swh.objstorage.backends.in_memory.InMemoryObjStorage method) (swh.objstorage.backends.libcloud.CloudObjStorage method) (swh.objstorage.backends.noop.NoopObjStorage method) (swh.objstorage.backends.pathslicing.PathSlicingObjStorage method) (swh.objstorage.backends.seaweedfs.objstorage.SeaweedFilerObjStorage method) (swh.objstorage.backends.winery.objstorage.WineryObjStorage method) (swh.objstorage.backends.winery.objstorage.WineryWriter method) (swh.objstorage.backends.winery.roshard.ROShardCreator method) (swh.objstorage.backends.winery.rwshard.RWShard method) (swh.objstorage.backends.winery.throttler.BandwidthCalculator method) (swh.objstorage.backends.winery.throttler.IOThrottler method) (swh.objstorage.backends.winery.throttler.LeakyBucket method) (swh.objstorage.backends.winery.WineryObjStorage method) (swh.objstorage.interface.ObjStorageInterface method) (swh.objstorage.multiplexer.MultiplexerObjStorage method) (swh.objstorage.proxies.readonly.ReadOnlyProxyObjStorage method) (swh.vault.cache.VaultCache method) add_alias() (swh.core.cli.AliasedGroup method) add_arguments() (swh.coarnotify.server.management.commands.createorganization.Command method) (swh.coarnotify.server.management.commands.createuser.Command method) (swh.web.mailmap.management.commands.sync_mailmaps.Command method) (swh.web.mailmap.management.commands.sync_masking_mailmaps.Command method) (swh.web.save_code_now.management.commands.dump_savecodenow_data.Command method) (swh.web.utils.management.commands.rename_app.Command method) add_backend_class() (swh.core.api.RPCServerApp method) add_batch() (swh.objstorage.api.client.RemoteObjStorage method) (swh.objstorage.factory.ObjStorage method) (swh.objstorage.interface.ObjStorageInterface method) (swh.objstorage.multiplexer.MultiplexerObjStorage method) (swh.objstorage.objstorage.ObjStorage method) add_content() (swh.loader.cvs.loader.CvsLoader method) add_directory() (swh.loader.svn.replay.DirEditor method) add_doc_route() (swh.web.api.apiurls.APIUrls method) add_edge() (swh.alter.subgraph.Subgraph method) add_error() (swh.storage.proxies.tenacious.RateQueue method) add_fieldsets (swh.coarnotify.server.admin.ActorAdmin attribute) add_file() (swh.loader.svn.replay.DirEditor method) add_files_to_s3_bucket() (in module swh.core.s3.pytest_plugin) add_forge_now_request_dashboard() (in module swh.web.add_forge_now.admin_views) add_forge_now_requests_moderation_dashboard() (in module swh.web.add_forge_now.admin_views) add_forge_request_list_datatables() (in module swh.web.add_forge_now.views) add_form (swh.coarnotify.server.admin.ActorAdmin attribute) add_id_keyword() (swh.loader.cvs.cvs2gitdump.cvs2gitdump.RcsKeywords method) add_list() (in module swh.indexer.metadata_dictionary.utils) add_map() (in module swh.indexer.metadata_dictionary.utils) add_notif_email() (swh.vault.backend.VaultBackend method) add_ok() (swh.storage.proxies.tenacious.RateQueue method) add_pagination_clause() (swh.storage.postgresql.db.QueryBuilder method) add_provenance() (in module swh.scanner.data) add_ProvenanceServiceServicer_to_server() (in module swh.provenance.grpc.swhprovenance_pb2_grpc) add_query_part() (swh.storage.postgresql.db.QueryBuilder method) add_redirect_for_checksum_args() (swh.web.utils.urlsindex.UrlsIndex method) add_swhid() (swh.alter.subgraph.Subgraph method) add_swhids() (swh.alter.subgraph.Subgraph method) add_TraversalServiceServicer_to_server() (in module swh.graph.grpc.swhgraph_pb2_grpc) add_url_if_valid() (in module swh.indexer.metadata_dictionary.utils) add_url_pattern() (swh.web.utils.urlsindex.UrlsIndex method) add_vertex() (swh.alter.subgraph.Subgraph method) AddForgeNowConfig (class in swh.web.add_forge_now.apps) AddForgeNowRequestForm (class in swh.web.add_forge_now.api_views) AddForgeNowRequestForm.Meta (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistoryForm (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistoryForm.Meta (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistoryPublicSerializer (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistoryPublicSerializer.Meta (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistorySerializer (class in swh.web.add_forge_now.api_views) AddForgeNowRequestHistorySerializer.Meta (class in swh.web.add_forge_now.api_views) AddForgeNowRequestPublicSerializer (class in swh.web.add_forge_now.api_views) AddForgeNowRequestPublicSerializer.Meta (class in swh.web.add_forge_now.api_views) AddForgeNowRequestSerializer (class in swh.web.add_forge_now.api_views) AddForgeNowRequestSerializer.Meta (class in swh.web.add_forge_now.api_views) additional_checks() (swh.deposit.api.common.APIBase method) (swh.deposit.api.private.deposit_update_status.APIUpdateStatus method) ADDRESS_SIGNER_DEFAULT (in module swh.web.inbound_email.utils) ADDRESS_SIGNER_SEP (in module swh.web.inbound_email.utils) ADDRESS_SIGNER_SUPPORTED (in module swh.web.inbound_email.utils) AddressMatch (class in swh.web.inbound_email.utils) admin_alteration() (in module swh.web.alter.views) admin_dashboard() (in module swh.web.alter.views) admin_deposit() (in module swh.web.deposit.urls) admin_deposit_list() (in module swh.web.deposit.urls) admin_event() (in module swh.web.alter.views) admin_mailmap() (in module swh.web.mailmap.views) admin_message() (in module swh.web.alter.views) admin_origin() (in module swh.web.alter.views) admin_origin_save_add_authorized_url() (in module swh.web.save_code_now.admin_views) admin_origin_save_add_unauthorized_url() (in module swh.web.save_code_now.admin_views) admin_origin_save_authorized_urls_list() (in module swh.web.save_code_now.admin_views) admin_origin_save_filters() (in module swh.web.save_code_now.admin_views) admin_origin_save_remove_authorized_url() (in module swh.web.save_code_now.admin_views) admin_origin_save_remove_unauthorized_url() (in module swh.web.save_code_now.admin_views) admin_origin_save_request_accept() (in module swh.web.save_code_now.admin_views) admin_origin_save_request_reject() (in module swh.web.save_code_now.admin_views) admin_origin_save_request_remove() (in module swh.web.save_code_now.admin_views) admin_origin_save_requests() (in module swh.web.save_code_now.admin_views) admin_origin_save_requests_csv_dump() (in module swh.web.save_code_now.admin_views) admin_origin_save_unauthorized_urls_list() (in module swh.web.save_code_now.admin_views) AdminConfig (class in swh.web.admin.apps) age_decrypt() (in module swh.alter.recovery_bundle) age_decrypt_from_identity() (in module swh.alter.recovery_bundle) age_encrypt() (in module swh.alter.recovery_bundle) age_encrypt_armored() (in module swh.alter.recovery_bundle) aggregate_datasets_path (swh.datasets.luigi.aggregate_datasets.AggregateContentDatasets attribute) (swh.datasets.luigi.aggregate_datasets.ExportNodesTable attribute) (swh.datasets.luigi.aggregate_datasets.RunAggregatedDatasets attribute) (swh.datasets.luigi.aggregate_datasets.UploadAggregatedContentDataset attribute) (swh.datasets.luigi.aggregate_datasets.UploadNodesTable attribute) aggregate_tarballs() (in module swh.deposit.api.private.deposit_read) (in module swh.loader.package.deposit.loader) AggregateContentDatasets (class in swh.datasets.luigi.aggregate_datasets) aid (swh.loader.package.maven.loader.ArtifactDict attribute) (swh.loader.package.maven.loader.MavenPackageInfo attribute) algorithm (swh.graph.libs.luigi.topology.TopoSort attribute) ALGORITHMS (in module swh.model.hashutil) ALIAS (swh.model.model.SnapshotTargetType attribute) alias (swh.web.utils.typing.SnapshotBranchInfo attribute) (swh.web.utils.typing.SnapshotReleaseInfo attribute) AliasedGroup (class in swh.core.cli) aliases_followed (swh.storage.interface.SnapshotBranchByNameResponse attribute) all() (swh.objstorage.backends.winery.rwshard.RWShard method) all_packages_count (swh.lister.elm.lister.ElmListerState attribute) ALL_PACKAGES_URL_PATTERN (swh.lister.elm.lister.ElmLister attribute) all_tokens() (in module swh.core.github.pytest_plugin) allow_request() (swh.web.api.throttling.SwhWebRateThrottle method) (swh.web.api.throttling.SwhWebUserRateThrottle method) allowed_next_statuses() (swh.web.add_forge_now.models.RequestStatus method) Alteration (class in swh.web.alter.models) alteration (swh.web.alter.models.Event attribute) (swh.web.alter.models.Origin attribute) (swh.web.alter.models.Token attribute) Alteration.DoesNotExist Alteration.MultipleObjectsReturned alteration_access() (in module swh.web.alter.views) alteration_details() (in module swh.web.alter.views) alteration_id (swh.web.alter.models.Event attribute) (swh.web.alter.models.Origin attribute) (swh.web.alter.models.Token attribute) alteration_link() (in module swh.web.alter.views) alteration_message() (in module swh.web.alter.views) AlterationAccessForm (class in swh.web.alter.forms) AlterationAdminForm (class in swh.web.alter.forms) AlterationAdminForm.Meta (class in swh.web.alter.forms) AlterationCategory (class in swh.web.alter.models) AlterationForm (class in swh.web.alter.forms) AlterationForm.Meta (class in swh.web.alter.forms) AlterationManager (class in swh.web.alter.models) AlterationSearchForm (class in swh.web.alter.forms) AlterationSearchForm.Meta (class in swh.web.alter.forms) AlterationStatus (class in swh.web.alter.models) AlterConfig (class in swh.web.alter.apps) anchor (swh.model.swhids.QualifiedSWHID attribute) (swh.web.utils.typing.SWHIDContext attribute) anonymize() (swh.alter.notifications.RemovalNotification method) (swh.indexer.storage.model.BaseRow method) (swh.journal.writer.interface.ValueProtocol method) (swh.model.model.BaseModel method) (swh.model.model.Person method) (swh.model.model.Release method) (swh.model.model.Revision method) AnonymousAuthBackend (class in swh.graphql.server) any_permission_required() (in module swh.web.auth.utils) api_add_forge_request_create() (in module swh.web.add_forge_now.api_views) api_add_forge_request_get() (in module swh.web.add_forge_now.api_views) api_add_forge_request_list() (in module swh.web.add_forge_now.api_views) api_add_forge_request_update() (in module swh.web.add_forge_now.api_views) API_BASE_URL (swh.lister.cpan.lister.CpanLister attribute) api_check_content_known() (in module swh.web.api.views.content) api_content_filetype() (in module swh.web.api.views.content) api_content_language() (in module swh.web.api.views.content) api_content_license() (in module swh.web.api.views.content) api_content_metadata() (in module swh.web.api.views.content) api_content_raw() (in module swh.web.api.views.content) api_directory() (in module swh.web.api.views.directory) api_doc() (in module swh.web.api.apidoc) api_endpoints() (in module swh.web.api.views.utils) api_exception (swh.core.api.RPCClient attribute) (swh.indexer.storage.api.client.RemoteStorage attribute) (swh.objstorage.api.client.RemoteObjStorage attribute) (swh.storage.api.client.RemoteStorage attribute) api_extid() (in module swh.web.api.views.extid) api_extid_target() (in module swh.web.api.views.extid) api_graph() (in module swh.web.api.views.graph) api_graph_proxy() (in module swh.web.api.views.graph) api_home() (in module swh.web.api.views.utils) API_INDEX_URL (swh.lister.nuget.lister.NugetLister attribute) api_lookup() (in module swh.web.api.views.utils) api_origin() (in module swh.web.api.views.origin) api_origin_extrinsic_metadata() (in module swh.web.api.views.origin) api_origin_intrinsic_metadata() (in module swh.web.api.views.origin) api_origin_intrinsic_metadata_legacy() (in module swh.web.api.views.origin) api_origin_metadata_search() (in module swh.web.api.views.origin) api_origin_save_bulk() (in module swh.web.save_bulk.api_views) api_origin_save_bulk_request_info() (in module swh.web.save_bulk.api_views) api_origin_save_bulk_requests() (in module swh.web.save_bulk.api_views) api_origin_search() (in module swh.web.api.views.origin) api_origin_visit() (in module swh.web.api.views.origin) api_origin_visit_latest() (in module swh.web.api.views.origin) api_origin_visits() (in module swh.web.api.views.origin) api_origins() (in module swh.web.api.views.origin) API_PROJECTS_ENDPOINT (swh.lister.pagure.lister.PagureLister attribute) api_provenance_whereare() (in module swh.web.provenance.api_views) api_provenance_whereis() (in module swh.web.provenance.api_views) api_raw_extrinsic_metadata_get() (in module swh.web.api.views.metadata) api_raw_extrinsic_metadata_origin_authorities() (in module swh.web.api.views.metadata) api_raw_extrinsic_metadata_swhid() (in module swh.web.api.views.metadata) api_raw_extrinsic_metadata_swhid_authorities() (in module swh.web.api.views.metadata) api_raw_intrinsic_citation_origin_get() (in module swh.web.api.views.citation) api_raw_intrinsic_citation_swhid_get() (in module swh.web.api.views.citation) api_raw_object() (in module swh.web.api.views.raw) api_release() (in module swh.web.api.views.release) API_REPOSITORY_PATH (swh.lister.phabricator.lister.PhabricatorLister attribute) api_request() (swh.lister.golang.lister.GolangLister method) (swh.lister.packagist.lister.PackagistLister method) api_resolve_swhid() (in module swh.web.api.views.identifiers) api_revision() (in module swh.web.api.views.revision) api_revision_directory() (in module swh.web.api.views.revision) api_revision_log() (in module swh.web.api.views.revision) api_revision_raw_message() (in module swh.web.api.views.revision) api_route() (in module swh.web.api.apiurls) api_save_bulk_origins_list() (in module swh.web.save_bulk.views) api_save_origin() (in module swh.web.save_code_now.api_views) api_snapshot() (in module swh.web.api.views.snapshot) api_stats() (in module swh.web.api.views.stat) api_swhid_known() (in module swh.web.api.views.identifiers) API_URL (swh.lister.bitbucket.lister.BitbucketLister attribute) (swh.lister.bower.lister.BowerLister attribute) (swh.lister.github.lister.GitHubLister attribute) api_vault_cook_directory() (in module swh.web.vault.api_views) api_vault_cook_flat() (in module swh.web.vault.api_views) api_vault_cook_git_bare() (in module swh.web.vault.api_views) api_vault_cook_gitfast() (in module swh.web.vault.api_views) api_vault_cook_revision_gitfast() (in module swh.web.vault.api_views) api_vault_download_directory() (in module swh.web.vault.api_views) api_vault_download_flat() (in module swh.web.vault.api_views) api_vault_download_revision_git_bare() (in module swh.web.vault.api_views) api_vault_download_revision_gitfast() (in module swh.web.vault.api_views) api_view() (in module swh.deposit.api.urls) APIBase (class in swh.deposit.api.common) ApiClient (class in swh.loader.package.deposit.loader) APIConfig (class in swh.deposit.config) APIDelete (class in swh.deposit.api.common) apidoc_routes (swh.web.api.apiurls.APIUrls attribute) APIDocException APIError , [1] APIGet (class in swh.deposit.api.common) APIList (class in swh.deposit.api.private.deposit_list) APIPost (class in swh.deposit.api.common) APIPrivateView (class in swh.deposit.api.private) APIPut (class in swh.deposit.api.common) APIReadArchives (class in swh.deposit.api.private.deposit_read) APIReadMetadata (class in swh.deposit.api.private.deposit_read) APIReleases (class in swh.deposit.api.private.deposit_releases) APIUpdateStatus (class in swh.deposit.api.private.deposit_update_status) APIUploadURLs (class in swh.deposit.api.private.deposit_upload_urls) APIUrls (class in swh.web.api.apiurls) application_name (swh.objstorage.backends.winery.settings.Database attribute) apply_migrations() (in module swh.storage.cassandra.migrations) apply_options() (in module swh.core.db.common) apply_textdelta() (swh.loader.svn.replay.FileEditor method) ARCH_API_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) ARCH_PACKAGE_DOWNLOAD_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) ARCH_PACKAGE_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) ARCH_PACKAGE_VERSIONS_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) archive Archive (class in swh.graphql.backends.archive) archive (swh.deposit.models.DepositRequest attribute) archive_get() (swh.deposit.client.PrivateApiDepositClient method) ARCHIVE_URL_PATTERN (swh.lister.conda.lister.CondaLister attribute) ARCHIVED (swh.web.alter.models.AlterationStatus attribute) ArchiveDir (class in swh.fuse.fs.mountpoint) ArchiveDiscoveryInterface (class in swh.model.discovery) ArchiveLoader (class in swh.loader.package.archive.loader) ArchivePackageInfo (class in swh.loader.package.archive.loader) ArchLister (class in swh.lister.arch.lister) ArchLoader (class in swh.loader.package.arch.loader) ArchPackageInfo (class in swh.loader.package.arch.loader) args (swh.scheduler.model.TaskArguments attribute) argument_error_handler() (in module swh.indexer.storage.api.server) (in module swh.objstorage.api.server) (in module swh.scheduler.api.server) (in module swh.vault.api.server) arguments (swh.scheduler.model.Task attribute) ark ARM_PACKAGE_DOWNLOAD_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) ARM_PACKAGE_URL_PATTERN (swh.lister.arch.lister.ArchLister attribute) artifact Artifact (class in swh.lister.nixguix.lister) ARTIFACT (swh.lister.nixguix.lister.ArtifactType attribute) artifact_to_listed_origin() (swh.lister.nixguix.lister.NixGuixLister method) ArtifactDict (class in swh.loader.package.maven.loader) ArtifactNatureMistyped ArtifactNatureUndetected artifacts (swh.lister.gnu.tree.GNUTree property) ArtifactType (class in swh.lister.nixguix.lister) ArtifactWithoutExtension ASC (swh.storage.interface.ListOrder attribute) assert_all_objects_consumed() (in module swh.journal.pytest_plugin) assert_exported_objects() (in module swh.export.test.test_journal_processor) assert_never() (in module swh.vault.cookers.git_bare) assert_origin_exists() (in module swh.storage.migrate_extrinsic_metadata) assistant_category() (in module swh.web.alter.views) assistant_email() (in module swh.web.alter.views) assistant_email_verification() (in module swh.web.alter.views) assistant_origins() (in module swh.web.alter.views) assistant_reasons() (in module swh.web.alter.views) assistant_summary() (in module swh.web.alter.views) athena_db_name (swh.export.luigi.CreateAthena attribute) (swh.export.luigi.RunExportAll attribute) (swh.graph.luigi.subdataset.CreateSubdatasetOnAthena attribute) athena_parent_db_name (swh.graph.luigi.subdataset.CreateSubdatasetOnAthena attribute) AthenaDatabaseTarget (class in swh.export.luigi) atomic_csv_zstd_writer() (in module swh.datasets.luigi.blobs_datasets) atomic_zstd_writer() (in module swh.datasets.luigi.blobs_datasets) AtomicFileSink (class in swh.graph.shell) AurLister (class in swh.lister.aur.lister) AurLoader (class in swh.loader.package.aur.loader) AurPackageInfo (class in swh.loader.package.aur.loader) auth_token (swh.coarnotify.server.models.Actor attribute) AuthConfig (class in swh.web.auth.apps) authenticate() (swh.auth.django.backends.OIDCAuthorizationCodePKCEBackend method) (swh.auth.django.backends.OIDCBearerTokenAuthentication method) (swh.auth.starlette.backends.BearerTokenAuthBackend method) (swh.graphql.server.AnonymousAuthBackend method) authenticate_credentials() (swh.deposit.auth.KeycloakBasicAuthentication method) author (swh.loader.cvs.rlog.revtuple attribute) (swh.loader.mercurial.identify.HgRevision attribute) (swh.loader.package.cpan.loader.CpanPackageInfo attribute) (swh.loader.package.hackage.loader.HackagePackageInfo attribute) (swh.loader.package.hex.loader.HexPackageInfo attribute) (swh.loader.package.pubdev.loader.PubDevPackageInfo attribute) (swh.model.model.Release attribute) (swh.model.model.Revision attribute) (swh.storage.cassandra.model.ReleaseRow attribute) (swh.storage.cassandra.model.RevisionRow attribute) (swh.web.alter.models.Event attribute) (swh.web.utils.typing.ReleaseMetadata attribute) (swh.web.utils.typing.RevisionMetadata attribute) author() (in module swh.loader.package.pypi.loader) author_date (swh.loader.package.deposit.loader.DepositPackageInfo attribute) author_to_db() (in module swh.storage.postgresql.converters) author_url (swh.web.utils.typing.ReleaseMetadata attribute) (swh.web.utils.typing.RevisionMetadata attribute) authority (swh.model.model.RawExtrinsicMetadata attribute) authority_type (swh.storage.cassandra.model.RawExtrinsicMetadataByIdRow attribute) (swh.storage.cassandra.model.RawExtrinsicMetadataRow attribute) authority_url (swh.storage.cassandra.model.RawExtrinsicMetadataByIdRow attribute) (swh.storage.cassandra.model.RawExtrinsicMetadataRow attribute) authorization_code() (swh.auth.keycloak.KeycloakOpenIDConnect method) authorization_url() (swh.auth.keycloak.KeycloakOpenIDConnect method) available (swh.web.alter.models.Origin attribute) aware_datetimes() (in module swh.model.hypothesis_strategies) AwsCloudObjStorage (class in swh.objstorage.backends.libcloud) AzureCloudObjStorage (class in swh.objstorage.backends.azure) B b32digest() (swh.core.nar.Nar method) b64digest() (swh.core.nar.Nar method) b64e() (in module swh.export.test.test_edges) BACKEND swh-scrubber-check-init command line option backend_class (swh.core.api.RPCClient attribute) (swh.counters.api.client.RemoteCounters attribute) (swh.indexer.storage.api.client.RemoteStorage attribute) (swh.objstorage.api.client.RemoteObjStorage attribute) (swh.provenance.api.client.RemoteProvenance attribute) (swh.scheduler.api.client.RemoteScheduler attribute) (swh.search.api.client.RemoteSearch attribute) (swh.storage.api.client.RemoteStorage attribute) (swh.vault.api.client.RemoteVaultClient attribute) backend_id (swh.scheduler.model.TaskRun attribute) (swh.scheduler.simulator.common.Task attribute) backend_name (swh.scheduler.model.TaskType attribute) backoff_factor (swh.scheduler.model.TaskType attribute) BACKOFF_SPLAY (in module swh.scheduler.celery_backend.recurrent_visits) backslashescape_errors() (in module swh.core.utils) backup_swhids() (swh.alter.recovery_bundle.RecoveryBundleCreator method) BadInputExc BadLogLevel BadPathException BandwidthCalculator (class in swh.objstorage.backends.winery.throttler) base_cgroup() (in module swh.graph.shell) base_directory (swh.objstorage.backends.winery.settings.DirectoryShardsPool attribute) base_fields (swh.coarnotify.server.forms.ActorChangeForm attribute) (swh.coarnotify.server.forms.ActorCreationForm attribute) (swh.web.add_forge_now.api_views.AddForgeNowRequestForm attribute) (swh.web.add_forge_now.api_views.AddForgeNowRequestHistoryForm attribute) (swh.web.alter.forms.AlterationAccessForm attribute) (swh.web.alter.forms.AlterationAdminForm attribute) (swh.web.alter.forms.AlterationForm attribute) (swh.web.alter.forms.AlterationSearchForm attribute) (swh.web.alter.forms.CategoryForm attribute) (swh.web.alter.forms.ConfirmationForm attribute) (swh.web.alter.forms.EmailVerificationForm attribute) (swh.web.alter.forms.EventAdminForm attribute) (swh.web.alter.forms.MessageAdminForm attribute) (swh.web.alter.forms.MessageForm attribute) (swh.web.alter.forms.OriginAdminForm attribute) (swh.web.alter.forms.OriginSearchForm attribute) (swh.web.alter.forms.OriginSelectForm attribute) (swh.web.alter.forms.ReasonsForm attribute) (swh.web.inbound_email.views.InboundEmailForm attribute) BASE_REPO_URL (swh.lister.conda.lister.CondaLister attribute) BASE_URL (swh.lister.arch.lister.ArchLister attribute) (swh.lister.aur.lister.AurLister attribute) (swh.lister.crates.lister.CratesLister attribute) (swh.lister.dlang.lister.DlangLister attribute) (swh.lister.elm.lister.ElmLister attribute) (swh.lister.hackage.lister.HackageLister attribute) (swh.lister.pubdev.lister.PubDevLister attribute) (swh.lister.puppet.lister.PuppetLister attribute) base_url (swh.lister.rpm.lister.RPMSourceData attribute) (swh.loader.package.maven.loader.ArtifactDict attribute) (swh.loader.package.maven.loader.MavenPackageInfo attribute) BaseApiDepositClient (class in swh.deposit.client) BaseChecker (class in swh.scrubber.base_checker) BaseConnection (class in swh.graphql.resolvers.base_connection) BaseContent (class in swh.model.model) BaseContentNode (class in swh.graphql.resolvers.content) BaseCreateDepositClient (class in swh.deposit.client) BaseDb (class in swh.core.db) BaseDepositClient (class in swh.deposit.client) BaseDirectoryEntryNode (class in swh.graphql.resolvers.directory_entry) BaseDirectoryLoader (class in swh.loader.core.loader) BaseDirectoryNode (class in swh.graphql.resolvers.directory) BaseDiscoveryGraph (class in swh.model.discovery) BaseExtrinsicMapping (class in swh.indexer.metadata_dictionary.base) BaseGitLoader (class in swh.loader.git.base) BaseHashableModel (class in swh.model.model) BaseIndexer (class in swh.indexer.indexer) BaseIntrinsicMapping (class in swh.indexer.metadata_dictionary.base) BaseList (class in swh.graphql.resolvers.base_connection) BaseLoader (class in swh.loader.core.loader) BaseMapping (class in swh.indexer.metadata_dictionary.base) BaseMetadataFetcher (class in swh.loader.metadata.base) BaseModel (class in swh.model.model) (class in swh.web.alter.models) BaseModel.Meta (class in swh.web.alter.models) basename_sortkey() (in module swh.core.utils) BaseNode (class in swh.graphql.resolvers.base_node) BaseOriginNode (class in swh.graphql.resolvers.origin) BaseOutput (class in swh.scanner.output) BasePackageInfo (class in swh.loader.package.loader) BasePartitionChecker (class in swh.scrubber.base_checker) BaseReleaseNode (class in swh.graphql.resolvers.release) BaseRevisionNode (class in swh.graphql.resolvers.revision) BaseRow (class in swh.indexer.storage.model) (class in swh.storage.cassandra.model) BaseSchedulerModel (class in swh.scheduler.model) BaseSnapshotBranchNode (class in swh.graphql.resolvers.snapshot_branch) BaseSnapshotNode (class in swh.graphql.resolvers.snapshot) BaseSWHNode (class in swh.graphql.resolvers.base_node) BaseTargetNode (class in swh.graphql.resolvers.target) BaseVaultCooker (class in swh.vault.cookers.base) BaseVisitNode (class in swh.graphql.resolvers.visit) BaseVisitStatusNode (class in swh.graphql.resolvers.visit_status) batch_cook() (swh.vault.api.client.RemoteVaultClient method) (swh.vault.backend.VaultBackend method) (swh.vault.in_memory_backend.InMemoryVaultBackend method) (swh.vault.interface.VaultInterface method) batch_progress() (swh.vault.api.client.RemoteVaultClient method) (swh.vault.backend.VaultBackend method) (swh.vault.in_memory_backend.InMemoryVaultBackend method) (swh.vault.interface.VaultInterface method) batch_size (swh.graph.luigi.compressed_graph.CompressGraph attribute) BAZAAR (swh.lister.sourceforge.lister.VcsNames attribute) (swh.model.model.RevisionType attribute) BazaarLoader (class in swh.loader.bzr.loader) BearerTokenAuthBackend (class in swh.auth.starlette.backends) BECH32 (swh.alter.bech32.Encoding attribute) bech32_create_checksum() (in module swh.alter.bech32) bech32_decode() (in module swh.alter.bech32) bech32_encode() (in module swh.alter.bech32) bech32_hrp_expand() (in module swh.alter.bech32) bech32_polymod() (in module swh.alter.bech32) bech32_verify_checksum() (in module swh.alter.bech32) BECH32M (swh.alter.bech32.Encoding attribute) BEFORE swh-storage-remove-old-object-reference-partitions command line option best_mimetype() (swh.core.api.negotiation.Negotiator method) (swh.core.api.Negotiator method) Bfs (class in swh.graph.luigi.compressed_graph) BFS (swh.graph.webgraph.CompressionStep attribute) BFS_DCF (swh.graph.webgraph.CompressionStep attribute) BFS_EF (swh.graph.webgraph.CompressionStep attribute) BFS_ROOTS (swh.graph.webgraph.CompressionStep attribute) BfsDcf (class in swh.graph.luigi.compressed_graph) BfsEf (class in swh.graph.luigi.compressed_graph) BFSRevisionsWalker (class in swh.storage.algos.revisions_walker) BfsRoots (class in swh.graph.luigi.compressed_graph) BibTeXCitationError BibTeXWithMacroWriter (class in swh.indexer.bibtex) binaries_without_bytes() (in module swh.model.hypothesis_strategies) binary_string_base64_resolver() (in module swh.graphql.resolvers.resolvers) binary_string_text_resolver() (in module swh.graphql.resolvers.resolvers) binhash() (in module swh.export.test.test_edges) BIOCONDUCTOR_HOMEPAGE (swh.lister.bioconductor.lister.BioconductorLister attribute) BioconductorLister (class in swh.lister.bioconductor.lister) BioconductorListerState (class in swh.lister.bioconductor.lister) BioconductorLoader (class in swh.loader.package.bioconductor.loader) BioconductorPackageInfo (class in swh.loader.package.bioconductor.loader) BitbucketLister (class in swh.lister.bitbucket.lister) BitbucketListerState (class in swh.lister.bitbucket.lister) BitbucketOriginSaveWebhookReceiver (class in swh.web.save_origin_webhooks.bitbucket) blake2s256 (swh.model.hashutil.HashDict attribute) (swh.model.hashutil.TotalHashDict attribute) (swh.model.model.Content attribute) (swh.model.model.SkippedContent attribute) (swh.s
2026-01-13T09:29:19
https://git-scm.com/book/fr/v2/Git-sur-le-serveur-HTTP-intelligent
Git - HTTP intelligent About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Démarrage rapide 1.1 À propos de la gestion de version 1.2 Une rapide histoire de Git 1.3 Rudiments de Git 1.4 La ligne de commande 1.5 Installation de Git 1.6 Paramétrage à la première utilisation de Git 1.7 Obtenir de l’aide 1.8 Résumé 2. Les bases de Git 2.1 Démarrer un dépôt Git 2.2 Enregistrer des modifications dans le dépôt 2.3 Visualiser l’historique des validations 2.4 Annuler des actions 2.5 Travailler avec des dépôts distants 2.6 Étiquetage 2.7 Les alias Git 2.8 Résumé 3. Les branches avec Git 3.1 Les branches en bref 3.2 Branches et fusions : les bases 3.3 Gestion des branches 3.4 Travailler avec les branches 3.5 Branches de suivi à distance 3.6 Rebaser (Rebasing) 3.7 Résumé 4. Git sur le serveur 4.1 Protocoles 4.2 Installation de Git sur un serveur 4.3 Génération des clés publiques SSH 4.4 Mise en place du serveur 4.5 Démon (Daemon) Git 4.6 HTTP intelligent 4.7 GitWeb 4.8 GitLab 4.9 Git hébergé 4.10 Résumé 5. Git distribué 5.1 Développements distribués 5.2 Contribution à un projet 5.3 Maintenance d’un projet 5.4 Résumé 6. GitHub 6.1 Configuration et paramétrage d’un compte 6.2 Contribution à un projet 6.3 Maintenance d’un projet 6.4 Gestion d’un regroupement 6.5 Écriture de scripts pour GitHub 6.6 Résumé 7. Utilitaires Git 7.1 Sélection des versions 7.2 Indexation interactive 7.3 Remisage et nettoyage 7.4 Signer votre travail 7.5 Recherche 7.6 Réécrire l’historique 7.7 Reset démystifié 7.8 Fusion avancée 7.9 Rerere 7.10 Déboguer avec Git 7.11 Sous-modules 7.12 Empaquetage (bundling) 7.13 Replace 7.14 Stockage des identifiants 7.15 Résumé 8. Personnalisation de Git 8.1 Configuration de Git 8.2 Attributs Git 8.3 Crochets Git 8.4 Exemple de politique gérée par Git 8.5 Résumé 9. Git et les autres systèmes 9.1 Git comme client 9.2 Migration vers Git 9.3 Résumé 10. Les tripes de Git 10.1 Plomberie et porcelaine 10.2 Les objets de Git 10.3 Références Git 10.4 Fichiers groupés 10.5 La refspec 10.6 Les protocoles de transfert 10.7 Maintenance et récupération de données 10.8 Les variables d’environnement 10.9 Résumé A1. Annexe A: Git dans d’autres environnements A1.1 Interfaces graphiques A1.2 Git dans Visual Studio A1.3 Git dans Visual Studio Code A1.4 Git dans IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git dans Sublime Text A1.6 Git dans Bash A1.7 Git dans Zsh A1.8 Git dans PowerShell A1.9 Résumé A2. Annexe B: Embarquer Git dans vos applications A2.1 Git en ligne de commande A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Commandes Git A3.1 Installation et configuration A3.2 Obtention et création des projets A3.3 Capture d’instantané basique A3.4 Création de branches et fusion A3.5 Partage et mise à jour de projets A3.6 Inspection et comparaison A3.7 Débogage A3.8 Patchs A3.9 Courriel A3.10 Systèmes externes A3.11 Administration A3.12 Commandes de plomberie 2nd Edition 4.6 Git sur le serveur - HTTP intelligent HTTP intelligent Nous avons à présent un accès authentifié par SSH et un accès non authentifié par git:// , mais il existe aussi un protocole qui peut faire les deux à la fois. La configuration d’un HTTP intelligent revient simplement à activer sur le serveur un script CGI livré avec Git qui s’appelle git-http-backend . Ce CGI va lire le chemin et les entêtes envoyés par un git fetch ou un git push à une URL donnée et déterminer si le client peut communiquer sur HTTP (ce qui est vrai pour tout client depuis la version 1.6.6). Si le CGI détecte que le client est intelligent, il va commencer à communiquer par protocole intelligent, sinon il repassera au comportement du protocole idiot (ce qui le rend de ce fait compatible avec les vieux clients). Détaillons une installation de base. Nous la réaliserons sur un serveur web Apache comme serveur CGI. Si Apache n’est pas installé sur votre PC, vous pouvez y remédier avec une commande : $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env Cela a aussi pour effet d’activer les modules mod_cgi , mod_alias , et mod_env qui sont nécessaires au fonctionnement du serveur. Vous allez aussi avoir besoin de paramétrer le groupe Unix des répertoire /srv/git à www-data pour que votre serveur web puisse lire et écrire dans les dépôts, parce que l’instance Apache lançant le script CGI aura (par défaut) cet utilisateur : $ chgrp -R www-data /srv/git Ensuite, nous devons ajouter quelques lignes à la configuration d’Apache pour qu’il lance git-http-backend comme gestionnaire de tous les chemins du serveur web sous /git . SetEnv GIT_PROJECT_ROOT /srv/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/lib/git-core/git-http-backend/ Si vous ne définissez pas la variable d’environnement GIT_HTTP_EXPORT_ALL , Git ne servira aux utilisateurs non authentifiés que les dépôts comprenant le fichier git-daemon-export-ok , de la même manière que le daemon Git. Puis, nous allons indiquer à Apache qu’il doit accepter les requêtes sur ce chemin avec quelque chose comme : <Files "git-http-backend"> AuthType Basic AuthName "Git Access" AuthUserFile /srv/git/.htpasswd Require expr !(%{QUERY_STRING} -strmatch '*service=git-receive-pack*' || %{REQUEST_URI} =~ m#/git-receive-pack$#) Require valid-user </Files> Il vous sera nécessaire de créer un fichier .htpasswd contenant les mots de passe de tous les utilisateurs valides. Voici un exemple d’ajout d’un utilisateur schacon au fichier : $ htpasswd -c /srv/git/.htpasswd schacon Il existe des milliers de façons d’authentifier des utilisateurs avec Apache, il suffira d’en choisir une et de la mettre en place. L’exemple ci-dessus n’est que le plus simple. Vous désirerez sûrement gérer tout ceci sous SSL pour que vos données soient chiffrées. Nous ne souhaitons pas nous appesantir spécifiquement sur la configuration d’Apache, car on peut utiliser un serveur différent ou avoir besoin d’une authentification différente. L’idée générale reste que Git est livré avec un CGI appelé git-http-backend qui, après authentification, va gérer toute la négociation pour envoyer et recevoir les données sur HTTP. Il ne gère pas l’authentification par lui-même, mais peut être facilement contrôlé à la couche serveur web qui l’invoque. Cela peut être réalisé avec n’importe quel serveur web gérant le CGI, donc celui que vous connaissez le mieux. Note Pour plus d’informations sur la configuration de l’authentification dans Apache, référez-vous à la documentation d’Apache : https://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
http://www.trello.com/butler-automation
Trello Automation: Automate your workflow with Trello | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Automate your workflow with Trello Powerful no-code automation is built into every Trello board. With Automation, you can focus on the work that matters most. Start automating today — It’s free! Learn more about Trello’s plans and pricing. Create rules, buttons, and commands to automate almost any action in Trello. Rule your boards Setting rules means important tasks won’t fall through the cracks. Just set a trigger and the actions to be performed, and let Automation run the show. Doesn’t that rule? + How to try it When a card is created in a "To Do" list, add the "Steps" checklist so that you don’t miss a beat. When you enter a card name ending with a date, set the card to be due on that date. Perfect for when emailing cards to your board. When a card is marked as “Done,” automatically mark the due date as complete, cross off checklist items, and remove members. Move work forward Get to the next step faster with custom card and board buttons. Card buttons appear on the back of every card, and board buttons are in the top right corner. They both perform a series of actions in a single click! + How to try it Create a card button named “Done!” that moves the card to the “Done” list, and marks the date complete. Add a “Start” card button that when pressed moves the card to list "In Progress,” and adds you to the card with a due date. Create a board button that sorts cards by Custom Field story points, due date, and votes, so you know what to prioritize next sprint. Integrate your favorite apps The power of Trello’s automation can be extended beyond the boundary of your boards with integrations for Slack, Jira, and email. Trello can post messages to Slack channels, create new Jira tickets, post comments to existing tickets, and send automated emails right from your boards. + Learn more Add integrations to a “Ready for development” card button to: Notify the #development Slack channel Create a new ticket for the task in Jira And send an email to any project stakeholders with an update. Intelligent automation tips As you use a Trello board, Automation will recognize repetitive actions being performed and suggest automations based on those actions that can be enabled in a single click. Companies of all shapes and sizes use Trello. Check out some of our customer stories here . Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://sre.google/workbook/foreword-II/
Google SRE - learn everything about SRE DevOps engineering Foreword II Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Foreword II Andrew Clay Shafer When I found out people were working on a second SRE book, I reached out and asked if I could write a few words. The principles from the first SRE book align so well with what I always imagined DevOps to be, and the practices are insightful, even when they aren’t 100% applicable outside of Google. After reading the principles from the first SRE book for the first time— embracing risk (Chapter 3) , service level objectives (Chapter 4) , and eliminating toil (Chapter 5) —I wanted to shout that message from the rooftops. “Embracing risk” resonated so much because I had used similar language many times to help traditional organizations motivate change. Chapter 6 was always an implicit DevOps goal, both to allow humans more time for creative higher-order work and to allow them to be more human. But I really fell in love with “service level objectives.” I love that the language and the process create a dispassionate contract between operational considerations and delivering new functionality. The SRE, SWE (software engineer), and business all agree that the service has to be up to be valuable, and the SRE solution quantifies objectives to drive actions and priorities. The solution—make the service level a target, and when you are below the target prioritize reliability over features—eliminates a classic conflict between operations and developers. This is a simple and elegant reframing that solves problems by not having them. I give these three chapters as a homework assignment to almost everyone I’ve met since. They are that good. Everyone should know. Tell all your friends. I’ve told all mine. The last decade of my career has been focused on helping people deliver software with better tools and process. Sometimes people say I contributed to inventing DevOps, but I was just in the position to borrow and steal successful patterns from across many different organizations and projects. I get embarrassed when people say “DevOps” was invented by anyone, but especially by me. I don’t consider myself an expert in anything but being inquisitive. My idealized DevOps always patterned off whatever information I could extract or infer from my friends, and my friends happened to be building the internet. I had the privilege of behind-the-scenes access to people deploying and operating a representative sample of the world’s most incredible infrastructures and applications. DevOps symbolizes aspects of the emergent and existential optimizations required to rapidly deliver highly available software over the internet. The shift from software delivered on physical media to software delivered as a service forced an evolution of tools and processes. This evolution elevated operations’ contribution to the value chain. If the systems are down, the software has no value. The good news is, you don’t have to wait for shipping the next shrink-wrapped box to change the software. For some, this is also the bad news. I simply had the opportunity and perspective to articulate the most successful patterns of the new way to a receptive audience. In 2008, before we used the word DevOps like we do now, I’d been through the dot-com collapse, grad school, and a couple of venture-funded rollercoaster rides as a developer—searching Google for answers daily the whole time. I was working on Puppet full-time and I was fascinated by the potential for automation to transform IT organizations. Puppet thrust me into solving problems in the operations domain. At this time, Google used Puppet to manage their corporate Linux and OS X workstations at a scale that pushed the capabilities of the Puppet server. We had a great working relationship with Google, but Google kept certain details of their internal operations secret as a matter of policy. I know this because I’m naturally curious and was constantly seeking more information. I always knew Google must have great internal tools and processes, but what these tools and processes were wasn’t always apparent. Eventually, I accepted that asking deep questions about Borg probably meant the current conversation wasn’t going very far. I would have loved to know more about how Google did everything, but this simply wasn’t allowed at the time. The significance of 2008 also includes the first O’Reilly Velocity conference and the year I met Patrick Debois. “DevOps” wasn’t a thing yet, but it was about to be. The time was right. The world was ready. DevOps symbolized a new way, a better way. If Site Reliability Engineering had been published then, I believe the community that formed would have rallied to fly the “eliminate toil” flag and the term DevOps might have never existed. Counterfactuals notwithstanding, I know the first SRE book personally advanced my understanding of the possible, and I already helped many others just with the SRE principles. In the early days of the DevOps movement, we consciously avoided codifying practices because everything was evolving so rapidly and we didn’t want to set limits on what DevOps could become. Plus, we explicitly didn’t want anyone to “own” DevOps. When I wrote about DevOps in 2010, I made three distinct points. First, developers and operations can and should work together. Second, system administration will become more and more like software development. Finally, sharing with a global community of practice accelerates and multiplies our collective capabilities. Around the same time, my friends Damon Edward and John Willis coined the acronym CAMS for Culture, Automation, Metrics, and Sharing. Jez Humble later expanded this acronym to CALMS by adding Lean continuous improvement. What each of these words might mean in context deserves to be a full book, but I mention them here because Site Reliability Engineering explicitly references Culture, Automation, Metrics, and Sharing alongside anecdotes about Google’s journey to continuously improve. By publishing the first SRE book, Google shared their principles and practices with the global community. Now I define DevOps simply as “optimizing human performance and experience operating software, with software, and with humans.” I don’t want to put words in anyone’s mouth, but that seems like a great way to describe SRE as well. Ultimately, I know DevOps when I see it and I see SRE at Google, in theory and practice, as one of the most advanced implementations. Good IT operations has always depended on good engineering, and solving operations problems with software has always been central to DevOps. Site Reliability Engineering makes the engineering aspect even more explicit. I cringe when I hear someone say “SRE versus DevOps.” For me, they are inseparable in time and space, as labels describing the sociotechnical systems that deliver modern infrastructure with software. I consider DevOps a loose generic set of principles and SRE an advanced explicit implementation. A parallel analogy would be the relationship between Agile and Extreme Programming (XP). True, XP is Agile, arguably the best of Agile, but not all organizations are capable of or willing to adopt XP. Some say “software is eating the world,” and I understand why they do, but “software” alone is not the right framing. Without the ubiquity of computational hardware connected with high-speed networks, much of what we take for granted as “software” would not be possible. This is an undeniable truth. What I think many miss in this conversation about technology are the humans. Technology exists because of humans and hopefully for humans, but if you look a little deeper, you also realize that the software we rely on, and probably take for granted, is largely dependent on humans. We rely on software, but software also relies on us. This is a single interconnected system of imperfect hardware—software and humans relying on themselves to build the future. Reliability is eating the world. Reliability is not just about technology, though, but also about people. The people and the technology form a single technosocial system. One nice feature about having Google share SRE with the rest of the industry is that any excuses about what kind of processes work at scale became invalid. Google set the highest standard for both reliability and scale. There might be valid arguments about why someone can’t adopt Google SRE practices directly, but the guiding principles should still apply. As I look at the landscape of possibilities to build the future and the ambition to transform human experience with software, I see a lot of ambitious projects to quite literally connect everything to the internet. My math says that the successful projects will find themselves ingesting and indexing incredible amounts of data. Few, if any, will surpass the scale of Google today, but some will be the same size Google was when they started SRE and will need to solve the same reliability problems. I contend that in these cases, adopting tools and process that look suspiciously like SRE is not optional but existential—though there is no need to wait for that crisis because SRE principles and practices apply at any scale. SRE is usually framed as how Google does operations, but that misses the bigger picture: SRE in practice enables software engineering, but also transforms architecture, security, governance, and compliance. When we leverage the SRE focus on providing a platform of services, all these other considerations get to have first-class emphasis, but where and how that happens may be quite different. Just like SRE (and hopefully DevOps) shifted more and more of the burden to software engineering, modern architecture and security practices evolve from slides, checklists, and hope to enabling the right behaviors with running code. Organizations adopting SRE principles and practices without revisiting these other aspects lose a huge opportunity to improve, and will also probably meet with internal resistance if the people who consider themselves responsible for those aspects are not converted into allies. I always enjoy learning. I read every word of the first SRE book straight through. I loved the language. I loved the anecdotes. I loved understanding more about how Google sees itself. But the question for me is always, “What behavior will I change?” Learning isn’t collecting information. Learning is changing behavior. This is easy to determine or even quantify in certain disciplines. You have learned to play a new song when you can play the song. You are better at chess when you win games against stronger players. Site Reliability Engineering, like DevOps, should not just be changing titles, but making definitive behavior changes, focusing on outcomes and obviously reliability. The Site Reliability Workbook promises to move forward from an enumeration of principles and practices by Google for Google toward more contextual actions and behaviors. Site reliability is for everyone, but reliability doesn’t come from reading books. Here’s to embracing risk and eliminating toil. Previous Foreword I Next Preface Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://sre.google/workbook/slo-engineering-case-studies/
Google SRE - SLO Implementation: Evernote and Home Depot Chapter 3 - SLO Engineering Case Studies Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon SLO Engineering Case Studies By Ben McCormack (Evernote) and William Bonnell (The Home Depot) with Garrett Plasky (Evernote), Alex Hidalgo, Betsy Beyer, and Dave Rensin While many tenets of SRE were shaped within the walls of Google, its principles have long lived outside our gates. Many standard Google SRE practices have been discovered in parallel or otherwise been adopted by many other organizations across the industry. SLOs are fundamental to the SRE model. Since we launched the Customer Reliability Engineering (CRE) team—a group of experienced SREs who help Google Cloud Platform (GCP) customers build more reliable services—almost every customer interaction starts and ends with SLOs. Here we present two stories, told by two very different companies, that outline their journeys toward adopting an SLO and error budget–based approach while working with the Google CRE team. For a more general discussion about SLOs and error budgets, see Implementing SLOs in this book, and Chapter 3 in our first book. Evernote’s SLO Story by Ben McCormack, Evernote Evernote is a cross-platform app that helps individuals and teams create, assemble, and share information. With more than 220 million users worldwide, we store over 12 billion pieces of information—a mix of text-based notes, files, and attachments/images—within the platform. Behind the scenes, the Evernote service is supported by 750+ MySQL instances. We introduced the concept of SLOs to Evernote as part of a much wider technology revamp aimed at increasing engineering velocity while maintaining quality of service. Our goals included: Move engineering focus away from undifferentiated heavy lifting in datacenters and toward product engineering work that customers actually cared about. To that end, we stopped running our physical datacenters and moved to a public cloud. Revise the working model of operations and software engineers to support an increase in feature velocity while maintaining overall quality of service. Revamp how we look at SLAs to ensure that we increase focus on how failures impact our large global customer base. These goals may look familiar to organizations across many industries. While no single approach to making these types of changes will work across the board, we hope that sharing our experience will provide valuable insights for others facing similar challenges. Why Did Evernote Adopt the SRE Model? At the outset of this transition, Evernote was characterized by a traditional ops/dev split: an operations team protected the sanctity of the production environment, while a development team was tasked with developing new product features for customers. These objectives were usually in conflict: the dev team felt constrained by lengthy operational requirements, while the ops team became frustrated when new code introduced new issues in production. As we swung wildly between these two goals, the ops and dev teams developed a frustrated and strained relationship. We wanted to reach a happier medium that better balanced the varying needs of the teams involved. We attempted to address the gaps in this traditional dichotomy in various ways over the course of five-plus years. After trying out a “You wrote it, you run it” (development) model, and a “You wrote it, we run it for you” (operations) model, we moved toward an SLO-centric SRE approach. So what motivated Evernote to move in this direction? At Evernote, we view the core disciplines of operations and development as separate professional tracks in which engineers can specialize. One track is concerned with the nearly 24/7 ongoing delivery of a service to customers. The other is concerned with the extension and evolution of that service to meet customer needs in the future. These two disciplines have moved toward each other in recent years as movements like SRE and DevOps emphasize software development as applied to operations. (This convergence has been furthered by advances in datacenter automation and the growth of public clouds, both of which give us a datacenter that can be fully controlled by software.) On the other side of the spectrum, full-stack ownership and continuous deployment are increasingly applied to software development. We were drawn to the SRE model because it fully embraces and accepts the differences between operations and development while encouraging teams to work toward a common goal. It does not try to transform operations engineers into application developers, or vice versa. Instead, it gives both a common frame of reference. In our experience, an error budget/SLO approach has led both teams to make similar decisions when presented with the same facts, as it removes a good deal of subjectivity from the conversation. Introduction of SLOs: A Journey in Progress The first part of our journey was the move from physical datacenters to Google Cloud Platform. 1 Once the Evernote service was up and running on GCP and stabilized, we introduced SLOs. Our objectives here were twofold: Align teams internally around Evernote SLOs, ensuring that all teams were working within the new framework. Incorporate Evernote’s SLO into how we work with the Google Cloud team, who now had responsibility for our underlying infrastructure. Since we now had a new partner within the overall model, we needed to ensure the move to GCP did not dilute or mask our commitment to our users. After actively using SLOs for about nine months, Evernote is already on version 3 of its SLO practice! Before getting into the technical details of an SLO, it is important to start the conversation from your customers’ point of view: what promises are you trying to uphold? Similar to most services, Evernote has many features and options that our users put to use in a variety of creative ways. We wanted to ensure we initially focused on the most important and common customer need: the availability of the Evernote service for users to access and sync their content across multiple clients . Our SLO journey started from that goal. We kept our first pass simple by focusing on uptime. Using this simple first approach, we could clearly articulate what we were measuring, and how. Our first SLOs document contained the following: A definition of the SLOs This was an uptime measure: 99.95% uptime measured over a monthly window, set for certain services and methods. We chose this number based upon discussions with our internal customer support and product teams and—more importantly—user feedback. We deliberately chose to bind our SLOs to a calendar month versus a rolling period to keep us focused and organized when running service reviews. What to measure, and how to measure it What to measure We specified a service endpoint we could call to test whether the service was functioning as expected. In our case, we have a status page built into our service that exercises most of our stack and returns a 200 status code if all is well. How to measure We wanted a prober that called the status page periodically. We wanted that prober to be located completely outside of and independent from our environment so we could test all our components, including our load balancing stack. Our goal here was to make sure that we were measuring any and all failures of both the GCP service and the Evernote application. However, we did not want random internet issues to trigger false positives. We chose to use a third-party company that specializes in building and running such probers. We selected Pingdom , but there are many others in the market. We conduct our measurements as follows: Frequency of probe: We poll our frontend nodes every minute. Location of probers: This setting is configurable; we currently use multiple probes in North America and Europe. Definition of “down”: If a prober check fails, the node is marked as Unconfirmed Down and then a second geographically separate prober performs a check. If the second check fails, the node is marked down for SLO calculation purposes. The node will continue to be marked as down as long as consecutive probe requests register errors. How to calculate SLOs from monitoring data Finally, we carefully documented how we calculate the SLO from the raw data we received from Pingdom. For example, we specified how to account for maintenance windows: we could not assume that all of our hundreds of millions of users knew about our published maintenance windows. Uninformed users would therefore experience these windows as generic and unexplained downtime, so our SLO calculations treated maintenance as downtime. Once we defined our SLOs, we had to do something with them. We wanted the SLOs to drive software and operations changes that would make our customers happier and keep them happy. How best to do this? We use the SLO/error budget concept as a method to allocate resources going forward. For example, if we missed the SLO for last month, that behavior helps us prioritize relevant fixes, improvements, and bug fixes. We keep it simple: teams from both Evernote and Google conduct a monthly review of SLO performance. At this meeting, we review the SLO performance from the previous month and perform a deep dive on any outages. Based on this analysis of the past month, we set action items for improvements that may not have been captured through the regular root-cause-analysis process. Throughout this process, our guiding principle has been “Perfect is the enemy of good.” Even when SLOs aren’t perfect, they’re good enough to guide improvements over time. A “perfect” SLO would be one that measures every possible user interaction with our service and accounts for all edge cases. While this is a great idea on paper, it would take many months to achieve (if achieving perfection were even possible)—time which we could use to improve the service. Instead, we selected an initial SLO that covered most, but not all, user interactions, which was a good proxy for quality of service. Since we began, we have revised our SLOs twice, according to signals from both our internal service reviews and in response to customer-impacting outages. Because we weren’t aiming for perfect SLOs at the outset, we were comfortable with making changes to better align with the business. In addition to our monthly Evernote/Google review of SLO performance, we’ve settled on a six-month SLO review cycle, which strikes the right balance between changing SLOs too often and letting them become stale. In revising our SLOs, we’ve also learned that it's important to balance what you would like to measure with what’s possible to measure. Since introducing SLOs, the relationship between our operations and development teams has subtly but markedly improved. The teams now have a common measure of success: removing the human interpretation of quality of service (QoS) has allowed both teams to maintain the same view and standards. To provide just one example, SLOs provided a common ground when we had to facilitate multiple releases in a compressed timeline in 2017. While we chased down a complex bug, product development requested that we apportion our normal weekly release over multiple separate windows, each of which would potentially impact customers. By applying an SLO calculation to the problem and removing human subjectivity from the scenario, we were able to better quantify customer impact and reduce our release windows from five to two to minimize customer pain. Breaking Down the SLO Wall Between Customer and Cloud Provider A virtual wall between customer and cloud provider concerns might seem natural or inevitable. While Google has SLOs and SLAs (service level agreements) for the GCP platforms we run Evernote on, Evernote has its own SLOs and SLAs. It’s not always expected that two such engineering teams would be informed about each other’s SLAs. Evernote never wanted such a wall. Of course, we could have engineered to a dividing wall, basing our SLOs and SLAs on the underlying GCP metrics. Instead, from the beginning, we wanted Google to understand which performance characteristics were most important to us, and why. We wanted to align Google’s objectives with ours, and for both companies to view Evernote’s reliability successes and failures as shared responsibilities. To achieve this, we needed a way to: Align objectives Ensure our partner (in this case, Google) really understood what’s important to us Share both successes and failures Most service providers manage to the published SLO/SLAs for their cloud services. Working within this context is important, but it can’t holistically represent how well our service is running within the cloud provider’s environment. For example, a given cloud provider probably runs hundreds of thousands of virtual machines globally, which they manage for uptime and availability. GCP promises 99.95% availability for Compute Engine (i.e., its virtual machines). Even when GCP SLO graphs are green (i.e., above 99.95%), Evernote’s view of the same SLO might be very different: because our virtual machine footprint is only a small percentage of the global GCP number, outages isolated to our region (or isolated for other reasons) may be “lost” in the overall rollup to a global level. To correct for scenarios like this, we share our SLO and real-time performance against SLO with Google. As a result, both the Google CRE team and Evernote work with same performance dashboards. This may seem like a very simple point, but has proven a rather powerful way to drive truly customer-focused behavior. As a result, rather than receiving generic “Service X is running slow”–type notifications, Google provides us with notifications that are more specific to our environment. For example, in addition to a generic “GCP load balancing environment is running slow today,” we’ll also be informed that this issue is causing a 5% impact to Evernote’s SLO. This relationship also helps teams within Google, who can see how their actions and decisions impact customers. This two-way relationship has also given us a very effective framework to support major incidents. Most of the time, the usual model of P1–P5 tickets and regular support channels works well and allows us to maintain good service and a good relationship with Google. But we all know there are times when a P1 ticket (“major impact to our business”) is not enough—the times when your whole service is on the line and you face extended business impact. At times like these, our shared SLOs and relationship with the CRE team come to fruition. We have a common understanding that if the SLO impact is high enough, both parties will treat the issue as a P1 ticket with special handling. Quite often, this means that Evernote and Google’s CRE Team rapidly mobilize on a shared conference bridge. The Google CRE team monitors the SLO we jointly defined and agreed upon, allowing us to remain in sync in terms of prioritization and appropriate responses. Current State After actively using SLOs for about nine months, Evernote is already on version 3 of its SLO practice. The next version of SLOs will progress from our simple uptime SLO. We plan to start probing individual API calls and accounting for the in-client view of metrics/performance so we can represent user QoS even better. By providing a standard and defined way of measuring QoS, SLOs have allowed Evernote to better focus on how our service is running. We can now have data-driven conversations—both internally, and with Google—about the impact of outages, which enables us to drive service improvements, ultimately making for more effective support teams and happier customers. The Home Depot’s SLO Story by William Bonnell, The Home Depot The Home Depot (THD) is the world’s largest home improvement retailer: we have more than 2,200 stores across North America, each filled with more than 35,000 unique products (and supplemented with over 1.5 million products online). Our infrastructure hosts a variety of software applications that support nearly 400,000 associates and process more than 1.5 billion customer transactions per year. The stores are deeply integrated with a global supply chain and an ecommerce website that receives more than 2 billion visits per year. In a recent refresh to our operations approach aimed at increasing the velocity and quality of our software development, THD both pivoted to Agile software development and changed how we design and manage our software. We moved from centralized support teams that supported large, monolithic software packages to a microservices architecture led by small, independently operated software development teams. As a result, our system now had smaller chunks of constantly changing software, which also needed to be integrated across the stack. Our move to microservices was complemented by a move to a new “freedom and responsibility culture” of full-stack ownership. This approach gives developers freedom to push code when they want, but also makes them jointly responsible for the operations of their service. For this model of joint ownership to work, operations and development teams need to speak a common language that promotes accountability and cuts across complexity: service level objectives. Services that depend upon each other need to know information like: How reliable is your service? Is it built for three 9s, three and a half 9s, or four 9s (or better)? Is there planned downtime? What kind of latency can I expect at the upper bounds? Can you handle the volume of requests I am going to send? How do you handle overload? Has your service achieved its SLOs over time? If every service could provide transparent and consistent answers to these questions, teams would have a clear view into their dependencies, which allows for better communication and increased trust and accountability between teams. The SLO Culture Project Before we began this shift in our service model, The Home Depot didn’t have a culture of SLOs. Monitoring tools and dashboards were plentiful, but were scattered everywhere and didn’t track data over time. We weren’t always able to pinpoint the service at the root of a given outage. Often, we began troubleshooting at the user-facing service and worked backward until we found the problem, wasting countless hours. If a service required planned downtime, its dependent services were surprised. If a team needed to build a three and a half 9s service, they wouldn’t know if the service they had a hard dependency on could support them with even better uptime (four 9s). These disconnects caused confusion and disappointment between our software development and operations teams. We needed to address these disconnects by building a common culture of SLOs. Doing so required an overarching strategy to influence people, process, and technology. Our efforts spanned four general areas: Common vernacular Define SLOs in the context of THD. Define how to measure them in a consistent way. Evangelism Spread the word across the company. Create training material to sell why SLOs matter, road shows across the company, internal blogs, and promotional materials like t-shirts and stickers. Enlist a few early adopters to implement SLOs and demonstrate the value to others. Establish a catchy acronym (VALET; as discussed later) to help the idea spread. Create a training program (FiRE Academy: Fundamentals in Reliability Engineering) to train developers on SLOs and other reliability concepts. 2 Automation To reduce the friction of adoption, implement a metric collection platform to automatically collect service level indicators for any service deployed to production. These SLIs can later be more easily turned into SLOs. Incentive Establish annual goals for all development managers to set and measure SLOs for their services. Establishing a common vernacular was critical to getting everybody on the same page. We also wanted to keep this framework as simple as possible to help the idea spread faster. To get started, we took a critical look at the metrics we monitored across our various services and discovered some patterns. Every service monitored some form of its traffic volume , latency , errors , and utilization —metrics that map closely to Google SRE’s Four Golden Signals . In addition, many services monitored uptime or availability distinctly from errors. Unfortunately, across the board, all categories of metrics were inconsistently monitored, were named differently, or had insufficient data. None of our services had SLOs. The closest metric our production systems had to a customer-facing SLO was support tickets. The primary (and often only) way we measured the reliability of the applications deployed to our stores was by tracking the number of support calls our internal support desk receives. Our First Set of SLOs We couldn’t create SLOs for every aspect of our systems that could be measured, so we had to decide which metrics or SLIs should also have SLOs. Availability and latency for API calls We decided that each microservice had to have availability and latency SLOs for its API calls that were called by other microservices. For example, the Cart microservice called the Inventory microservice. For those API calls, the Inventory microservice published SLOs that the Cart microservice (and other microservices that needed Inventory) could consult to determine if the Inventory microservice could meet its reliability requirements. Infrastructure utilization Teams at THD measure infrastructure utilization in different ways, but the most typical measurement is real-time infrastructure utilization at one minute granularity. We decided against setting utilization SLOs for a few reasons. To begin with, microservices aren’t overly concerned with this metric—your users don’t really care about utilization as long as you can handle the traffic volume, your microservice is up, it’s responding quickly, it’s not throwing errors, and you’re not in danger of running out of capacity. Additionally, our impending move to the cloud meant that utilization would be less of a concern, so cost planning would overshadow capacity planning. (We’d still need to monitor utilization and perform capacity planning, but we didn’t need to include it in our SLO framework.) Traffic volume Because THD didn’t already have a culture of capacity planning, we needed a mechanism for software and operations teams to communicate how much volume their service could handle. Traffic was easy to define as requests to a service, but we needed to decide if we should track average requests per second, peak requests per second, or the volume of requests over the reporting time period. We decided to track all three and let each service select the most appropriate metric. We debated whether or not to set an SLO for traffic volume because this metric is determined by user behavior, rather than internal factors that we can control. Ultimately, we decided that as a retailer we needed to size our service for peaks like Black Friday, so we set an SLO according to expected peak capacity. Latency We let each service define its SLO for latency and determine where to best measure it. Our only request was that a service should supplement our common white-box performance monitoring with black-box monitoring to catch issues caused by the network or other layers like caches and proxies that fail outside the microservice. We also decided that percentiles were more appropriate than arithmetic averages. At minimum, services needed to hit a 90th percentile target; user-facing services had a preferred target of 95th and/or 99th percentile. Errors Errors were somewhat complicated to account for. Since we were primarily dealing with web services, we had to standardize what constitutes an error and how to return errors. If a web service encountered an error, we naturally standardized on HTTP response codes: A service should not indicate an error in the body of a 2xx response; rather, it should throw either a 4xx or a 5xx. An error caused by a problem with the service (for example, out of memory) should throw a 5xx error. An error caused by the client (for example, sending a malformed request) should throw a 4xx error. After much deliberation, we decided to track both 4xx and 5xx errors, but used 5xx errors only to set SLOs. Similar to our approach for other SLO-related elements, we kept this dimension generic so that different applications could leverage it for different contexts. For example, in addition to HTTP errors, errors for a batch processing service might be the number of records that failed to process. Tickets As previously mentioned, tickets were originally the primary way we evaluated most of our production software. For historical reasons, we decided to continue to track tickets alongside our other SLOs. You can consider this metric as analogous to something like “software operation level.” VALET We summed up our new SLOs into a handy acronym: VALET. Volume (traffic) How much business volume can my service handle? Availability Is the service up when I need it? Latency Does the service respond fast when I use it? Errors Does the service throw an error when I use it? Tickets Does the service require manual intervention to complete my request? Evangelizing SLOs Armed with an easy-to-remember acronym, we set out to evangelize SLOs to the enterprise: Why SLOs are important How SLOs support our “freedom and responsibility” culture What should be measured What to do with the results Since developers were now responsible for the operation of their software, they needed to establish SLOs to demonstrate their ability to build and support reliable software, and also to communicate with the consumers of their services and product managers for customer-facing services. However, most of this audience was unfamiliar with concepts like SLAs and SLOs, so they needed to be educated on this new VALET framework. As we needed to secure executive backing for our move to SLOs, our education campaign started with senior leadership. We then met with development teams one by one to espouse the values of SLOs. We encouraged teams to move from their custom metric-tracking mechanisms (which were often manual) to the VALET framework. To keep the momentum going, we sent a weekly SLO report in VALET format, which we paired with commentary around general reliability concepts and lessons learned from internal events, to senior leadership. This also helped frame business metrics like purchase orders created (Volume) or purchase orders that failed to process (Errors) in terms of VALET. We also scaled our evangelism in a number of ways: We set up an internal WordPress site to host blogs about VALET and reliability, outlinking to useful resources. We conducted internal Tech Talks (including a Google SRE guest speaker) to discuss general reliability concepts and how to measure with VALET. We conducted a series of VALET training workshops (which would later evolve into FiRE Academy), and opened the invite to whomever wanted to attend. The attendance for these workshops stayed strong for several months. We even created VALET laptop stickers and t-shirts to support a comprehensive internal marketing campaign. Soon everybody in the company knew VALET, and our new culture of SLOs began to take hold. SLO implementation even began to officially factor into THD’s annual performance reviews for development managers. While roughly 50 services were regularly capturing and reporting on their SLOs on a weekly basis, we were storing the metrics ad hoc in a spreadsheet. Although the idea of VALET had caught on like wildfire, we needed to automate data collection to foster widespread adoption. Automating VALET Data Collection While our culture of SLOs now had a strong foothold, automating VALET data collection would accelerate SLO adoption. TPS Reports We built a framework to automatically capture VALET data for any service that was deployed to our new GCP environment. We called this framework TPS Reports , a play on the term we used for volume and performance testing (transactions per second), and, of course, to poke fun 3 at the idea that multiple managers might want to review this data. We built the TPS Reports framework on top of GCP’s BigQuery database platform. All of the logs generated by our web-serving frontend were fed into BigQuery for processing by TPS Reports. We eventually also included metrics from a variety of other monitoring systems such as Stackdriver’s probe for availability. TPS Reports transformed this data into hourly VALET metrics that anyone could query. Newly created services were automatically registered into TPS Reports and therefore could be immediately queried. Since the data was all stored in BigQuery, we could efficiently report on VALET metrics across time frames. We used this data to build a variety of automated reports and alerts. The most interesting integration was a chatbot that let us directly report on the VALET of services in a commercial chat platform. For example, any service could display VALET for the last hour, VALET versus previous week, services out of SLO, and a variety of other interesting data right inside the chat channel. VALET service Our next step was to create a VALET application to store and report on SLO data. Because SLOs are best leveraged as a trending tool, the service tracks SLOs at daily, weekly, and monthly granularity. Note that our SLOs are a trending tool that we can use for error budgets, but aren’t directly connected to our monitoring systems. Instead, we have a variety of disparate monitoring platforms, each with its own alerting. Those monitoring systems aggregate their SLOs on a daily basis and publish to the VALET service for trending. The downside of this setup is that alerting thresholds set in the monitoring systems aren’t integrated with SLOs; however, we have the flexibility to change out monitoring systems as needed. Anticipating the need to integrate VALET with other applications not running in GCP, we created a VALET integration layer that provides an API to collect aggregated VALET data for a service daily. TPS Reports was the first system to integrate with the VALET service, and we eventually integrated with a variety of on-premises application platforms (more than half of the services registered in VALET). VALET Dashboard The VALET Dashboard (shown in Figure 3-1 ) is our UI to visualize and report on this data and is relatively straightforward. It allows users to: Register a new service. This typically means assigning the service to one or more URLs, which may already have VALET data collected. Set SLO objectives for any of the five VALET categories. Add new metrics types under each of the VALET categories. For example, one service may track latency at the 99th percentile, while another tracks latency at the 90th percentile (or both). Or, a backend processing system may track volume at a daily level (purchase orders created in a day), whereas a customer-serving frontend may track peak transactions per second. The VALET Dashboard lets users report on SLOs for many services at once, and to slice and dice the data in a variety of ways. For example, a team can view stats for all of their services that missed SLO in the past week. A team seeking to review service performance can view latency across all of their services and the services they depend upon. The VALET Dashboard stores the data in a simple Cloud SQL database, and developers use a popular commercial reporting tool to build reports. These reports became the foundation for a new best practice with developers: regular SLO reviews of their services (typically, either weekly or monthly). Based upon these reviews, developers can create action items to return a service to its SLO, or perhaps decide that an unrealistic SLO needs to be adjusted. Figure 3-1. The VALET Dashboard The Proliferation of SLOs Once SLOs were firmly cemented in the organization’s collective mind and effective automation and reporting were in place, new SLOs proliferated quickly. After tracking SLOs for about 50 services at the beginning of the year, by the end of the year we were tracking SLOs for 800 services, with about 50 new services per month being registered with VALET. Because VALET allowed us to scale SLO adoption across THD, the time effort required to develop automation was well worth it. However, other companies shouldn’t be scared away from adopting an SLO-based approach if they can’t develop similarly complex automation. While automation provided THD extra benefits, there are benefits to just writing SLOs in the first place. Applying VALET to Batch Applications As we developed robust reporting around SLOs, we discovered some additional uses for VALET. With a little adjusting, batch applications can fit into this framework as follows: Volume The volume of records processed Availability How often (as a percentage) the job completed by a certain time Latency The amount of time it takes for the job to run Errors The records that failed to process Tickets The number of times an operator has to manually fix data and reprocess a job Using VALET in Testing Since we were developing an SRE culture at the same time, we found that VALET supported our destructive testing (chaos engineering) automation in our staging environments. With the TPS Reports framework in place, we could automatically run destructive tests and record the impact (or hopefully lack of impact) to the service’s VALET data. Future Aspirations With 800 services (and growing) collecting VALET data, we have a lot of useful operational data at our disposal. We have several aspirations for the future. Now that we are effectively collecting SLOs, we want to use this data to take action. Our next step is an error budget culture similar to Google, whereby a team stops pushing new features (other than improvements to reliability) when a service is out of SLO. To protect the velocity demands of our business, we’ll have to strive to find a good balance between the SLO reporting time frame (weekly or monthly) and the frequency of SLOs being breached. Like many companies adopting error budgets, we’re weighing the pros and cons of rolling windows versus fixed windows. We want to further refine VALET to track detailed endpoints and the consumers of a service. Currently, even if a particular service has multiple endpoints, we track VALET only across the entire service. As a result, it’s difficult to distinguish between different operations (for example, a write to the catalog versus a read to the catalog; while we monitor and alert on these operations separately, we don’t track SLOs). Similarly, we’d also like to differentiate VALET results for different consumers of a service. Although we currently track latency SLOs at the web-serving layer, we’d also like to track a latency SLO for end users. This measurement would capture how factors like third-party tags, internet latency, and CDN caching affect how long it takes a page to start rendering and to complete rendering. We’d also like to extend VALET data to application deployments. Specifically, we’d like to use automation to verify that VALET is within tolerance before rolling out a change to the next server, zone, or region. We’ve started to collect information about service dependencies, and have prototyped a visual graph that shows where we’re not hitting VALET metrics along a call tree. This type of analysis will become even easier with emerging service mesh platforms. Finally, we strongly believe that the SLOs for a service should be set by the business owner of the service (often called a product manager) based on its criticality to the business. At the very least, we want the business owners to set the requirement for a service’s uptime and use that SLO as a shared objective between product management and development. Although technologists found VALET intuitive, the concept wasn’t so intuitive for product managers. We are striving to simplify the concepts of VALET using terminology relevant to them: we’ve both simplified the number of choices for uptime and provided example metrics. We also emphasize the significant investment required to move from one level to another. Here’s an example of simplified VALET metrics we might provide: 99.5%: Applications that are not used by store associates or an MVP of a new service 99.9%: Adequate for the majority of nonselling systems at THD 99.95%: Selling systems (or services that support selling systems) 99.99%: Shared infrastructure services Casting metrics in business terms and sharing a visible goal (an SLO!) between product and development will reduce a lot of misaligned expectations about reliability often seen in large companies. Summary Introducing a new process, let alone a new culture, to a large company takes a good strategy, executive buy-in, strong evangelism, easy adoption patterns, and—most of all—patience. It might take years for a significant change like SLOs to become firmly established at a company. We’d like to emphasize that The Home Depot is a traditional enterprise; if we can introduce such a large change successfully, you can too. You also don’t have to approach this task all at once. While we implemented SLOs piece by piece, developing a comprehensive evangelism strategy and clear incentive structure facilitated a quick transformation: we went from 0 to 800 SLO-supported services in less than a year. Conclusion SLOs and error budgets are powerful concepts that help address many different problem sets. These case studies from Evernote and The Home Depot present very real examples of how implementing an SLO culture can bring product development and operations closer together. Doing so can facilitate communication and better inform development decisions. It will ultimately result in better experiences for your customers—whether those customers are internal, external, humans, or other services. These two case studies highlight that SLO culture is an ongoing process and not a one-time fix or solution. While they share philosophical underpinnings, THD’s and Evernote’s measurement styles, SLIs, SLOs, and implementation details are markedly different. Both stories complement Google’s own take on SLOs by demonstrating that SLO implementation need not be Google-specific. Just as these companies tailored SLOs to their own unique environments, so can other companies and organizations. 1 But that’s a story for another book—see more details at https://bit.ly/2spqgcl . 2 Training options range from a one-hour primer to half-day workshops to intense four-week immersion with a mature SRE team, complete with a graduation ceremony and a FiRE badge. 3 As made famous in the 1999 film Office Space . Previous Chapter 2 - Implementing SLOs Next Chapter 4 - Monitoring Copyright © 2018 Google, Inc. Published by O'Reilly Media, Inc. Licensed under CC BY-NC-ND 4.0
2026-01-13T09:29:19
https://sre.google/mobaa/frayed/
Google SRE - The Museum of Borgmon Abstract Art Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content The Museum of Borgmon Abstract Art Thrilling Tapestry Shaded Notes k8s Elephant Fraying Clownfish Capture Bars of Color Within Rectangles 3D Books Thrilling Tapestry Shaded Notes k8s Elephant Fraying Clownfish Capture Bars of Color Within Rectangles 3D Books Interested in joining SRE? Google strives to cultivate an inclusive workplace. We believe diversity of perspectives and ideas leads to better discussions, decisions, and outcomes for everyone. Explore SRE opportunities at Google Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
https://git-scm.com/book/fa/v2/%d9%85%d9%82%d8%af%d9%85%d8%a7%d8%aa-%da%af%db%8c%d8%aa-git-basics-chapter-%da%af%d8%b1%d9%81%d8%aa%d9%86-%db%8c%da%a9-%d9%85%d8%ae%d8%b2%d9%86-%da%af%db%8c%d8%aa-Getting-a-Git-Repository
Git - گرفتن یک مخزن گیت (Getting a Git Repository) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. شروع به کار (getting started) 1.1 درباره ورژن کنترل (About Version Control) 1.2 تاریخچه کوتاهی از گیت (A Short History of Git) 1.3 گیت چیست؟ (What is Git) 1.4 نصب گیت (Installing Git) 1.5 ستاپ اولیه گیت (First-Time Git Setup) 1.6 دریافت کمک (Getting Help) 1.7 خلاصه (summary) 2. مقدمات گیت (git basics chapter) 2.1 گرفتن یک مخزن گیت (Getting a Git Repository) 2.2 ثبت تغییرات در مخزن (Recording Changes to the Repository) 2.3 مشاهده تاریخچه کامیت‌ها (Viewing the Commit History) 2.4 بازگرداندن تغییرات (Undoing Things) 2.5 کار کردن با ریموت ها (Working with Remotes) 2.6 تگ کردن (Tagging) 2.7 نام مستعار گیت (Git Aliases) 2.8 خلاصه (summary) 3. انشعاب‌گیری در گیت (Git Branching) 3.1 شاخه‌ها در یک نگاه (Branches in a Nutshell) 3.2 شاخه‌بندی و ادغام پایه‌ای (Basic Branching and Merging) 3.3 مدیریت شاخه‌ها (Branch Management) 3.4 روندهای کاری شاخه‌ها (Branching Workflows) 3.5 شاخه‌های راه دور (Remote Branches) 3.6 بازپایه‌گذاری (Rebasing) 3.7 خلاصه (Summary) 4. گیت روی سرور (Git on the server) 4.1 پروتکل‌ها (The Protocols) 4.2 راه‌اندازی گیت روی یک سرور (Getting Git on a Server) 4.3 ایجاد کلید عمومی SSH شما (Generating Your SSH Public Key) 4.4 نصب و راه‌اندازی سرور (Setting up server) 4.5 سرویس‌دهنده گیت (Git Daemon) 4.6 HTTP هوشمند (Smart HTTP) 4.7 گیت‌وب (GitWeb) 4.8 گیت‌لب (GitLab) 4.9 گزینه‌های میزبانی شخص ثالث (Third Party Hosted Options) 4.10 خلاصه (Summary) 5. گیت توزیع‌شده (Distributed git) 5.1 جریان‌های کاری توزیع‌شده (Distributed Workflows) 5.2 مشارکت در یک پروژه (Contributing to a Project) 5.3 نگهداری یک پروژه (Maintaining a Project) 5.4 خلاصه (Summary) 6. گیت هاب (GitHub) 6.1 راه‌اندازی و پیکربندی حساب کاربری (Account Setup and Configuration) 6.2 مشارکت در یک پروژه (Contributing to a Project) 6.3 نگهداری یک پروژه (Maintaining a Project) 6.4 مدیریت یک سازمان (Managing an organization) 6.5 اسکریپتنویسی در گیتهاب (Scripting GitHub) 6.6 خلاصه (Summary) 7. ابزارهای گیت (Git Tools) 7.1 انتخاب بازبینی (Revision Selection) 7.2 مرحله‌بندی تعاملی (Interactive Staging) 7.3 ذخیره موقت و پاک‌سازی (Stashing and Cleaning) 7.4 امضای کارهای شما (Signing Your Work) 7.5 جستجو (Searching) 7.6 بازنویسی تاریخچه (Rewriting History) 7.7 بازنشانی به زبان ساده (Reset Demystified) 7.8 ادغام پیشرفته (Advanced Merging) 7.9 بازاستفاده خودکار از حل تضادها (Rerere) 7.10 اشکال‌زدایی با گیت (Debugging with Git) 7.11 سابماژول ها (Submodules) 7.12 بسته‌بندی (Bundling) 7.13 جایگزینی (Replace) 7.14 ذخیره‌سازی اطلاعات ورود (Credential Storage) 7.15 خلاصه (Summary) 8. سفارشی‌سازی Git (Customizing Git) 8.1 پیکربندی گیت (Git Configuration) 8.2 ویژگی‌های گیت (Git Attributes) 8.3 هوک‌های گیت (Git Hooks) 8.4 یک نمونه سیاست اعمال شده توسط گیت (An Example Git-Enforced Policy) 8.5 خلاصه (Summary) 9. گیت و سیستم‌های دیگر (Git and Other Systems) 9.1 گیت به‌عنوان کلاینت (Git as a Client) 9.2 مهاجرت به گیت (Migrating to Git) 9.3 خلاصه (Summary) 10. مباحث درونی گیت (Git Internals) 10.1 ابزارها و دستورات سطح پایین (Plumbing and Porcelain) 10.2 اشیا گیت (Git Objects) 10.3 مراجع گیت (Git References) 10.4 فایل‌های بسته (Packfiles) 10.5 نگاشت (The Refspec) 10.6 پروتکل‌های انتقال (Transfer Protocols) 10.7 نگهداری و بازیابی داده‌ها (Maintenance and Data Recovery) 10.8 متغیرهای محیطی (Environment Variables) 10.9 خلاصه (Summary) A1. پیوست A: گیت در محیط‌های دیگر (Git in Other Environments) A1.1 رابط های گرافیکی (Graphical Interfaces) A1.2 گیت در ویژوال استودیو (Git in Visual Studio) A1.3 گیت در Visual Studio Code (Git in Visual Studio Code) A1.4 گیت در IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine (Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine) A1.5 گیت در Sublime Text (Git in Sublime Text) A1.6 گیت در بش (Git in Bash) A1.7 گیت در Zsh (Git in Zsh) A1.8 گیت در PowerShell (Git in PowerShell) A1.9 خلاصه (Summary) A2. پیوست B: گنجاندن گیت در برنامه‌های شما (Embedding Git in your Applications) A2.1 خط فرمان گیت (Command-line Git) A2.2 کتابخانهٔ گیت به زبان سی (Libgit2) A2.3 کتابخانه گیت برای زبان جاوا (JGit) A2.4 کتابخانه گیت برای زبان گو (go-git) A2.5 کتابخانه گیت پایتون (Dulwich) A3. پیوست C: دستورات گیت (Git Commands) A3.1 تنظیم و پیکربندی (Setup and Config) A3.2 گرفتن و ایجاد پروژه‌ها (Getting and Creating Projects) A3.3 نمونه‌برداری پایه‌ای (Basic Snapshotting) A3.4 انشعاب‌گیری و ادغام (Branching and Merging) A3.5 به‌اشتراک‌گذاری و به‌روزرسانی پروژه‌ها (Sharing and Updating Projects) A3.6 بازرسی و مقایسه (Inspection and Comparison) A3.7 عیب‌یابی (Debugging) A3.8 اعمال تغییرات به صورت پچ (Patching) A3.9 ایمیل (Email) A3.10 سیستم‌های خارجی (External Systems) A3.11 مدیریت (Administration) A3.12 دستورات سطح پایین گیت (Plumbing Commands) 2nd Edition 2.1 مقدمات گیت (git basics chapter) - گرفتن یک مخزن گیت (Getting a Git Repository) اگر فقط می‌توانید یک فصل از این کتاب را بخوانید تا با گیت آشنا شوید و راه بیوفتید، دقیقاً همینجا است. این فصل شامل تمامی دستورات پایه گیت است که شما در اکثر اوقات از آن‌ها استفاده می‌کنید. در آخر این فصل شما قادر خواهید بود یک مخزن را پیکربندی و راه‌اندازی کنید، پیگیری فایل‌ها را شروع کنید یا متوقف کنید و تغییرات را استیج و ثبت کنید. همچنین به شما نشان خواهیم داد که چگونه گیت را برای نادیده گرفتن برخی از فایل‌ها و الگوهای فایل‌ها تنظیم کنید، اشتباهات را به سرعت و راحتی بازگردانی کنید، در تاریخچه پروژه‌ جستو‌جو کنیم و تغییرات ثبت‌ شده بین کامیت‌ها را مشاهده کنیم و پوش و پول از مخزن ریموت انجام دهیم. گرفتن یک مخزن گیت (Getting a Git Repository) شما معمولاً یک مخزن گیت را به یکی از دو روش بدست می آورید: شما می توانید یک دایرکتوری محلی را که در حال حاضر تحت کنترل نسخه نیست، و آن را به یک مخزن گیت تبدیل کنید، یا شما می توانید یک مخزن گیت موجود را از جای دیگر کلان کنید. در هر صورت، شما یک مخزن گیت در ماشین محلی خود دارید که آماده کار است. ایجاد مخزن در یک پوشه موجود (Initializing a Repository in an Existing Directory) اگر یک دایرکتوری پروژه دارید که در حال حاضر تحت کنترل نسخه نیست و می خواهید با گیت شروع به کنترل آن کنید، ابتدا باید به دایرکتوری آن پروژه بروید. اگر شما هرگز این کار را انجام نداده اید، بسته به اینکه کدام سیستم را اجرا می کنید، کمی متفاوت به نظر می رسد: برای لینوکس: $ cd /home/user/my_project for macOS: $ cd /Users/user/my_project for Windows: $ cd C:/Users/user/my_project و تایپ کنید: $ git init این یک زیر دایرکتوری جدید به نام .git ایجاد می کند که شامل تمام فایل های ذخیره سازی مورد نیاز شما است — یک اسکلت ذخیره سازی گیت. در این مرحله، هیچ چیز در پروژه شما هنوز ردیابی نشده است. برای اطلاعات بیشتر در مورد اینکه دقیقاً چه فایل هایی در دایرکتوری .git که تازه ایجاد کرده اید وجود دارد، به مباحث درونی گیت (Git Internals) مراجعه کنید. اگر می خواهید نسخه فایل های موجود را کنترل کنید (در مقابل یک دایرکتوری خالی) ، احتمالا باید ردیابی این فایل ها را شروع کنید و یک commit اولیه انجام دهید. شما می توانید این کار را با چند دستور git add انجام دهید که فایل هایی را که می خواهید ردیابی کنید مشخص می کند، و سپس یک git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' ما در یک دقیقه به آنچه که این دستورات انجام می دهند می پردازیم. در این مرحله، شما یک مخزن گیت با فایل های ردیابی شده و یک commit اولیه دارید. کلون کردن یک مخزن موجود (Cloning an Existing Repository) اگر می خواهید یک کپی از یک مخزن Git موجود را دریافت کنید - به عنوان مثال، پروژه ای که می خواهید در آن مشارکت کنید - دستور مورد نیاز شما git clone است. اگر با سایر سیستم های VCS مانند Subversion آشنا هستید، متوجه خواهید شد که دستور "clone" است و نه "checkout". این یک تمایز مهم است — به جای دریافت فقط یک کپی کاری، گیت یک کپی کامل از تقریبا تمام داده هایی که سرور دارد را دریافت می کند. هر نسخه از هر فایل برای تاریخچه پروژه به طور پیش فرض زمانی که شما git clone را اجرا می کنید، پایین می آید. در واقع، اگر دیسک سرور شما خراب شود، اغلب می توانید تقریباً از کلون ها در هر کلاینت استفاده کنید تا سرور را به حالت زمانی که کلون شده بود برگردانید (ممکن است برخی از قلاب های طرف سرور را از دست بدهید، اما تمام داده های نسخه ای وجود دارد - برای جزئیات بیشتر به راه‌اندازی گیت روی یک سرور (Getting Git on a Server) مراجعه کنید). شما یک مخزن را با <git clone <url> `. به عنوان مثال، اگر می خواهید کتابخانه قابل پیوند گیت به نام `libgit2 را کلان کنید، می توانید این کار را به این شکل انجام دهید: $ git clone https://github.com/libgit2/libgit2 این یک دایرکتوری با نام libgit2 ایجاد می کند، دایرکتوری .git را در داخل آن شروع می کند، تمام داده های آن مخزن را پایین می آورد، و یک نسخه کار از آخرین نسخه را بررسی می کند. اگر به دایرکتوری جدید libgit2 که تازه ایجاد شده است بروید، فایل های پروژه را در آنجا خواهید دید، آماده برای کار یا استفاده. اگر می خواهید مخزن را به یک دایرکتوری با نام دیگری به غیر از libgit2 شبیه سازی کنید، می توانید نام دایرکتوری جدید را به عنوان یک استدلال اضافی مشخص کنید: $ git clone https://github.com/libgit2/libgit2 mylibgit این دستور همان کار قبلی را انجام می دهد، اما دایرکتوری هدف به نام mylibgit است. گیت پروتکل های انتقال مختلفی دارد که می توانید از آنها استفاده کنید. مثال قبلی از پروتکل https:// استفاده می کند، اما شما همچنین می توانید git:// یا user@server:path/to/repo.git را ببینید که از پروتکل انتقال SSH استفاده می کند. راه‌اندازی گیت روی یک سرور (Getting Git on a Server) تمام گزینه های موجود را که سرور می تواند برای دسترسی به مخزن گیت شما تنظیم کند و مزایا و معایب هر یک را معرفی می کند. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/mobaa/2024/
Museum of Borgmon Abstract Art - Google - Site Reliability Engineering Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content 2024 Gallery Page Welcome to the MoBAA Biennial. This year, our featured artists muse on the convergence of nature and monitoring graphs. At times, as we dive into the dashboards showing the quantitative data collected from Google’s production systems, shapes and patterns emerge — and our imagination takes charge, leading us to name these images after familiar animals and natural occurrences. The Crocodile’s Jaw GMail Squid Swarm Halfdome Shark Vs. Whale Werewolf Stacked Rays Build Label Cats The Swordfish The Crocodile’s Jaw GMail Squid Swarm Halfdome Shark Vs. Whale Werewolf Stacked Rays Build Label Cats The Swordfish Interested in joining SRE? Google strives to cultivate an inclusive workplace. We believe diversity of perspectives and ideas leads to better discussions, decisions, and outcomes for everyone. Explore SRE opportunities at Google Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19
http://www.trello.com/tour
What is Trello: Learn Features, Uses & More | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Your productivity powerhouse Welcome to Trello, the ultimate tool for getting things done. Whether you're managing daily to-dos, setting personal goals, or organizing your life, Trello helps you focus on what matters most and always stay on top of your game. Sign up—it’s free! Join a community of millions of users globally who are using Trello to get more done. Your to-dos, organized Stay organized and efficient with Inbox, Boards, and Planner. Every to-do, idea, or responsibility—no matter how small—finds its place, keeping you at the top of your game. Inbox When it’s on your mind, it goes in your Inbox. Capture your to-dos from anywhere, anytime. Boards Your to-do list may be long, but it can be manageable! Keep tabs on everything from "to-dos to tackle" to "mission accomplished!” Planner Drag, drop, get it done. Snap your top tasks into your calendar and make time for what truly matters. Inbox When it’s on your mind, it goes in your Inbox. Capture your to-dos from anywhere, anytime. Boards Your to-do list may be long, but it can be manageable! Keep tabs on everything from "to-dos to tackle" to "mission accomplished!” Planner Drag, drop, get it done. Snap your top tasks into your calendar and make time for what truly matters. Simplify your to-do list Trello cards keep all of your important information organized and accessible in one place. Add due dates, attach files, and mark to-do’s as “Done” when completed. Plus, AI helps you refine your text for a polished finish! Due dates Due dates go hand-in-hand with staying organized. In Trello, they’re easy to set, hard to miss (with reminders!), and oh-so-satisfying to mark as “done.” Attachments No more digging through endless email chains to find attachments. Just drag and drop them onto a card so the right files stay with the right tasks. Checklists Your best tool to overpower overwhelming asks. Break big tasks into small ones, check things off the list, and watch that status bar go to 100% complete. Powered by AI Harness the power of AI. Brainstorm new ideas, organize action items, and condense lengthy texts instantly with AI INTEGRATIONS Simplify your workflow and stay on top of your to-dos effortlessly! Integrate seamlessly with the apps and services your team relies on like Slack, Microsoft Teams, Outlook, Gmail, Confluence, and hundreds more. And if you use Jira, you can seamlessly sync and manage your tasks right within Trello using Jira Lists. Learn more > Unlock your potential with Trello Experience a new level of efficiency and productivity with Trello. Your work, your way Need to structure information on Trello cards specific to the task at hand? Custom Fields lets you do all that and more— providing process and formality to your workflow to meet your unique needs. Automate your workflow Create a foolproof process for moving work forward with Trello’s built-in automation system. Run commands and set automated rules for almost any action in Trello so that you can focus on what matters most. Learn more > Get peace of mind Easily manage users in real-time and control content-management permissions with just a few clicks. With 99.99% uptime YoY and industry-leading compliance certifications, you can rest assured we’ll never leave you hanging. Try Premium free for 14 days Supercharge your productivity with AI and unlimited automation Sign up By entering my email, I acknowledge the Atlassian Privacy Policy Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://git-scm.com/book/ms/v2/Git-on-the-Server-Smart-HTTP
Git - Smart HTTP About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Getting Started 1.1 About Version Control 1.2 A Short History of Git 1.3 What is Git? 1.4 The Command Line 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Getting Help 1.8 Summary 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.6 Git on the Server - Smart HTTP Smart HTTP We now have authenticated access through SSH and unauthenticated access through git:// , but there is also a protocol that can do both at the same time. Setting up Smart HTTP is basically just enabling a CGI script that is provided with Git called git-http-backend on the server. This CGI will read the path and headers sent by a git fetch or git push to an HTTP URL and determine if the client can communicate over HTTP (which is true for any client since version 1.6.6). If the CGI sees that the client is smart, it will communicate smartly with it; otherwise it will fall back to the dumb behavior (so it is backward compatible for reads with older clients). Let’s walk through a very basic setup. We’ll set this up with Apache as the CGI server. If you don’t have Apache setup, you can do so on a Linux box with something like this: $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env This also enables the mod_cgi , mod_alias , and mod_env modules, which are all needed for this to work properly. You’ll also need to set the Unix user group of the /srv/git directories to www-data so your web server can read- and write-access the repositories, because the Apache instance running the CGI script will (by default) be running as that user: $ chgrp -R www-data /srv/git Next we need to add some things to the Apache configuration to run the git-http-backend as the handler for anything coming into the /git path of your web server. SetEnv GIT_PROJECT_ROOT /srv/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/lib/git-core/git-http-backend/ If you leave out GIT_HTTP_EXPORT_ALL environment variable, then Git will only serve to unauthenticated clients the repositories with the git-daemon-export-ok file in them, just like the Git daemon did. Finally you’ll want to tell Apache to allow requests to git-http-backend and make writes be authenticated somehow, possibly with an Auth block like this: <Files "git-http-backend"> AuthType Basic AuthName "Git Access" AuthUserFile /srv/git/.htpasswd Require expr !(%{QUERY_STRING} -strmatch '*service=git-receive-pack*' || %{REQUEST_URI} =~ m#/git-receive-pack$#) Require valid-user </Files> That will require you to create a .htpasswd file containing the passwords of all the valid users. Here is an example of adding a “schacon” user to the file: $ htpasswd -c /srv/git/.htpasswd schacon There are tons of ways to have Apache authenticate users, you’ll have to choose and implement one of them. This is just the simplest example we could come up with. You’ll also almost certainly want to set this up over SSL so all this data is encrypted. We don’t want to go too far down the rabbit hole of Apache configuration specifics, since you could well be using a different server or have different authentication needs. The idea is that Git comes with a CGI called git-http-backend that when invoked will do all the negotiation to send and receive data over HTTP. It does not implement any authentication itself, but that can easily be controlled at the layer of the web server that invokes it. You can do this with nearly any CGI-capable web server, so go with the one that you know best. Note For more information on configuring authentication in Apache, check out the Apache docs here: https://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://sre.google/workbook/eliminating-toil/
Google SRE - Operational Efficiency: Eliminating Toil Chapter 6 - Eliminating Toil Table of Contents Foreword I Foreword II Preface 1. How SRE Relates to DevOps Part I - Foundations 2. Implementing SLOs 3. SLO Engineering Case Studies 4. Monitoring 5. Alerting on SLOs 6. Eliminating Toil 7. Simplicity Part II - Practices 8. On-Call 9. Incident Response 10. Postmortem Culture: Learning from Failure 11. Managing Load 12. Introducing Non-Abstract Large System Design 13. Data Processing Pipelines 14. Configuration Design and Best Practices 15. Configuration Specifics 16. Canarying Releases Part III - Processes 17. Identifying and Recovering from Overload 18. SRE Engagement Model 19. SRE: Reaching Beyond Your Walls 20. SRE Team Lifecycles 21. Organizational Change Management in SRE Conclusion Appendix A. Example SLO Document Appendix B. Example Error Budget Policy Appendix C. Results of Postmortem Analysis Index About the Editors Colophon Eliminating Toil By David Challoner, Joanna Wijntjes, David Huska, Matthew Sartwell, Chris Coykendall, Chris Schrier, John Looney, and Vivek Rau with Betsy Beyer, Max Luebbe, Alex Perry, and Murali Suriar Google SREs spend much of their time optimizing—squeezing every bit of performance from a system through project work and developer collaboration. But the scope of optimization isn’t limited to compute resources: it’s also important that SREs optimize how they spend their time. Primarily, we want to avoid performing tasks classified as toil . For a comprehensive discussion of toil, see Chapter 5 in Site Reliability Engineering . For the purposes of this chapter, we’ll define toil as the repetitive, predictable, constant stream of tasks related to maintaining a service. Toil is seemingly unavoidable for any team that manages a production service. System maintenance inevitably demands a certain amount of rollouts, upgrades, restarts, alert triaging, and so forth. These activities can quickly consume a team if left unchecked and unaccounted for. Google limits the time SRE teams spend on operational work (including both toil- and non-toil-intensive work) at 50% (for more context on why, see Chapter 5 in our first book). While this target may not be appropriate for your organization, there’s still an advantage to placing an upper bound on toil, as identifying and quantifying toil is the first step toward optimizing your team’s time. What Is Toil? Toil tends to fall on a spectrum measured by the following characteristics, which are described in our first book. Here, we provide a concrete example for each toil characteristic: Manual When the tmp directory on a web server reaches 95% utilization, engineer Anne logs in to the server and scours the filesystem for extraneous log files to delete. Repetitive A full tmp directory is unlikely to be a one-time event, so the task of fixing it is repetitive. Automatable 1 If your team has remediation documents with content like “log in to X, execute this command, check the output, restart Y if you see…,” these instructions are essentially pseudocode to someone with software development skills! In the tmp directory example, the solution has been partially automated. It would be even better to fully automate the problem detection and remediation by not requiring a human to run the script. Better still, submit a patch so that the software no longer breaks in this way. Nontactical/reactive When you receive too many alerts along the lines of “disk full” and “server down,” they distract engineers from higher-value engineering and potentially mask other, higher-severity alerts. As a result, the health of the service suffers. Lacks enduring value Completing a task often brings a satisfying sense of accomplishment, but this repetitive satisfaction isn’t a positive in the long run. For example, closing that alert-generated ticket ensured that the user queries continued to flow and HTTP requests continued to serve with status codes Grows at least as fast as its source Many classes of operational work grow as fast as (or faster than) the size of the underlying infrastructure. For example, you can expect time spent performing hardware repairs to increase in lock-step fashion with the size of a server fleet. Physical repair work may unavoidably scale with the number of machines, but ancillary tasks (for example, making software/configuration changes) doesn’t necessarily have to. Sources of toil may not always meet all of these criteria, but remember that toil comes in many forms. In addition to the preceding traits, consider the effect a particular piece of work has on team morale. Do people enjoy doing a task and find it rewarding, or is it the type of work that’s often neglected because it’s viewed as boring or unrewarding? 2 Toil can slowly deflate team morale. Time spent working on toil is generally time not spent thinking critically or expressing creativity; reducing toil is an acknowledgment that an engineer’s effort is better utilized in areas where human judgment and expression are possible. Example: Manual Response to Toil by John Looney, Production Engineering Manager at Facebook, and always an SRE at heart It’s not always clear that a certain chunk of work is toil. Sometimes, a “creative” solution—writing a workaround—is not the right call. Ideally, your organization should reward root-cause fixes over fixes that simply mask a problem. My first assignment after joining Google (April 2005) was to log in to broken machines, investigate why they were broken, then fix them or send them to a hardware technician. This task seemed simple until I realized there were over 20,000 broken machines at any given time! The first broken machine I investigated had a root filesystem that was completely full with gigabytes of nonsense logs from a Google-patched network driver. I found another thousand broken machines with the same problem. I shared my plan to address the issue with my teammate: I’d write a script to ssh into all broken machines and check if the root filesystem was full. If the filesystem was full, the script would truncate any logs larger than a megabyte in /var/log and restart syslog. My teammate’s less-than-enthusiastic reaction to my plan gave me pause. He pointed out that it’s better to fix root causes when possible. In the medium to long term, writing a script that masked the severity of the problem would waste time (by not fixing the actual problem) and potentially cause more problems later. Analysis demonstrated that each server probably cost $1 per hour. According to my train of thought, shouldn’t cost be the most important metric? I hadn’t considered that if I fixed the symptom, there would be no incentive to fix the root cause: the kernel team’s release test suite didn’t check the volume of logs these machines produced. The senior engineer directed me at the kernel source so I could find the offensive line of code and log a bug against the kernel team to improve their test suite. My objective cost/benefit analysis showing that the problem was costing Google $1,000 per hour convinced the devs to fix the problem with my patch. My patch was turned into a new kernel release that evening, and the next day I rolled it out to the affected machines. The kernel team updated their test suite later the following week. Instead of the short-term endorphin hit of fixing those machines every morning, I now had the more cerebral pleasure of knowing that I’d fixed the problem properly. Measuring Toil How do you know how much of your operational work is toil? And once you’ve decided to take action to reduce toil, how do you know if your efforts were successful or justified? Many SRE teams answer these questions with a combination of experience and intuition. While such tactics might produce results, we can improve upon them. Experience and intuition are not repeatable, objective, or transferable . Members of the same team or organization often arrive at different conclusions regarding the magnitude of engineering effort lost to toil, and therefore prioritize remediation efforts differently. Furthermore, toil reduction efforts can span quarters or even years (as demonstrated by some of the case studies in this chapter), during which time team priorities and personnel can change. To maintain focus and justify cost over the long term, you need an objective measure of progress. Usually, teams must choose a toil-reduction project from several candidates. An objective measure of toil allows your team to evaluate the severity of the problems and prioritize them to achieve maximum return on engineering investment. Before beginning toil reduction projects, it’s important to analyze cost versus benefit and to confirm that the time saved through eliminating toil will (at minimum) be proportional to the time invested in first developing and then maintaining an automated solution ( Figure 6-1 ). Projects that look “unprofitable” from a simplistic comparison of hours saved versus hours invested might still be well worth undertaking because of the many indirect or intangible benefits of automation. Potential benefits could include: Growth in engineering project work over time, some of which will further reduce toil Increased team morale and decreased team attrition and burnout Less context switching for interrupts, which raises team productivity Increased process clarity and standardization Enhanced technical skills and career growth for team members Reduced training time Fewer outages attributable to human errors Improved security Shorter response times for user requests Figure 6-1. Estimate the amount of time you’ll spend on toil reduction efforts, and make sure that the benefits outweigh the cost (source: xkcd.com/1319/ ) So how do we recommend you measure toil? Identify it. Chapter 5 of the first SRE book offers guidelines for identifying the toil in your operations. The people best positioned to identify toil depend upon your organization. Ideally, they will be stakeholders, including those who will perform the actual work. Select an appropriate unit of measure that expresses the amount of human effort applied to this toil. Minutes and hours are a natural choice because they are objective and universally understood. Be sure to account for the cost of context switching. For efforts that are distributed or fragmented, a different well-understood bucket of human effort may be more appropriate. Some examples of units of measure include an applied patch, a completed ticket, a manual production change, a predictable email exchange, or a hardware operation. As long as the unit is objective, consistent, and well understood, it can serve as a measurement of toil. Track these measurements continuously before, during, and after toil reduction efforts. Streamline the measurement process using tools or scripts so that collecting these measurements doesn’t create additional toil! Toil Taxonomy Toil, like a crumbling bridge or a leaky dam, hides in the banal day to day. The categories in this section aren’t exhaustive, but represent some common categories of toil. Many of these categories seem like “normal” engineering work, and they are. It’s helpful to think of toil as a spectrum rather than a binary classification. Business Processes This is probably the most common source of toil. Maybe your team manages some computing resource—compute, storage, network, load balancers, databases, and so on—along with the hardware that supplies that resource. You deal with onboarding users, configuring and securing their machines, performing software updates, and adding and removing servers to moderate capacity. You also work to minimize cost or waste of that resource. Your team is the human interface to the machine, typically interacting with internal customers who file tickets for their needs. Your organization may even have multiple ticketing systems and work intake systems. Ticket toil is a bit insidious because ticket-driven business processes usually accomplish their goal. Users get what they want, and because the toil is typically dispersed evenly across the team, the toil doesn’t loudly and obviously call for remediation. Wherever a ticket-driven process exists, there’s a chance that toil is quietly accumulating nearby. Even if you’re not explicitly planning to automate a process, you can still perform process improvement work such as simplification and streamlining—the processes will be easier to automate later, and easier to manage in the meantime. Production Interrupts Interrupts are a general class of time-sensitive janitorial tasks that keep systems running. For example, you may need to fix an acute shortage of some resource (disk, memory, I/O) by manually freeing up disk space or restarting applications that are leaking memory. You may be filing requests to replace hard drives, “kicking” unresponsive systems, or manually tweaking capacity to meet current or expected loads. Generally, interrupts take attention away from more important work. Release Shepherding In many organizations, deployment tools automatically shepherd releases from development to production. Even with automation, thorough code coverage, code reviews, and numerous forms of automated testing, this process doesn’t always go smoothly. Depending on the tooling and release cadence, release requests, rollbacks, emergency patches, and repetitive or manual configuration changes, releases may still generate toil. Migrations You may find yourself frequently migrating from one technology to another. You perform this work manually or with limited scripting because, hopefully, you’re only going to move from X to Y once. Migrations come in many forms, but some examples include changes of data stores, cloud vendors, source code control systems, application libraries, and tooling. If you approach a large-scale migration manually, the migration quite likely involves toil . You may be inclined to execute the migration manually because it’s a one-time effort. While you might even be tempted to view it as “project work” rather than “toil”, migration work can also meet many of the criteria of toil. Technically, modifying backup tooling for one database to work with another is software development, but this work is basically just refactoring code to replace one interface with another. This work is repetitive, and to a large extent, the business value of the backup tooling is the same as before. Cost Engineering and Capacity Planning Whether you own hardware or use an infrastructure provider (cloud), cost engineering and capacity planning usually entail some associated toil. For example: Ensuring a cost-effective baseline or burstable capability for future needs across resources like compute, memory, or IOPS (input/output operations per second). This may translate into purchase orders, AWS Reserved Instances, or Cloud/Infrastructure as a Service contract negotiation. Preparing for (and recovering from) critical high-traffic events like a product launch or holiday. Reviewing downstream and upstream service levels/limits. Optimizing workload against different footprint configurations. (Do you want to buy one big box, or four smaller boxes?) Optimizing applications against the billing specifics of proprietary cloud service offerings (DynamoDB for AWS or Cloud Datastore for GCP). Refactoring tooling to make better use of cheaper “spot” or “preemptable” resources. Dealing with oversubscribed resources, either upstream with your infrastructure provider or with your downstream customers. Troubleshooting for Opaque Architectures Distributed microservice architectures are now common, and as systems become more distributed, new failure modes arise. An organization may not have the resources to build sophisticated distributed tracing, high-fidelity monitoring, or detailed dashboards. Even if the business does have these tools, they might not work with all systems. Troubleshooting may even require logging in to individual systems and writing ad hoc log analytics queries with scripting tools. Troubleshooting itself isn’t inherently bad, but you should aim to focus your energy on novel failure modes—not the same type of failure every week caused by brittle system architecture. With each new critical upstream dependency of availability P, availability decreases by 1 – P due to the combined chance of failure. A four 9s service that adds nine critical four 9s dependencies is now a three 9s service. 3 Toil Management Strategies We’ve found that performing toil management is critical if you’re operating a production system of any scale. Once you identify and quantify toil, you need a plan for eliminating it. These efforts may take weeks or quarters to accomplish, so it’s important to have a solid overarching strategy. Eliminating toil at its source is the optimal solution, but if doing so isn’t possible, then you must handle the toil by other means. Before we dive into the specifics of two in-depth case studies, this section provides some general strategies to consider when you’re planning a toil reduction effort. As you’ll observe across the two stories, the nuances of toil vary from team to team (and from company to company), but regardless of specificity, some common tactics ring true for organizations of any size or flavor. Each of the following patterns is illustrated in a concrete way in at least one of the subsequent case studies. Identify and Measure Toil We recommend that you adopt a data-driven approach to identify and compare sources of toil, make objective remedial decisions, and quantify the time saved (return on investment) by toil reduction projects. If your team is experiencing toil overload, treat toil reduction as its own project. Google SRE teams often track toil in bugs and rank toil according to the cost to fix it and the time saved by doing so. See the section Measuring Toil for techniques and guidance. Engineer Toil Out of the System The optimal strategy for handling toil is to eliminate it at the source. Before investing effort in managing the toil generated by your existing systems and processes, examine whether you can reduce or eliminate that toil by changing the system. A team that runs a system in production has invaluable experience with how that system works. They know the quirks and tedious bits that cause the most amount of toil. An SRE team should apply this knowledge by working with product development teams to develop operationally friendly software that is not only less toilsome, but also more scalable, secure, and resilient. Reject the Toil A toil-laden team should make data-driven decisions about how best to spend their time and engineering effort. In our experience, while it may seem counterproductive, rejecting a toil-intensive task should be the first option you consider. For a given set of toil, analyze the cost of responding to the toil versus not doing so. Another tactic is to intentionally delay the toil so that tasks accumulate for batch or parallelized processing. Working with toil in larger aggregates reduces interrupts and helps you identify patterns of toil, which you can then target for elimination. Use SLOs to Reduce Toil As discussed in Implementing SLOs , services should have a documented service level objective (SLO) . A well-defined SLO enables engineers to make informed decisions. For example, you might ignore certain operational tasks if doing so does not consume or exceed the service’s error budget. An SLO that focuses on overall service health, rather than individual devices, is more flexible and sustainable as the service grows. See Implementing SLOs for guidance on writing effective SLOs. Start with Human-Backed Interfaces If you have a particularly complex business problem with many edge cases or types of requests, consider a partially automated approach as an interim step toward full automation. In this approach, your service receives structured data—usually via a defined API—but engineers may still handle some of the resulting operations. Even if some manual effort remains, this “engineer behind the curtain” approach allows you to incrementally move toward full automation. Use customer input to progress toward a more uniform way of collecting this data; by decreasing free-form requests, you can move closer to handling all requests programmatically. This approach can save back and forth with customers (who now have clear indicators of the information you need) and save you from overengineering a big-bang solution before you’ve fully mapped and understood the domain. Provide Self-Service Methods Once you’ve defined your service offering via a typed interface (see Start with Human-Backed Interfaces ), move to providing self-service methods for users. You can provide a web form, binary or script, API, or even just documentation that tells users how to issue pull requests to your service’s configuration files. For example, rather than asking engineers to file a ticket to provision a new virtual machine for their development work, give them a simple web form or script that triggers the provisioning. Allow the script to gracefully degrade to a ticket for specialized requests or if a failure occurs. 4 Human-backed interfaces are a good start in the war against toil, but service owners should always aim to make their offerings self-service where possible. Get Support from Management and Colleagues In the short term, toil reduction projects reduce the staff available to address feature requests, performance improvements, and other operational tasks. But if the toil reduction is successful, in the long term the team will be healthier and happier, and have more time for engineering improvements. It is important for everyone in the organization to agree that toil reduction is a worthwhile goal. Manager support is crucial in defending staff from new demands. Use objective metrics about toil to make the case for pushback. Promote Toil Reduction as a Feature To create strong business cases for toil reduction, look for opportunities to couple your strategy with other desirable features or business goals. If a complementary goal—for example, security, scalability, or reliability—is compelling to your customers, they’ll be more willing to give up their current toil-generating systems for shiny new ones that aren’t as toil intentive. Then, reducing toil is just a nice side effect of helping users! Start Small and Then Improve Don’t try to design the perfect system that eliminates all toil. Automate a few high-priority items first, and then improve your solution using the time you gained by eliminating that toil, applying the lessons learned along the way. Pick clear metrics such as MTTR (Mean Time to Repair) to measure your success. Increase Uniformity At scale, a diverse production environment becomes exponentially harder to manage. Special devices require time-consuming and error-prone ongoing management and incident response. You can use the “pets versus cattle” approach 5 to add redundancy and enforce consistency in your environment. Choosing what to consider cattle depends on the needs and scale of an organization. It may be reasonable to evaluate network links, switches, machines, racks of machines, or even entire clusters as interchangeable units. Shifting devices to a cattle philosophy may have a high initial cost, but can reduce the cost of maintenance, disaster recovery, and resource utilization in the medium to long term. Equipping multiple devices with the same interface implies that they have consistent configuration, are interchangeable, and require less maintenance. A consistent interface (to divert traffic, restore traffic, perform a shutdown, etc.) for a variety of devices allows for more flexible and scalable automation. Google aligns business incentives to encourage engineering teams to unify across our ever-evolving toolkit of internal technologies and tools. Teams are free to choose their own approaches, but they have to own the toil generated by unsupported tools or legacy systems. Assess Risk Within Automation Automation can save countless hours in human labor, but in the wrong circumstances, it can also trigger outages. In general, defensive software is always a good idea; when automation wields admin-level powers, defensive software is crucial. Every action should be assessed for its safety before execution. This includes changes that might reduce serving capacity or redundancy. When you’re implementing automation, we recommend the following practices: Handle user input defensively, even if that input is flowing from upstream systems—that is, be sure to validate the input carefully in context. Build in safeguards that are equivalent to the types of indirect alerts that a human operator might receive. Safeguards might be as simple as command timeouts, or might be more sophisticated checks of current system metrics or the number of current outages. For this reason, monitoring, alerting, and instrumentation systems should be consumable by both machine and human operators. Be aware that even read operations, naively implemented, can spike device load and trigger outages. As automation scales, these safety checks can eventually dominate workload. Minimize the impact of outages caused by incomplete safety checks of automation. Automation should default to human operators if it runs into an unsafe condition. Automate Toil Response Once you identify a piece of toil as automatable, it’s worthwhile to consider how to best mirror the human workflow in software. You rarely want to literally transcribe a human workflow into a machine workflow. Also note that automation shouldn’t eliminate human understanding of what’s going wrong. Once your process is thoroughly documented, try to break down the manual work into components that can be implemented separately and used to create a composable software library that other automation projects can reuse later. As the upcoming datacenter repair case study illustrates, automation often provides the opportunity to reevaluate and simplify human workflows. Use Open Source and Third-Party Tools Sometimes you don’t have to do all of the work to reduce toil yourself. Many efforts like one-off migrations may not justify building their own bespoke tooling, but you’re probably not the first organization to tread this path. Look for opportunities to use or extend third-party or open source libraries to reduce development costs, or at least to help you transition to partial automation. Use Feedback to Improve It’s important to actively seek feedback from other people who interact with your tools, workflows, and automation. Your users will make different assumptions about your tools depending on their understanding of the underlying systems. The less familiar your users are with these systems, the more important it is to actively seek feedback from users. Leverage surveys, user experience (UX) studies, and other mechanisms to understand how your tools are used, and integrate this feedback to produce more effective automation in the future. Human input is only one dimension of feedback you should consider. Also measure the effectiveness of automated tasks according to metrics like latency, error rate, rework rate, and human time saved (across all groups involved in the process). Ideally, find high-level measures you can compare before and after any automation or toil reduction efforts. Legacy Systems Most engineers with SRE-like responsibilities have encountered at least one legacy system in their work. These older systems often introduce problems with respect to user experience, security, reliability, or scalability. They tend to operate like a magical black box in that they “mostly work,” but few people understand how they work. They’re scary and expensive to modify, and keeping them running often requires a good deal of toilsome operational ritual. The journey away from a legacy system usually follows this path: Avoidance: There are many reasons to not tackle this problem head on: you may not have the resources to replace this system. You judge the cost and risk to your business as not worth the cost of a replacement. There may not be any substantially better solutions commercially available. Avoidance is effectively choosing to accept technical debt and to move away from SRE principles and toward system administration. Encapsulation/augmentation: You can bring SREs on board to build a shell of abstracted APIs, automation, configuration management, monitoring, and testing around these legacy systems that will offload work from SAs. The legacy system remains brittle to change, but now you can at least reliably identify misbehavior and roll back when appropriate. This tactic is still avoidance, but is a bit like refinancing high-interest technical debt into low-interest technical debt. It’s usually a stopgap measure to prepare for an incremental replacement. Replacement/refactoring: Replacing a legacy system can require a vast amount of determination, patience, communication, and documentation. It’s best undertaken incrementally. One approach is to define a common interface that sits in front of and abstracts a legacy system. This strategy helps you slowly and safely migrate users to alternatives using release engineering techniques like canarying or blue-green deployments. Often, the “specification” of a legacy system is really defined only by its historical usage, so it’s helpful to build production-sized data sets of historical expected inputs and outputs to build confidence that new systems aren’t diverging from expected behavior (or are diverging in an expected way). Retirement/custodial ownership: Eventually the majority of customers or functionality is migrated to one or more alternatives. To align business incentives, stragglers who haven’t migrated can assume custodial ownership of remnants of the legacy system. Case Studies The following case studies illustrate the strategies for toil reduction just discussed. Each story describes an important area of Google’s infrastructure that reached a point at which it could no longer scale sublinearly with human effort; over time, an increasing number of engineer hours resulted in smaller returns on that investment. Much of that effort you’ll now recognize as toil. For each case study, we detail how the engineers identified, assessed, and mitigated that toil. We also discuss the results and the lessons we learned along the way. In the first case study, Google’s datacenter networking had a scaling problem: we had a massive number of Google-designed components and links to monitor, mitigate, and repair. We needed a strategy to minimize the toilsome nature of this work for datacenter technicians. The second case study focuses on a team running their own “outlier” specialized hardware to support toil-intensive business processes that had become deeply entrenched within Google. This case study illustrates benefits of reevaluating and replacing operationally expensive business processes. It demonstrates that with a little persistence and perseverance, it’s possible to move to alternatives even when constrained by the institutional inertia of a large organization. Taken together, these case studies provide a concrete example of each toil reduction strategy covered earlier. Each case study begins with a list of relevant toil reduction strategies. Case Study 1: Reducing Toil in the Datacenter with Automation Note Toil reduction strategies highlighted in Case Study 1: Engineer toil out of the system Start small and then improve Increase uniformity Use SLOs to reduce toil Assess risk within automation Use feedback to improve Automate toil response Background This case study takes place in Google’s datacenters. Similar to all datacenters, Google’s machines are connected to switches, which are connected to routers. Traffic flows in and out from these routers via links that in turn connect to other routers on the internet. As Google’s requirements for handling internet traffic grew, the number of machines required to serve that traffic increased dramatically. Our datacenters grew in scope and complexity as we figured out how to serve a large amount of traffic efficiently and economically. This growth changed the nature of datacenter manual repairs from occasional and interesting to frequent and rote—two signals of toil. When Google first began running its own datacenters, each datacenter’s network topology featured a small number of network devices that managed traffic to a large number of machines. A single network device failure could significantly impact network performance, but a relatively small team of engineers could handle troubleshooting the small number of devices. At this early stage, engineers debugged problems and shifted traffic away from failed components manually. Our next-generation datacenter had significantly more machines and introduced software-defined networking (SDN) with a folded Clos topology which greatly increased the number of switches. Figure 6-2 shows the complexity of traffic flow for a small datacenter Clos switch network. This proportionately larger number of devices meant that a larger number of components could now fail. While each individual failure had less impact on network performance than before, the sheer volume of issues began to overwhelm the engineering staff. In addition to introducing a heavy load of new problems to debug, the complex layout was confusing to technicians: Which exact links needed to be checked? Which line card 6 did they need to replace? What was a Stage 2 switch, versus a Stage 1 or Stage 3 switch? Would shutting down a switch create problems for users? Figure 6-2. A small Clos network, which supports 480 machines attached below Stage 1 Repairing failed datacenter line cards was one obvious growing work backlog, so we targeted this task as our first stage of creating datacenter network repair automation. This case study describes how we introduced repair automation for our first generation of line cards (named Saturn). We then discuss the improvements we introduced with the next generation of line cards for Jupiter fabrics. As shown in Figure 6-3 , before the automation project, each fix in the datacenter line-card repair workflow required an engineer to do the following: Check that it was safe to move traffic from the affected switch. Shift traffic away from the failed device (a “drain” operation). Perform a reboot or repair (such as replacing a line card). Shift traffic back to the device (an “undrain” operation). This unvarying and repetitive work of draining, undraining, and repairing devices is a textbook example of toil. The repetitive nature of the work introduced problems of its own—for example, engineers might multitask by working on a line card while also debugging more challenging problems. As a result, the distracted engineer might accidentally introduce an unconfigured switch back to the network. Figure 6-3. Datacenter (Saturn) line-card repair workflow before automation: all steps require manual work Problem Statement The datacenter repairs problem space had the following dimensions: We couldn’t grow the team fast enough to keep up with the volume of failures, and we couldn’t fix problems fast enough to prevent negative impact to the fabric. Performing the same steps repeatedly and frequently introduced too many human errors. Not all line-card failures had the same impact. We didn’t have a way to prioritize more serious failures. Some failures were transient. We wanted the option to restart the line card or reinstall the switch as a first pass at repair. Ideally, we could then programmatically capture the problem if it happened again and flag the device for replacement. The new topology required us to manually assess the risk of isolating capacity before we could take action. Every manual risk assessment was an opportunity for human error that could result in an outage. Engineers and technicians on the floor didn’t have a good way to gauge how many devices and links would be impacted by their planned repair. What We Decided to Do Instead of assigning every issue to an engineer for risk assessment, drain, undrain, and validation, we decided to create a framework for automation that, when coupled with an on-site technician where appropriate, could support these operations programmatically. Design First Effort: Saturn Line-Card Repair Our high-level goal was to build a system that would respond to problems detected on network devices, rather than relying on an engineer to triage and fix these problems. Instead of sending a “line card down” alert to an engineer, we wrote the software to request a drain (to remove traffic) and create a case for a technician. The new system had a few notable features: We leveraged existing tools where possible. As shown in Figure 6-3 , our alerting could already detect problems on the fabric line cards; we repurposed that alerting to trigger an automated repair. The new workflow also repurposed our ticketing system to support network repairs. We built in automated risk assessment to prevent accidental isolation of devices during a drain and to trigger safety mechanisms where required. This eliminated a huge source of human errors. We adopted a strike policy that was tracked by software: the first failure (or strike) only rebooted the card and reinstalled the software. A second failure triggered card replacement and full return to the vendor. Implementation The new automated workflow (shown in Figure 6-4 ) proceeded as follows: The problematic line card is detected and a symptom is added to a specific component in the database. The repair service picks up the problem and enables repairs on the switch. The service performs a risk assessment to confirm that no capacity will be isolated by the operation, and then: Drains traffic from the entire switch. Shuts down the line card. If this is a first failure, reboots the card and undrains the switch, restoring service to the switch. At this point, the workflow is complete. If this is the second failure, the workflow proceeds to step 3. The workflow manager detects the new case and sends it to a pool of repair cases for a technician to claim. The technician claims the case, sees a red “stop” in the UI (indicating that the switch needs to be drained before repairs are started), and executes the repair in three steps: Initiates the chassis drain via a “Prep component” button in the technician UI. Waits for the red “stop” to clear, indicating that the drain is complete and the case is actionable. Replaces the card and closes the case. The automated repair system brings the line card up again. After a pause to give the card time to initialize, the workflow manager triggers an operation to restore traffic to the switch and close the repair case. Figure 6-4. Saturn line-card repair workflow with automation: manual work required only to push a button and replace the line card The new system freed the engineering team from a large volume of toilsome work, giving them more time to pursue productive projects elsewhere: working on Jupiter, the next-generation Clos topology. Design Second Effort: Saturn Line-Card Repair Versus Jupiter Line-Card Repair Capacity requirements in the datacenter continued to double almost every 12 months. As a result, our next-generation datacenter fabric, Jupiter, was more than six times larger than any previous Google fabric. The volume of problems was also six times larger. Jupiter presented scaling challenges for repair automation because thousands of fiber links and hundreds of line cards could fail in each layer. Fortunately, the increase in potential failure points was accompanied by far greater redundancy, which meant we could implement more ambitious automation. As shown in Figure 6-5 we preserved some of the general workflow from Saturn and added a few important modifications: After an automated drain/reboot cycle determined that we wanted to replace hardware, we sent the hardware to a technician. However, instead of requiring a technician to initiate the drain with the “Push prep button to drain switch,” we automatically drained the entire switch when it failed. We added automation for installing and pushing the configuration that engages after component replacement. We enabled automation for verifying that the repair was successful before undraining the switch. We focused attention on recovering the switch without involving a technician unless absolutely necessary. Figure 6-5. Saturn line-card down automation (left) versus Jupiter automation (right) Implementation We adopted a simple and uniform workflow for every line-card problem on Jupiter switches: declare the switch down, drain it, and begin a repair. The automation carried out the following: The problem switch-down is detected and a symptom is added to the database. The repair service picks up the problem and enables repairs on the switch: drain the entire switch, and add a drain reason. If this is the second failure within six months, proceed to step 4. Otherwise, proceed to step 3. Attempt (via two distinct methods) to power-cycle the switch. If the power-cycle is successful, run automated verification, then install and configure the switch. Remove the repair reason, clear the problem from the database, and undrain the switch. If preceding sanity-checking operations fail, send the case to a technician with an instruction message. If this was the second failure, send the case directly to the technician, requesting new hardware. After the hardware change occurs, run automated verification and then install and configure the switch. Remove the repair reason, clear the problem from the database, and undrain the switch. This new workflow management was a complete rewrite of the previous repair system. Again, we leveraged existing tools when possible: The operations for configuring new switches (install and verify) were the same operations we needed to verify that a switch that had been replaced. Deploying new fabrics quickly required the ability to BERT 7 and cable-audit 8 programmatically. Before restoring traffic, we reused that capability to automatically run test patterns on links that had fallen into repairs. These tests further improved performance by identifying faulty links. The next logical improvement was to automate mitigation and repair of memory errors on Jupiter switch line cards. As shown in Figure 6-6 , prior to automation, this workflow depended heavily on an engineer to determine if the failure was hardware- or software-related, and then to drain and reboot the switch or arrange a repair if appropriate. Figure 6-6. Jupiter memory error repair workflow before automation Our automation simplified the repair workflow by no longer attempting to troubleshoot memory errors (see Sometimes imperfect automation is good enough for why this made sense). Instead, we treated memory errors the same way we handled failed line cards. To extend automation to memory errors, we simply had to add another symptom to a config file to make it act on the new problem type. Figure 6-7 depicts the automated workflow for memory errors. Figure 6-7. Jupiter memory error repair workflow with automation Lessons Learned During the several years we worked to automate network repair, we learned a lot of general lessons about how to effectively reduce toil. UIs should not introduce overhead or complexity For Saturn-based line cards, replacing a line card required draining the entire switch. Draining the entire switch early in the repair process meant losing the working capacity of all line cards on the switch while waiting for replacement parts and a technician. We introduced a button in the UI called “Prep component” that allowed a technician to drain traffic from the entire switch right before they were ready to replace the card, thereby eliminating unnecessary downtime for the rest of the switch (see “Push prep button to drain switch” in Figure 6-5 ). This aspect of the UI and repair workflow introduced a number of unexpected problems: After pressing the button, the technician did not get feedback on drain progress but instead simply had to wait for permission to proceed. The button didn’t reliably sync with the actual state of the switch. As a result, sometimes a drained switch did not get repaired, or a technician interrupted traffic by acting upon an undrained switch. Components that did not have automation enabled returned a generic “contact engineering” message when a problem arose. Newer technicians did not know the best way to reach someone who could help. Engineers who were contacted were not always immediately available. In response to user reports and problems with regressions caused by the complexity of the feature, we designed future workflows to ensure the switch was safe and ready for repair before the technician arrived at the switch. Don’t rely on human expertise We leaned too heavily on experienced datacenter technicians to identify errors in our system (for example, when the software indicated it was safe to proceed with repairs, but the switch was actually undrained). These technicians also had to perform several tasks manually, without being prompted by automation. Experience is difficult to replicate. In one particularly high-impact episode, a technician decided to expedite the “press button and wait for results” experience by initiating concurrent drains on every line card waiting for repairs at the datacenter, resulting in network congestion and user-visible packet loss. Our software didn’t anticipate and prevent this action because we didn’t test the automation with new technicians. Design reusable components Where possible, avoid monolithic designs. Build complex automation workflows from separable components, each of which handles a distinct and well-defined task. We could easily reuse or adapt key components of our early Jupiter automation for each successive generation of fabric, and it was easier to add new features when we could build on automation that already existed. Successive variations on Jupiter-type fabrics could leverage work done in earlier iterations. Don’t overthink the problem We overanalyzed the memory error problem for Jupiter line cards. In our attempts at precise diagnosis, we sought to distinguish software errors (fixable by reboots) from hardware errors (which required card replacement), and also to identify errors that impacted traffic versus errors that did not. We spent nearly three years (2012–2015) collecting data on over 650 discrete memory error problems before realizing this exercise was probably overkill, or at least shouldn’t block our repair automation project. Once we decided to act upon any error we detected, it was straightforward to use our existing repair automation to implement a simple policy of draining, rebooting, and reinstalling switches in response to memory errors. If the problem recurred, we concluded that the failure was likely hardware-based and requested component replacement. We gathered data over the course of a quarter and discovered that most of the errors were transient—most switches recovered after being rebooted and reinstalled. We didn’t need additional data to perform the repair, so the three-year delay in implementing the automation was unnecessary. Sometimes imperfect automation is good enough While the ability to verify links with BERT before undraining them was handy, BERT tooling didn’t support network management links. We added these links into the existing link repair automation with a check that allowed them to skip verification. We were comfortable bypassing verification because the links didn’t carry customer traffic, and we could add this functionality later if verification turned out to be important. Repair automation is not fire and forget Automation can have a very long lifetime, so make sure to plan for project continuity as people leave and join the team. New engineers should be trained on legacy systems so they can fix bugs. Due to parts shortages for Jupiter fabrics, Saturn-based fabrics lived on long after the originally targeted end-of-life date, requiring us to introduce some improvements quite late in Saturn’s overall lifespan. Once adopted, automation may become entrenched for a long time, with positive and negative consequences. When possible, design your automation to evolve in a flexible way. Relying on inflexible automation makes systems brittle to change. Policy-based automation can help by clearly separating intent from a generic implementation engine, allowing automation to evolve more transparently. Build in risk assessment and defense in depth After building new tools for Jupiter that determined the risk of a drain operation before executing it, the complexity we encountered led us to introduce a secondary check for defense in depth. The second check established an upper limit for the number of impacted links, and another limit for impacted devices. If we exceeded either threshold, a tracking bug to request further investigation opened automatically. We tuned these limits over time to reduce false positives. While we originally considered this a temporary measure until the primary risk assessment stabilized, the secondary check has proven useful for identifying atypical repair rates due to power outages and software bugs (for one example, see “Automation: Enabling Failure at Scale” in Site Reliability Engineering ). Get a failure budget and manager support Repair automation can sometimes fail, especially when first introdu
2026-01-13T09:29:19
http://www.trello.com/use-cases/project-management
Project Management Without the Mayhem | Trello Skip to main content Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Use case: Task management Track progress of tasks in one convenient place with a visual layout that adds ‘ta-da’ to your to-do’s. Use case: Resource hub Save hours when you give teams a well-designed hub to find information easily and quickly. Use case: Project management Keep projects organized, deadlines on track, and teammates aligned with Trello. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Back Navigation Features Solutions Plans Pricing Resources Explore the features that help your team succeed Inbox Capture every vital detail from emails, Slack, and more directly into your Trello Inbox. Planner Sync your calendar and allocate focused time slots to boost productivity. Automation Automate tasks and workflows with Trello. Power-Ups Power up your teams by linking their favorite tools with Trello plugins. Templates Give your team a blueprint for success with easy-to-use templates from industry leaders and the Trello community. Integrations Find the apps your team is already using or discover new ways to get work done in Trello. Meet Trello Trello makes it easy for your team to get work done. No matter the project, workflow, or type of team, Trello can help keep things organized. It’s simple – sign-up, create a board, and you’re off! Productivity awaits. Check out Trello Take a page out of these pre-built Trello playbooks designed for all teams Marketing teams Whether launching a new product, campaign, or creating content, Trello helps marketing teams succeed. Product management Use Trello’s management boards and roadmap features to simplify complex projects and processes. Engineering teams Ship more code, faster, and give your developers the freedom to be more agile with Trello. Design teams Empower your design teams by using Trello to streamline creative requests and promote more fluid cross-team collaboration. Startups From hitting revenue goals to managing workflows, small businesses thrive with Trello. Remote teams Keep your remote team connected and motivated, no matter where they’re located around the world. See all teams Our product in action Read though our use cases to make the most of Trello on your team. See all use cases Standard For teams that need to manage more work and scale collaboration. Premium Best for teams up to 100 that need to track multiple projects and visualize work in a variety of ways. Enterprise Everything your enterprise teams and admins need to manage projects. Free plan For individuals or small teams looking to keep work organized. Take a tour of Trello Compare plans & pricing Whether you’re a team of 2 or 2,000, Trello’s flexible pricing model means you only pay for what you need. View Trello pricing Learn & connect Trello guide Our easy to follow workflow guide will take you from project set-up to Trello expert in no time. Remote work guide The complete guide to setting up your team for remote work success. Webinars Enjoy our free Trello webinars and become a productivity professional. Customer stories See how businesses have adopted Trello as a vital part of their workflow. Developers The sky's the limit in what you can deliver to Trello users in your Power-Up! Help resources Need help? Articles and FAQs to get you unstuck. Helping teams work better, together Discover Trello use cases, productivity tips, best practices for team collaboration, and expert remote work advice. Check out the Trello blog Project management without the mayhem From initiation to completion, Trello makes it easy to monitor every aspect of your next project. Designed for teams of any size, anywhere. Get started with Trello Manage any project (big or small) with Trello Manage all of your projects in one powerful tool that makes project management ~~manageable~~ magical Custom fields: the only limitation is your imagination Whether it’s a simple to-do list or a complex workflow, custom fields are an effective way to save time and meet the demands of any project. Pro tip: combine custom fields with automation to supercharge your boards. Automation keeps work flowing Trello’s built in automation tool handles repetitive tasks such as creating agenda cards, adding team members, or tracking due dates. Save time with automation See your projects from any angle There are multiple ways to visualize your projects with Trello: view as a dashboard, a timeline, a table, and more. Using views helps visualize the work to be done, giving you the ability to see high-level project status down to the day-to-day tasks. Learn about Views Trello for Project Management Trello + your favorite tools Trello has 200+ integrations with the tools you know and love. Integrate the tools your team already uses with Trello to keep your work organized and in one place. Try a Trello Power-Up to connect all of your team’s (other) favorite apps to your boards. Go to Power-Ups gallery Time Tracking & Reporting Analytics & Reporting Help your team log time on Trello cards to track work. Create and export reports to see aggregated data. Slack Automation Keep your communication and collaboration apps connected. Google Drive File management Search Google Drive right from Trello and attach relevant files and folders. List Limit Board utilities Set a limit on your lists and we'll highlight the list if the number of cards in it passes this limit. Confluence Communication & collaboration Pull Confluence pages into your meeting board, or start a new one from a card. Jira Developer Tools Connect Jira and Trello to help all your teams work better together. No need to start from scratch. Use one of our customizable templates. Jumpstart your board with a well-proven template designed by our team. Customize it for yours. Project Management Big dreams turn into bigger results with a killer project plan. Use this basic structure to build your team’s ideal workflow, for projects big or small. Get organized Work Request and Intake Process This board uses Google Forms, Zapier, Butler rules, and Power-Ups to create a way for coworkers to request work from you. Improve your workflow Join a community of millions of users globally who are using Trello to get more done. Join a community of millions of users globally who are using Trello to get more done. Project management tips from the pros. Read the Trello blog to find ways to improve efficiency, productivity, and collaboration. The ultimate guide to team project management There’s no such thing as winging it when it comes to effective project management. Discover how to stay aligned as a team, identify risks, improve collaboration, and remove blockers. Read more How to unlock project management perfection with Trello Get your project off the ground in a matter of seconds by discovering the secret sauce to project planning. Hint: Trello is the main ingredient. Read more Your team’s project roadmap: Gantt charts in Trello Building a project roadmap in Trello gives you an automatic way to generate Gantt charts on demand. Plan your project, and your chart will follow. Read more The most common project blockers and ways to eliminate them Project blockers come in all shapes and sizes. The key to delivering work on time is knowing your project blockers or impediments. Read more Log In About Trello What’s behind the boards. Jobs Learn about open roles on the Trello team. Apps Download the Trello App for your Desktop or Mobile devices. Contact us Need anything? Get in touch and we can help. Čeština Deutsch English Español Français Italiano Magyar Nederlands Norsk (bokmål) Polski Português (Brasil) Suomi Svenska Tiếng Việt Türkçe Русский Українська ภาษาไทย 中文 (简体) 中文 (繁體) 日本語 Notice at Collection Privacy Policy Terms Copyright © 2024 Atlassian
2026-01-13T09:29:19
https://sre.google/
Google SRE - Site Reliability engineering Site Reliability Engineering Jump to Content Home Resources Latest resources Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Distributed PubSub Distributed Image Server The Art of SLO Latest resources Resources overview Google SRE Video Gallery New! Product-Focused Reliability for SRE Twentieth Anniversary Twenty years of SRE lessons learned Prodverbs SRE Fundamentals Measuring Reliability Why Heroism is Bad System Theoretic Process Analysis Books Books overview Building Secure & Reliable Systems The Site Reliability Workbook Site Reliability Engineering Mobaa Mobaa overview 2024 Gallery 2022 Gallery 2020 Gallery Vector Methods Classroom Classroom overview Distributed PubSub Distributed Image Server The Art of SLO Books Careers Cloud Local Prodcast Spotlight Site Reliability Engineering Jump to Content What is Site Reliability Engineering (SRE)? SRE is what you get when you treat operations as if it’s a software problem. Our mission is to protect, provide for, and progress the software and systems behind all of Google’s public services — Google Search, Ads, Gmail, Android, YouTube, and App Engine, to name just a few — with an ever-watchful eye on their availability, latency, performance, and capacity. SRE Video Gallery Explore SRE topics in depth in our video gallery! Read more SRE Video Gallery Explore SRE topics in depth in our video gallery! Read more Museum of Borgmon Abstract Art Explore more of the beauty that emerges from chaos, in this biennial collection of images from our monitoring systems. Read more Museum of Borgmon Abstract Art Explore more of the beauty that emerges from chaos, in this biennial collection of images from our monitoring systems. Read more SRE Careers Hear from some of our most senior engineers about their role at Google. Read more SRE Careers Hear from some of our most senior engineers about their role at Google. Read more SRE Resources A curated list of Site Reliability and Production Engineering resources. Read more SRE Resources A curated list of Site Reliability and Production Engineering resources. Read more What is SRE? Since 2004, SRE has evolved to become the industry-leading practice for service reliability. Hear from key figures about the history of SRE and what’s next for the SRE community. Watch video What is SRE? Since 2004, SRE has evolved to become the industry-leading practice for service reliability. Hear from key figures about the history of SRE and what’s next for the SRE community. Watch video Product-Focused Reliability for SRE Learn about how a product-focused reliability model can effectively support the overall reliability of a product. Read more Product-Focused Reliability for SRE Learn about how a product-focused reliability model can effectively support the overall reliability of a product. Read more SRE Books Read our SRE books online: Building Secure & Reliable Systems, The SRE Workbook, and the original SRE book. Read online SRE Books Read our SRE books online: Building Secure & Reliable Systems, The SRE Workbook, and the original SRE book. Read online STPA (System Theoretic Process Analysis) at Google Google is using STPA to analyze pure software systems and discover the unknown unknowns. Read more STPA (System Theoretic Process Analysis) at Google Google is using STPA to analyze pure software systems and discover the unknown unknowns. Read more 1 2 3 4 5 6 7 8 What we do as SRE Our job is a combination not found elsewhere in the industry. Like traditional operations groups, we keep important, revenue-critical systems up and running despite hurricanes, bandwidth outages, and configuration errors. How we SRE at Google As SRE, we flip between the fine-grained detail of disk driver IO scheduling to the big picture of continental-level service capacity, across a range of systems and a user population measured in billions. Interested in joining SRE? Google strives to cultivate an inclusive workplace. We believe diversity of perspectives and ideas leads to better discussions, decisions, and outcomes for everyone. Explore SRE opportunities at Google Follow us About Google Google products Privacy Terms Help
2026-01-13T09:29:19