url
stringlengths
11
2.25k
text
stringlengths
88
50k
ts
timestamp[s]date
2026-01-13 08:47:33
2026-01-13 09:30:40
https://ko-kr.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT0fEAEvmavVOfJ7JaieFNpYJkNKcijyZpM1gnYFIG_5bI6cR7Xju0L5R8rVAOJjpd0zxjpipZlDG7UeNudb-wpQIeTyqEqyu80aqlwXdSxEdsAgHUdw82crQnhdTXNkOto6zYYUP1fRvDY9
Facebook Facebook 이메일 또는 휴대폰 비밀번호 계정을 잊으셨나요? 새 계정 만들기 일시적으로 차단됨 일시적으로 차단됨 회원님의 이 기능 사용 속도가 너무 빠른 것 같습니다. 이 기능 사용에서 일시적으로 차단되었습니다. Back 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch 가입하기 로그인 Messenger Facebook Lite 동영상 Meta Pay Meta 스토어 Meta Quest Ray-Ban Meta Meta AI Meta AI 콘텐츠 더 보기 Instagram Threads 투표 정보 센터 개인정보처리방침 개인정보 보호 센터 정보 광고 만들기 페이지 만들기 개발자 채용 정보 쿠키 AdChoices 이용 약관 고객 센터 연락처 업로드 및 비사용자 설정 활동 로그 Meta © 2026
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-Hosting-oplossingen-van-derden
Git - Hosting oplossingen van derden About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.9 Git op de server - Hosting oplossingen van derden Hosting oplossingen van derden Als je niet al het werk wilt verzetten wat samenhangt met het opzetten van je eigen Git server, heb je een aantal opties om je Git projecten te laten hosten op een site van een externe partij die zich hierop volledig heeft toegelegd. Als je dit doet geeft dit je een aantal voordelen: een hosting site is over het algemeen snel opgezet en het is eenvoudig om hierop projecten te beginnen, er komt geen server onderhoud of bewaking bij kijken. Zelfs als je je eigen interne server opzet en draait, wil je misschien toch een openbare hosting site voor je open source code gebruiken - het is over het algemeen makkelijker voor de open source gemeenschap om je te vinden en je ermee te helpen. Vandaag de dag heb je een groot aantal hosting opties waaruit je kunt kiezen, elk met verschillende voor- en nadelen. Om een recentelijk bijgewerkte lijst te zien, kijk even op de GitHosting pagine van de hoofd Git wiki op https://git.wiki.kernel.org/index.php/GitHosting We zullen het gebruik van GitHub tot in detail bespreken in GitHub , daar dit zo’n beetje de grootste Git host is die er is en je waarschijnlijk toch in aanraking zult komen met projecten die daar gehost worden; maar er zijn nog tientallen waar je uit kunt kiezen als je niet je eigen Git server wilt inrichten. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/gr/v2/%ce%a4%ce%b1-%ce%b8%ce%b5%ce%bc%ce%b5%ce%bb%ce%b9%cf%8e%ce%b4%ce%b7-%cf%83%cf%84%ce%bf%ce%b9%cf%87%ce%b5%ce%af%ce%b1-%cf%84%ce%bf%cf%85-Git-%ce%94%ce%bf%cf%85%ce%bb%ce%b5%cf%8d%ce%bf%ce%bd%cf%84%ce%b1%cf%82-%ce%bc%ce%b5-%ce%b1%cf%80%ce%bf%ce%bc%ce%b1%ce%ba%cf%81%cf%85%cf%83%ce%bc%ce%ad%ce%bd%ce%b1-%ce%b1%cf%80%ce%bf%ce%b8%ce%b5%cf%84%ce%ae%cf%81%ce%b9%ce%b1
Git - Δουλεύοντας με απομακρυσμένα αποθετήρια About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Ξεκινώντας με το Git 1.1 Σχετικά με τον έλεγχο εκδόσεων 1.2 Σύντομο ιστορικό του Git 1.3 Τι είναι το Git; 1.4 Η γραμμή εντολών 1.5 Εγκατάσταση του Git 1.6 Ρύθμιση του Git για πρώτη φορά 1.7 Χρησιμοποιώντας τη βοήθεια 1.8 Ανακεφαλαίωση 2. Τα θεμελιώδη στοιχεία του Git 2.1 Απόκτηση αποθετηρίου Git 2.2 Καταγραφή αλλαγών στο αποθετήριο 2.3 Χρησιμοποιώντας το ιστορικό υποβολών 2.4 Αναιρέσεις (undoing) 2.5 Δουλεύοντας με απομακρυσμένα αποθετήρια 2.6 Ετικέτες 2.7 Συντομεύεσεις στο Git 2.8 Ανακεφαλαίωση 3. Διακλαδώσεις στο Git 3.1 Οι κλάδοι με λίγα λόγια 3.2 Βασικές έννοιες διακλαδώσεων και συγχωνεύσεων 3.3 Διαχείριση κλάδων 3.4 Ροές εργασίας με διακλαδώσεις 3.5 Απομακρυσμένοι κλάδοι 3.6 Αλλαγή βάσης (rebasing) 3.7 Ανακεφαλαίωση 4. Το Git στον διακομιστή 4.1 Τα πρωτόκολλα 4.2 Εγκατάσταση του Git σε διακομιστή 4.3 Δημιουργία δημόσιου κλειδιού SSH 4.4 Στήσιμο του διακομιστή 4.5 Δαίμονες του Git 4.6 Έξυπνο HTTP 4.7 GitWeb 4.8 GitLab 4.9 Επιλογές φιλοξενίας από τρίτους 4.10 Ανακεφαλαίωση 5. Κατανεμημένο Git 5.1 Κατανεμημένες ροές εργασίας 5.2 Συνεισφέροντας σε ένα έργο 5.3 Συντήρηση ενός έργου 5.4 Ανακεφαλαίωση 6. GitHub 6.1 Δημιουργία λογαριασμού και ρύθμισή του 6.2 Συνεισφορά σε έργο 6.3 Συντήρηση ενός έργου 6.4 Διαχείριση οργανισμώνν 6.5 Συγγραφή script στο GitHub 6.6 Ανακεφαλαίωση 7. Εργαλεία του Git 7.1 Επιλογή αναθεώρησης 7.2 Διαδραστική εργασία με το στάδιο καταχώρισης 7.3 Αποθέματα και Καθαρισμός 7.4 Υπογραφή της δουλειάς μας 7.5 Αναζήτηση 7.6 Η ιστορία ξαναγράφεται 7.7 Απομυθοποίηση της reset 7.8 Προχωρημένη Συγχώνευση 7.9 Rerere 7.10 Αποσφαλμάτωση με το Git 7.11 Υπομονάδες 7.12 Δεμάτιασμα δεδομένων 7.13 Replace 7.14 Αποθήκευση διαπιστευτηρίων 7.15 Ανακεφαλαίωση 8. Εξατομίκευση του Git 8.1 Διαμόρφωση Git 8.2 Γνωρίσματα του Git 8.3 Τα άγκιστρα του Git 8.4 Ένα παράδειγμα επιβολής πολιτικής από το Git 8.5 Ανακεφαλαίωση 9. Το Git και άλλα συστήματα 9.1 Το Git ως πελάτης 9.2 Μετανάστευση στο Git 9.3 Ανακεφαλαίωση 10. Εσωτερική λειτουργία του Git 10.1 Διοχετεύσεις και πορσελάνες 10.2 Αντικείμενα του Git 10.3 Αναφορές του Git 10.4 Πακετάρισμα αρχείων 10.5 Τα refspec 10.6 Πρωτόκολλα μεταφοράς 10.7 Διατήρηση και ανάκτηση δεδομένων 10.8 Μεταβλητές περιβάλλοντος 10.9 Ανακεφαλαίωση A1. Appendix A: Το Git σε άλλα περιβάλλοντα A1.1 Γραφικές διεπαφές A1.2 Το Git στο Visual Studio A1.3 Git στο Visual Studio Code A1.4 Git στο IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git στο Sublime Text A1.6 Το Git στο Bash A1.7 Το Git στο Zsh A1.8 Το Git στο Powershell A1.9 Ανακεφαλαίωση A2. Appendix B: Ενσωμάτωση του Git στις εφαρμογές μας A2.1 Γραμμή εντολών Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Εντολές Git A3.1 Ρύθμιση και διαμόρφωση A3.2 Λήψη και δημιουργία έργων A3.3 Βασική λήψη στιγμιοτύπων A3.4 Διακλάδωση και συγχώνευση A3.5 Κοινή χρήση και ενημέρωση έργων A3.6 Επιθεώρηση και σύγκριση A3.7 Αποσφαλμάτωση A3.8 Επιθέματα A3.9 Ηλεκτρονικό ταχυδρομείο A3.10 Εξωτερικά Συστήματα A3.11 Διοίκηση A3.12 Εντολές διοχέτευσης 2nd Edition 2.5 Τα θεμελιώδη στοιχεία του Git - Δουλεύοντας με απομακρυσμένα αποθετήρια Δουλεύοντας με απομακρυσμένα αποθετήρια Για να μπορούμε να συνεργαζόμαστε σε έργα του Git, θα πρέπει να γνωρίζουμε πώς να διαχειριζόμαστε τα απομακρυσμένα αποθετήριά μας. Τα απομακρυσμένα αποθετήρια (remote repositories) είναι εκδόσεις του έργου μας που βρίσκονται στο Διαδίκτυο ή σε κάποιο δίκτυο. Μπορούμε να δημιουργήσουμε όσα θέλουμε και καθένα από αυτά είναι προσβάσιμα από μας είτε ως για ανάγνωση-μόνο είτε για ανάγνωση/εγγραφή. Η συνεργασία με άλλους συμπεριλαμβάνει τη διαχείριση αυτών των απομακρυσμένων αποθετηρίων, ώθηση και ελκυσμό δεδομένων προς και από αυτά όποτε χρειάζεται να κοινοποιήσουμε τη δουλειά μας ή να ενημερωθούμε για τη δουλειά άλλων. Η διαχείριση τέτοιων αποθετηρίων χρειάζεται ικανότητες όπως: πρόσθεση καινούριων απομακρυσμένων αποθετηρίων, διαγραφή αποθετηρίων που δεν έχουν πια κάποια χρησιμότητα, διαχείριση απομακρυσμένων κλάδων και ορισμό τους ως υπο-παρακολούθηση ή τερματισμό της παρακολούθησής τους και άλλα. Σε αυτή την ενότητα, θα ασχοληθούμε με κάποιες από αυτές τις δεξιότητες. Note Τα απομακρυσμένα αποθετήρια μπορεί να βρίσκονται στον υπολογιστή μας. Είναι απολύτως δυνατό να εργάζεστε με ένα “απομακρυσμένο” αποθετήριο, που βρίσκεται στον ίδιο υπολογιστή στον οποίο βρίσκεστε και εσείς. Η λέξη “απομακρυσμένο” δεν υπονοεί απαραίτητα ότι το αποθετήριο βρίσκεται κάπου αλλού στο δίκτυο ή στο Διαδίκτυο, αλλά μόνο ότι βρίσκεται κάπου αλλού. Όταν εργάζεστε με ένα τέτοιο απομακρυσμένο αποθετήριο περιλαμβάνει επίσης τις συνήθεις λειτουργίες ώθησης (push), ελκυσμού (pull) και ανάκτησης (fetch), όπως με κάθε άλλο απομακρυσμένο αποθετήριο. Εμφάνιση των απομακρυσμένων αποθετηρίων μας Για να δούμε τους απομακρυσμένους διακομιστές που έχουμε παραμετροποιήσει, εκτελούμε την εντολή git remote . Η εντολή αυτή θα μας επιστρέψει μια λίστα με τα ονόματα των απομακρυσμένων που έχουμε ορίσει. Αν έχουμε κλωνοποιήσει το αποθετήριό μας, θα πρέπει να βλέπουμε τουλάχιστον το origin  — που είναι το προεπιλεγμένο όνομα που δίνει το Git στον διακομιστή από τον οποίο μόλις κλωνοποιήσαμε: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Η επιλογή -v , η οποία θα μας δείξει τα σύντομα ονόματα των απομακρυσμένων αποθετηρίων μας, μαζί με τις διευθύνσεις URL που είναι συσχετισμένες με αυτά: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Αν έχουμε περισσότερα από ένα απομακρυσμένα αποθετήρια, η εντολή αυτή θα τα παραθέσει όλα. Για παράδειγμα, ένα αποθετήριο με πολλά απομακρυσμένα αποθετήρια, ώστε να συνεργάζονται πολλά άτομα, θα φαίνεται κάπως έτσι: $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Αυτό σημαίνει ότι μπορούμε να τραβήξουμε τη συνεισφορά καθενός από αυτούς τους χρήστες πολύ εύκολα. Επιπλέον ενδεχομένως μπορούμε να ωθήσουμε αλλαγές σε κάποιο ή κάποια από αυτά τα απομακρυσμένα αποθετήρια, αν και αυτό δεν το γνωρίζουμε ακόμα. Παρατηρούμε ότι τα απομακρυσμένα αποθετήρια χρησιμοποιούν πολλά πρωτόκολλα· θα καλύψουμε αναλυτικά τα πρωτόκολλα αυτά στο Εγκατάσταση του Git σε διακομιστή . Προσθήκη απομακρυσμένων αποθετηρίων Έχουμε ήδη αναφέρει και έχουμε επιδείξει πώς η εντολή git clone έμμεσα προσθέτει το απομακρυσμένο αποθετήριο origin στο αποθετήριό μας. Ας δούμε πως μπορούμε να προσθέσουμε ένα νέο απομακρυσμένο αποθετήριο άμεσα . Για να προσθέσουμε ένα νέο απομακρυσμένο αποθτεήριο Git με ένα σύντομο όνομα, το οποίο μπορούμε να θυμάστε εύκολα, εκτελούμε την εντολή git remote add <shortname> <url> : $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Τώρα πλέον μπορούμε να χρησιμοποιούμε στη γραμμή εντολών τη συμβολοσειρά pb αντί για ολόκληρη τη διεύθυνση του αποθετηρίου. Για παράδειγμα, αν θέλουμε να ανακτήσουμε (fetch) όλες τις πληροφορίες που έχει ο Paul στο αποθετήριό του, μπορούμε να εκτελέσουμε την εντολή git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Ο κλάδος master του Paul είναι πλέον προσβάσιμος τοπικά σε μας ως pb/master  — μπορούμε να τον συγχωνεύσουμε σε κάποιον δικό μας κλάδο, ή να φτιάξουμε (check out) έναν τοπικό κλάδο σε αυτό το σημείο, αν θέλουμε να το επιθεωρήσουμε. Θα δούμε περισσότερα για τους κλάδους και πώς τους χρησιμοποιούμε στην ενότητα Διακλαδώσεις στο Git . Ανάκτηση δεδομένων από απομακρυσμένα αποθετήρια Όπως μόλις είδαμε, για να πάρουμε δεδομένα από απομακρυσμένα έργα, μπορούμε να εκτελέσουμε: $ git fetch <remote> Η εντολή αυτή θα πάει στο απομακρυσμένο έργο και θα τραβήξει όλα τα δεδομένα από αυτό το απομακρυσμένο έργο που δεν έχουμε ακόμα. Αφού γίνει αυτό, θα έχουμε πρόσβαση σε όλους τους κλάδους αυτού του απομακρυσμένου έργου, τους οποίους και μπορούμε να συγχωνεύσουμε ή να τους επιθεωρήσουμε περαιτέρω. Όταν κλωνοποιήσουμε ένα αποθετήριο, το αποθετήριο αυτό αποθηκεύεται με το όνομα “origin”. Συνεπώς, η εντολή git fetch origin ανακτά όλες τις νέες αλλαγές που έχουν γίνει από τότε που κλωνοποιήσαμε το αποθετήριο ή από τότε που ανακτήσαμε δεδομένα από αυτό για τελευταία φορά. Είναι σημαντικό να τονίσουμε ότι η εντολή git fetch απλά τραβά δεδομένα στο τοπικό μας αποθετήριο· δεν συγχωνεύει τα δεδομένα αυτά με διάφορες αλλαγές που μπορεί να έχουμε κάνει εμείς τοπικά. Θα πρέπει να κάνουμε τη συγχώνευση χειροκίνητα όταν είστε έτοιμοι. Αν ο κλάδος που παρακολουθεί έναν απομακρυσμένα κλάδο (περισσότερες λεπτομέρειες για αυτό στην επόμενη ενότητα και την ενότητα Διακλαδώσεις στο Git ), μπορούμε να χρησιμοποιήσουμε την εντολή git pull για να γίνει αυτόματη ανάκτηση και συγχώνευση του απομακρυσμένου κλάδου στο τρέχοντα δικό μας. Αυτή η ροή εργασιών ενδεχομένως μας φαίνεται πιο εύκολη· επιπλέον, η εντολή git clone θέτει αυτόματα τον τοπικό μας κλάδο master να παρακολουθεί τον απομακρυσμένο κλάδου master (ή όπως ονομάζεται ο προεπιλεγμένος κλάδος) στον διακομιστή από τον οποίο κλωνοποιήσαμε. Η εκτέλεση της εντολής git pull , γενικά ανακτά τα δεδομένα από τον διακομιστή, τον οποίο είχαμε αρχικά κλωνοποιήσει και προσπαθεί να συγχωνεύσει αυτά τα δεδομάνα στον κώδικα πάνω στον οποίο εργαζόμαστε. Note Από την έκδοση 2.27 του Git και μετά, η εντολή git pull σας προειδοποιεί αν η μεταβλητή pull.rebase δεν έχει καθοριστεί. Το Git θα συνεχίσει να σας δίνει προειδοποιήσεις μέχρι να καθορίσετε αυτή τη μεταβλητή. Αν θέλετε την προεπιλεγμένη συμπεριφορά του Git (fast-forward αν είναι δυνατό, αλλιώς δημιούργησε μία υποβολή συγχώνευσης (merge commit)): git config --global pull.rebase "false" Αν θέλετε να κάνετε rebase όταν έλκουμε: git config --global pull.rebase "true" Ώθηση δεδομένων σε απομακρυσμένα αποθετήρια Όταν έχουμε φέρει κάποιο έργο μας σε σημείο που θέλουμε να το κοινοποιήσουμε, θα πρέπει να το ωθήσουμε. Η εντολή είναι απλή: git push <remote> <branch> . Για παράδειγμα, αν θέλουμε να ωθήσουμε τον τοπικό μας κλάδο master στον απομακρυσμένο διακομιστή origin (επαναλαμβάνουμε ότι η κλωνοποίηση παραμετροποιεί αυτόματα αυτά τα ονόματα), τότε μπορούμε να εκτελέσουμε αυτή την εντολή έτσι ώστε να ωθήσουμε τις υποβολές (commits) που έχουμε κάνει στον διακομιστή: $ git push origin master Η εντολή αυτή θα εκτελεστεί επιτυχώς μόνο αν έχουμε κλωνοποιήσει από έναν διακομιστή στον οποίο έχουμε δικαίωμα εγγραφής και αν κανείς άλλος δεν έχει ωθήσει δεδομένα στο μεσοδιάστημα. Αν εμείς και κάποιος άλλος έχουμε κλωνοποιήσει το έργο ενώ αυτό βρίσκεται στην ίδια κατάσταση και αυτός ωθήσει δεδομένα στον διακομιστή και μετά ωθήσουμε εμείς δεδομένα στον διακομιστή, η δική μας εντολή για ώθηση δεδομένων θα απορριφθεί. Αυτό που θα πρέπει να κάνουμε είναι να ανακτήσουμε τις αλλαγές του άλλου και να τις ενσωματώσουμε στις δικές μας, ώστε να μας επιτραπεί να ωθήσουμε. Στην ενότητα Διακλαδώσεις στο Git θα δούμε περισσότερες πληροφορίες σχετικά με την ώθηση δεδομένων σε απομακρυσμένους διακομιστές. Επιθεώρηση απομακρυσμένου αποθετηρίου Αν θέλουμε να δούμε περισσότερες πληροφορίες σχετικά με ένα απομακρυσμένο αποθετήριο, μπορούμε να χρησιμοποιήσουμε την εντολή git remote show <remote> . Αν εκτελέσουμε αυτή την εντολή με κάποιο σύντομο όνομα, όπως το origin , θα δούμε κάτι σαν: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Παρατίθεται το URL του απομακρυσμένου αποθετηρίου καθώς και οι κλάδοι τους οποίους παρακολουθούμε. Επίσης μας ενημερώνει ότι αν βρισκόμαστε στον κλάδο master και εκτελέσουμε git pull , ο κλάδος master του απομακρυσμένου αποθετηρίου θα συγχωνευτεί αυτόματα στον τοπικό κλάδο master αφότου ανακτηθεί. Επίσης παραθέτει τις απομακρυσμένες αναφορές που έχει κατεβάσει. Το παραπάνω είναι ένα απλό παράδειγμα που ενδεχομένως θα συναντήσουμε. Όταν αρχίσουμε να χρησιμοποιούμε πιο εκτεταμένα το Git, μπορεί να δούμε πολύ περισσότερες πληροφορίες όταν εκτελούμε την εντολή git remote show . $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Αυτή η εντολή μας δείχνει σε αυτή την περίπτωση σε ποιον κλάδο ωθούμε δεδομένα όταν βρισκόμαστε σε συγκεκριμένους κλάδους και εκτελούμε την εντολή git push . Επίσης μας δείχνει ποιους απομακρυσμένους κλάδους του διακομιστή δεν έχουμε ακόμα, ποιους απομακρυσμένους κλάδους έχουμε αλλά έχουν αφαιρεθεί από τον διακομιστή, καθώς και τους κλάδους που θα συγχωνευτούν αυτόματα στους τοπικούς κλάδους που τους παρακολουθούν αν εκτελέσουμε την εντολή git pull . Μετονομασία και διαγραφή απομακρυσμένων αποθετηρίων Αν θέλουμε να μετονομάσουμε το σύντομο όνομα ενός απομακρυσμένου αποθετηρίου, εκτελούμε την εντολή git remote rename . Για παράδειγμα, αν θέλουμε να μετονομάσουμε το pb σε paul , μπορούμε να χρησιμοποιήσουμε την git remote rename : $ git remote rename pb paul $ git remote origin paul Αξίζει να σημειωθεί ότι η εντολή αυτή αλλάζει επίσης τα ονόματα των απομακρυσμένων κλάδων που παρακολουθούμε. Στον κλάδο στον οποίο αναφερόμασταν ως pb/master πλέον θα αναφερόμαστε ως paul/master . Αν θελήσουμε να διαγράψουμε για κάποιο λόγο ένα απομακρυσμένο αποθετήριο· έχουμε μετακίνησει τον διακομιστή σε άλλη διεύθυνση ή δεν χρησιμοποιούμε καθόλου το συγκεκριμένο αποθετήριο, ή απλά κάποιος συνεργάτης έχει εγκαταλείψει — μπορούμε να χρησιμοποιήσουμε είτε την εντολή git remote remove ή την git remote rm : $ git remote remove paul $ git remote origin Μόλις σβήσουμε την αναφορά σε κάποιο απομακρυσμένο αποθετήριο με αυτό το τρόπο, όλοι οι απομακρυσμένοι κλάδοι και όλες οι ρυθμίσεις που είναι σχετικές με αυτό το απομακρυσμένο αποθετήριο, θα σβηστούν επίσης. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/id/v2/Git-Basics-Mendapatkan-Repository-Git
Git - Mendapatkan Repository Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Memulai 1.1 Tentang Version Control 1.2 Sejarah Singkat Git 1.3 Dasar-dasar Git 1.4 Command Line 1.5 Memasang Git 1.6 Pengaturan Awal Git 1.7 Mendapatkan Bantuan 1.8 Kesimpulan 2. Git Basics 2.1 Mendapatkan Repository Git 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Alias Git 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git di Server 4.1 Protokol 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Ringkasan 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Pengaturan dan Konfigurasi Akun 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Mengelola Organization 6.5 Scripting GitHub 6.6 Ringkasan 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Kostumisasi Git 8.1 Konfigurasi Git 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Ringkasan 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.1 Git Basics - Mendapatkan Repository Git If you can read only one chapter to get going with Git, this is it. This chapter covers every basic command you need to do the vast majority of the things you’ll eventually spend your time doing with Git. By the end of the chapter, you should be able to configure and initialize a repository, begin and stop tracking files, and stage and commit changes. We’ll also show you how to set up Git to ignore certain files and file patterns, how to undo mistakes quickly and easily, how to browse the history of your project and view changes between commits, and how to push and pull from remote repositories. Mendapatkan Repository Git Anda bisa mendapatkan project Git menggunakan dua pendekatan utama. Pendekatan yang pertama mengambil project atau direktori yang ada dan mengimpornya ke Git. Pendekatan yang kedua dengan melakukan clone ke repositori Git yang ada dari server lain. Menginisialisasi Repositori di Direktori yang Ada Jika Anda mulai melacak project yang ada di Git, Anda perlu masuk ke direktori dan jenis proyek $ git init Ini akan membuat sub direktori baru bernama .git yang berisi semua file repositori yang diperlukan - kerangka penyimpanan repositori Git. Pada titik ini, tidak ada yang dilacak dalam project Anda. (Lihat Git Internals untuk informasi lebih lanjut tentang file apa saja yang terdapat di direktori .git yang baru saja Anda buat.) Jika Anda ingin memulai mengendalikan-versi file yang ada (berlawanan dengan direktori kosong), Anda mungkin harus mulai melacak file-file tersebut dan melakukan commit awal. Anda bisa mencapainya dengan beberapa perintah git add yang menentukan file yang ingin Anda lacak, diikuti dengan git commit : $ git add *.c $ git add LICENSE $ git commit -m 'initial project version' Kita akan membahas apa yang diperintahkan commands ini hanya dalam satu menit. Pada titik ini, Anda memiliki repositori Git dengan file yang dilacak dan commit awal. Menduplikat Repositori yang Ada Jika Anda ingin mendapatkan duplikat repositori Git yang ada - misalnya, sebuah project yang ingin Anda kontribusikan - command yang Anda butuhkan adalah git clone . Jika Anda terbiasa dengan sistem VCS lain seperti Subversion , Anda akan melihat bahwa command -nya adalah "clone" dan bukan "checkout". Setiap versi setiap file untuk sejarah project ditarik secara default saat Anda menjalankan git clone . Sebenarnya, jika disk server Anda rusak, Anda dapat sering menggunakan hampir semua duplikat pada klien mana pun untuk mengatur server kembali ke keadaan saat duplikat (Anda mungkin kehilangan beberapa kait sisi server dan semacamnya, tapi semua data berversi akan ada di sana - lihat Getting Git on a Server untuk lebih jelasnya). Anda menduplikat sebuah repositori dengan git clone [url] . Misalnya, jika Anda ingin menduplikat pustaka yang terhubung dengan Git yang disebut libgit2, Anda dapat melakukannya seperti ini: $ git clone https://github.com/libgit2/libgit2 Ini membuat sebuah direktori bernama ``libgit2``, menginisialisasi direktori .git di dalamnya, menarik semua data untuk repositori tersebut, dan memeriksa salinan tugas versi terbaru. Jika Anda masuk ke direktori libgit2 yang baru, Anda akan melihat file project di sana, siap untuk dikerjakan atau digunakan. Jika Anda ingin menduplikat repositori ke dalam direktori yang bernama sesuatu selain ``libgit2``, Anda dapat menentukannya sebagai opsi baris command berikutnya: $ git clone https://github.com/libgit2/libgit2 mylibgit Command itu melakukan hal yang sama seperti yang sebelumnya, tapi direktori targetnya disebut 'mylibgit`. Git memiliki sejumlah protokol transfer yang berbeda yang dapat Anda gunakan. Contoh sebelumnya menggunakan protokol https:// , tetapi Anda juga bisa melihat git:// atau user@server:path/to/repo.git , yang menggunakan protokol transfer SSH. Getting Git on a Server akan memperkenalkan semua pilihan yang tersedia yang bisa diatur server untuk mengakses repositori Git Anda dengan membahas kelebihan dan kekurangan masing-masing pilihan. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/pl/v2/Git-na-serwerze-GitWeb
Git - GitWeb About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pierwsze kroki 1.1 Wprowadzenie do kontroli wersji 1.2 Krótka historia Git 1.3 Podstawy Git 1.4 Linia poleceń 1.5 Instalacja Git 1.6 Wstępna konfiguracja Git 1.7 Uzyskiwanie pomocy 1.8 Podsumowanie 2. Podstawy Gita 2.1 Pierwsze repozytorium Gita 2.2 Rejestrowanie zmian w repozytorium 2.3 Podgląd historii rewizji 2.4 Cofanie zmian 2.5 Praca ze zdalnym repozytorium 2.6 Tagowanie 2.7 Aliasy 2.8 Podsumowanie 3. Gałęzie Gita 3.1 Czym jest gałąź 3.2 Podstawy rozgałęziania i scalania 3.3 Zarządzanie gałęziami 3.4 Sposoby pracy z gałęziami 3.5 Gałęzie zdalne 3.6 Zmiana bazy 3.7 Podsumowanie 4. Git na serwerze 4.1 Protokoły 4.2 Uruchomienie Git na serwerze 4.3 Generowanie Twojego publicznego klucza SSH 4.4 Konfigurowanie serwera 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Inne opcje hostowania przez podmioty zewnętrzne 4.10 Podsumowanie 5. Rozproszony Git 5.1 Rozproszone przepływy pracy 5.2 Wgrywanie zmian do projektu 5.3 Utrzymywanie projektu 5.4 Podsumowanie 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Narzędzia Gita 7.1 Wskazywanie rewizji 7.2 Interaktywne używanie przechowali 7.3 Schowek i czyszczenie 7.4 Signing Your Work 7.5 Searching 7.6 Przepisywanie historii 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugowanie z Gitem 7.11 Moduły zależne 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Podsumowanie 8. Dostosowywanie Gita 8.1 Konfiguracja Gita 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git i inne systemy 9.1 Git jako klient 9.2 Migracja do Gita 9.3 Podsumowanie 10. Mechanizmy wewnętrzne w Git 10.1 Komendy typu plumbing i porcelain 10.2 Obiekty Gita 10.3 Referencje w Git 10.4 Spakowane pliki (packfiles) 10.5 Refspec 10.6 Protokoły transferu 10.7 Konserwacja i odzyskiwanie danych 10.8 Environment Variables 10.9 Podsumowanie A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.7 Git na serwerze - GitWeb GitWeb Teraz, gdy już podstawy odczytu i zapisu są dostępne tylko dla Twojego projektu, możesz założyć prostą internetową wizualizacje. Do tego celu Git wyposażony jest w skrypt CGI o nazwie GitWeb Figure 49. Internetowy interfejs użytkownika GitWeb Jeśli chcesz zobaczyć jak GitWeb będzie wyglądał dla Twojego projektu, Git posiada polecenie do uruchamiania tymczasowej instancji, pod warunkiem, że posiadasz lekki serwer taki jak lighttpd lub webrick . Na komputerach z zainstalowanym linuxem lighttpd jest bardzo często instalowany więc należy go uruchomić wpisując git instaweb w katalogu projektu. Jeśli używasz komputera Mac, Leopard jest automatycznie instalowany z Ruby więc webrick może być najlepszym rozwiązaniem. Aby rozpocząć instaweb bez tymczasowej instancji, należy uruchomić go z opcją --httpd . $ git instaweb --httpd=webrick [2009-02-21 10:02:21] INFO WEBrick 1.3.1 [2009-02-21 10:02:21] INFO ruby 1.8.6 (2008-03-03) [universal-darwin9.0] Powyższe polecenie uruchamia serwer HTTPD na porcie 1234, a następnie automatycznie uruchamia przeglądarkę internetową, która otwiera się na tej stronie. Kiedy skończysz i chcesz wyłączyć serwer, użyj tego samego polecenia z opcją --stop : $ git instaweb --httpd=webrick --stop Jeśli chcesz aby uruchomiony interfejs WWW był cały czas dostępny dla Twojego zespołu lub projektu open source, będziesz musiał skonfigurować skrypt CGI dla normalnego serwera WWW. Niektóre dystrybucje linuxa mają pakiet gitweb , który można zainstalować przez apt lub yum , więc warto spróbować tego w pierwszej kolejności. Jeśli się nie uda to musimy zainstalować GitWeb ręcznie, co trwa tylko chwile. Najpierw musimy pobrać kod źródłowy GitWeb i wygenerować własny skrypt CGI: $ git clone git://git.kernel.org/pub/scm/git/git.git $ cd git/ $ make GITWEB_PROJECTROOT="/opt/git" prefix=/usr gitweb SUBDIR gitweb SUBDIR ../ make[2]: `GIT-VERSION-FILE' is up to date. GEN gitweb.cgi GEN static/gitweb.js $ sudo cp -Rf gitweb /var/www/ Zwróć uwagę że musisz ręcznie podać lokalizacje swoich repozytoriów gita w zmiennej GITWEB_PROJECTROOT . Następnie należy stworzyć serwer Apache używający skryptu CGI, na którym można dodać wirtualnego hosta: <VirtualHost *:80> ServerName gitserver DocumentRoot /var/www/gitweb <Directory /var/www/gitweb> Options ExecCGI +FollowSymLinks +SymLinksIfOwnerMatch AllowOverride All order allow,deny Allow from all AddHandler cgi-script cgi DirectoryIndex gitweb.cgi </Directory> </VirtualHost> GitWeb można używać z dowolnym serwerem CGI. Jeśli wolisz korzystać z czegoś innego to nie powinno być trudne do skonfigurowania. W tym momencie powinieneś być w stanie odwiedzić http://gitserver/ w celu przeglądania repozytoriów online. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/commands/cargo-publish.html#name
cargo publish - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book cargo-publish(1) NAME cargo-publish — Upload a package to the registry SYNOPSIS cargo publish [ options ] DESCRIPTION This command will create a distributable, compressed .crate file with the source code of the package in the current directory and upload it to a registry. The default registry is https://crates.io . This performs the following steps: Performs a few checks, including: Checks the package.publish key in the manifest for restrictions on which registries you are allowed to publish to. Create a .crate file by following the steps in cargo-package(1) . Upload the crate to the registry. The server will perform additional checks on the crate. The client will poll waiting for the package to appear in the index, and may timeout. In that case, you will need to check for completion manually. This timeout does not affect the upload. This command requires you to be authenticated using cargo-login(1) or environment variables of the registry.token and registries.<name>.token config fields. See the reference for more details about packaging and publishing. OPTIONS Publish Options --dry-run Perform all checks without uploading. --no-verify Don’t verify the contents by building them. --allow-dirty Allow working directories with uncommitted VCS changes to be packaged. --index index The URL of the registry index to use. --registry registry Name of the registry to publish to. Registry names are defined in Cargo config files . If not specified, and there is a package.publish field in Cargo.toml with a single registry, then it will publish to that registry. Otherwise it will use the default registry, which is defined by the registry.default config key which defaults to crates-io . Package Selection By default, when no package selection options are given, the packages selected depend on the selected manifest file (based on the current working directory if --manifest-path is not given). If the manifest is the root of a workspace then the workspaces default members are selected, otherwise only the package defined by the manifest will be selected. The default members of a workspace can be set explicitly with the workspace.default-members key in the root manifest. If this is not set, a virtual workspace will include all workspace members (equivalent to passing --workspace ), and a non-virtual workspace will include only the root crate itself. -p spec … --package spec … Publish only the specified packages. See cargo-pkgid(1) for the SPEC format. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. --workspace Publish all members in the workspace. --all Deprecated alias for --workspace . --exclude SPEC … Exclude the specified packages. Must be used in conjunction with the --workspace flag. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. Compilation Options --target triple Publish for the specified target architecture. Flag may be specified multiple times. The default is the host architecture. The general format of the triple is <arch><sub>-<vendor>-<sys>-<abi> . Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. This may also be specified with the build.target config value . Note that specifying this flag makes Cargo run in a different mode where the target artifacts are placed in a separate directory. See the build cache documentation for more details. --target-dir directory Directory for all generated artifacts and intermediate files. May also be specified with the CARGO_TARGET_DIR environment variable, or the build.target-dir config value . Defaults to target in the root of the workspace. Feature Selection The feature flags allow you to control which features are enabled. When no feature options are given, the default feature is activated for every selected package. See the features documentation for more details. -F features --features features Space or comma separated list of features to activate. Features of workspace members may be enabled with package-name/feature-name syntax. This flag may be specified multiple times, which enables all specified features. --all-features Activate all available features of all selected packages. --no-default-features Do not activate the default feature of the selected packages. Manifest Options --manifest-path path Path to the Cargo.toml file. By default, Cargo searches for the Cargo.toml file in the current directory or any parent directory. --locked Asserts that the exact same dependencies and versions are used as when the existing Cargo.lock file was originally generated. Cargo will exit with an error when either of the following scenarios arises: The lock file is missing. Cargo attempted to change the lock file due to a different dependency resolution. It may be used in environments where deterministic builds are desired, such as in CI pipelines. --offline Prevents Cargo from accessing the network for any reason. Without this flag, Cargo will stop with an error if it needs to access the network and the network is not available. With this flag, Cargo will attempt to proceed without the network if possible. Beware that this may result in different dependency resolution than online mode. Cargo will restrict itself to crates that are downloaded locally, even if there might be a newer version as indicated in the local copy of the index. See the cargo-fetch(1) command to download dependencies before going offline. May also be specified with the net.offline config value . --frozen Equivalent to specifying both --locked and --offline . --lockfile-path PATH Changes the path of the lockfile from the default ( <workspace_root>/Cargo.lock ) to PATH . PATH must end with Cargo.lock (e.g. --lockfile-path /tmp/temporary-lockfile/Cargo.lock ). Note that providing --lockfile-path will ignore existing lockfile at the default path, and instead will either use the lockfile from PATH , or write a new lockfile into the provided PATH if it doesn’t exist. This flag can be used to run most commands in read-only directories, writing lockfile into the provided PATH . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #14421 ). Miscellaneous Options -j N --jobs N Number of parallel jobs to run. May also be specified with the build.jobs config value . Defaults to the number of logical CPUs. If negative, it sets the maximum number of parallel jobs to the number of logical CPUs plus provided value. If a string default is provided, it sets the value back to defaults. Should not be 0. --keep-going Build as many crates in the dependency graph as possible, rather than aborting the build on the first one that fails to build. For example if the current package depends on dependencies fails and works , one of which fails to build, cargo publish -j1 may or may not build the one that succeeds (depending on which one of the two builds Cargo picked to run first), whereas cargo publish -j1 --keep-going would definitely run both builds, even if the one run first fails. Display Options -v --verbose Use verbose output. May be specified twice for “very verbose” output which includes extra output such as dependency warnings and build script output. May also be specified with the term.verbose config value . -q --quiet Do not print cargo log messages. May also be specified with the term.quiet config value . --color when Control when colored output is used. Valid values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. May also be specified with the term.color config value . Common Options + toolchain If Cargo has been installed with rustup, and the first argument to cargo begins with + , it will be interpreted as a rustup toolchain name (such as +stable or +nightly ). See the rustup documentation for more information about how toolchain overrides work. --config KEY=VALUE or PATH Overrides a Cargo configuration value. The argument should be in TOML syntax of KEY=VALUE , or provided as a path to an extra configuration file. This flag may be specified multiple times. See the command-line overrides section for more information. -C PATH Changes the current working directory before executing any specified operations. This affects things like where cargo looks by default for the project manifest ( Cargo.toml ), as well as the directories searched for discovering .cargo/config.toml , for example. This option must appear before the command name, for example cargo -C path/to/my-project build . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #10098 ). -h --help Prints help information. -Z flag Unstable (nightly-only) flags to Cargo. Run cargo -Z help for details. ENVIRONMENT See the reference for details on environment variables that Cargo reads. EXIT STATUS 0 : Cargo succeeded. 101 : Cargo failed to complete. EXAMPLES Publish the current package: cargo publish SEE ALSO cargo(1) , cargo-package(1) , cargo-login(1)
2026-01-13T09:29:19
https://il.linkedin.com/company/fhenix?trk=organization_guest_main-feed-card-text#main-content
Fhenix | LinkedIn דילוג לתוכן הראשי ‏LinkedIn מאמרים אנשים למידה משרות משחקים התחברות הצטרפות Fhenix Software Development Programmable. Scalable. Encrypted. מעקב הצגת ‏כל ‏54‏ העובדים‏ דווח על חברה זו עלינו Fhenix extends Ethereum’s capabilities with Fully Homomorphic Encryption (FHE) Coprocessors. Leveraging computation over encrypted data, Fhenix gives blockchain developers the power to build end-to-end encrypted applications for the first time אתר אינטרנט https://www.fhenix.io/ קישור חיצוני עבור Fhenix תעשייה Software Development גודל החברה 11-50 עובדים משרדים ראשיים Tel Aviv סוג בבעלות פרטית הקמה 2023 מיקומים הראשי Tel Aviv, IL קבלת הוראות הגעה עובדים ב- Fhenix Guy Itzhaki Doron Zarchy Tovi Levis Giancarlo Di Bella צפייה בכל עובדים עדכונים ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Optalysys ‏‏3,604‏ עוקבים‏ ‏2‏חודש דיווח על פוסט הזה DevConnect ARG is loading... ↻ and we're proud to be partnering with Fhenix for Encryption Day. Buenos Aires, here we come #privacybydesign Our sponsor cohort: Arbitrum | Ethereum Foundation | Horizen | Artifact Systems | Hyli | Silhouette Eyes on for more details coming soon... X 🔗 https://lnkd.in/e4uRuzHh #devconnect #encryptionday #privacybydesign 15 ‏1‏ ‏תגובה‏ לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש נערך דיווח על פוסט הזה Encryption Day Returns November 19! 👩💻 Live coding sessions and demos from the Ethereum Foundation , Fhenix , and more from 10:00–12:00 🎤 Panels and presentations from some of the brightest minds in the Ethereum ecosystem all day! ( RAILGUN DAO , MetaCartel (metacartel.org) 1kx , + more) 🍪 Argentinian treats and good company! 🔗 Get your ticket and share with friends! https://luma.com/e9ces3lr 🙌 Brought to you by Optalysys , Artifact Staking , Hyli , zkVerify , Arbitrum , & Silhouette Exchange ! Encryption Day · Luma luma.com 5 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש דיווח על פוסט הזה Every ecosystem has a moment where privacy stops being optional. For Japan, that moment starts now... Japan Joins the Encrypted Frontier: Our Strategic Raise with BIPROGY & TransLink Capital ‏Fhenix‏ ב-LinkedIn 24 ‏2‏ ‏תגובות‏ לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש נערך דיווח על פוסט הזה Come hack at ETHSamba 's ETHLatam Hackathon, with a total bounty of 25k+ USD in prizes! 📅 November 6 - 8 🔗 https://lnkd.in/gq9rJrkR ETHSamba ‏‏466‏ עוקבים‏ ‏2‏חודש נערך Privacidade não morreu — está sendo reprogramada. Com computação criptografada para Ethereum, a Fhenix traz lógica privada por padrão. 20 000× mais throughput, 37× menos latência — performance digna do playground fintech que o Brasil se tornou. Bora pra ETH LATAM 4 לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Optalysys ‏‏3,604‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה 🗞️ In the news: FHE research from the Fhenix camp! Guy Itzhaki CEO, Fhenix; Bringing Encryption to Ethereum; @GuyItzh ‏3‏חודש Today marks a significant milestone for both Fhenix and the broader FHE community — our newest research on highly efficient FHE decryption is now live. Why does this matter? Because it represents a breakthrough in how FHE decryption can scale securely and efficiently, paving the way for real-world, high-throughput confidential computation. The challenge we solved: For years, delivering a secure and efficient FHE decryption method has been one of the hardest open problems in the field. Existing approaches were slow, costly, or required significant computational resources, making them impractical for DeFi-scale workloads. Our new Threshold FHE Decryption protocol eliminates these limitations — delivering both security and speed at levels previously thought unattainable. We achieved: • 20,000× higher throughput • 37× lower latency • Real-time encrypted computation — unlocking Private DeFi Next Week, we will present our paper, “High-Throughput Universally Composable Threshold FHE Decryption,” at ACM CCS 2025 — one of the world’s top cryptography conferences. If you're attending, be sure to hear Guy Zyskind talk. Link to the paper: https://lnkd.in/gPkM3RAd High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 9 לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Guy Itzhaki ‏3‏חודש דיווח על פוסט הזה Today marks a significant milestone for both Fhenix and the broader FHE community — our newest research on highly efficient FHE decryption is now live. Why does this matter? Because it represents a breakthrough in how FHE decryption can scale securely and efficiently, paving the way for real-world, high-throughput confidential computation. The challenge we solved: For years, delivering a secure and efficient FHE decryption method has been one of the hardest open problems in the field. Existing approaches were slow, costly, or required significant computational resources, making them impractical for DeFi-scale workloads. Our new Threshold FHE Decryption protocol eliminates these limitations — delivering both security and speed at levels previously thought unattainable. We achieved: • 20,000× higher throughput • 37× lower latency • Real-time encrypted computation — unlocking Private DeFi Next Week, we will present our paper, “High-Throughput Universally Composable Threshold FHE Decryption,” at ACM CCS 2025 — one of the world’s top cryptography conferences. If you're attending, be sure to hear Guy Zyskind talk. Link to the paper: https://lnkd.in/gPkM3RAd High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 33 ‏4‏ ‏תגובות‏ לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Blockchain Futurist Conference ‏‏6,284‏ עוקבים‏ ‏3‏חודש נערך דיווח על פוסט הזה ⚡ Your newly added speakers are here! ⚡ 👇 Just 5 weeks until we hear from these Web3 & AI legends 👇 ◾ Iggy Azalea, Rapper & Songwriter ◾ Jake Antifaev , Co-Founder & CEO, Thrust ◾ Webster Barnaby , Representative for HD-29, Florida House of Representatives ◾ Abhay M., SVP of Corporate Development and Investment, MoonPay ◾ André Beganski , Staff Reporter, Decrypt ◾ AnewbiZ 0x , Chief of Operations, Secret Network Foundation ◾ Jan Hans Georg Pachner , Secretary General, One Ocean Foundation ◾ Lauren Dutton , Devrel, Fhenix ◾ Richard Ells , CEO, Electroneum ◾ Lucía Gallardo , Founder, Emerge , Co-Founder, AeraTech ◾ Elina Litvak Cohen, Esq. , General Counsel, QuickNode ◾ Kyle K. , Co-Founder, Blockstreet™ Let's go! Web3, AI, Crypto, DeFi, RWAs, NFTs, Privacy, and more under one roof. November 5-6 at the Hard Rock Guitar and DAER. More announcements coming... see you all soon! 🎟️ https://lnkd.in/dhuz8gu 53 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה Three things just changed: 1/ FHE decryption got 20,000× faster 2/ Private DeFi is unlocked 3/ Fhenix made it happen Guy Zyskind 's paper detailing Fhenix's advancements in Threshold FHE Decryption has been featured ACM Special Interest Group on Security, Audit and Control (SIGSAC) 's 2025 conference. This acknowledgment places Fhenix’s research alongside the work of Microsoft, Google, Meta, Stanford, and MIT — validating that our progress isn’t just fast… it’s foundational. View the paper: https://lnkd.in/gbWuFihV Recap: https://lnkd.in/gtAzuPa7 High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 6 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה Join Guy Itzhaki amongst others tomorrow 9am ET / 13:00 UTC for a roundtable discussion on confidential AI! 🔗: https://lnkd.in/gVNgBFeR Optalysys ‏‏3,604‏ עוקבים‏ ‏3‏חודש This week: 𝗖𝗘𝗢𝘀 𝗗𝗲𝗰𝗼𝗱𝗲 𝗔𝗜 𝗖𝗼𝗻𝗳𝗶𝗱𝗲𝗻𝘁𝗶𝗮𝗹: 𝗭𝗞, 𝗧𝗘𝗘, 𝗙𝗛𝗘 | an X Space Roundtable Matt Garrow-Fisher will moderate this unmissable CEO roundtable with: Nick New – CEO, Optalysys Guy Itzhaki – CEO, Fhenix Robert Viglione – CEO, @HorizenLabs Marvin Tong – CEO, @PhalaNetwork 📌 Be there: https://lnkd.in/eBXugGXc 🗓️ Add to your calendar now: Wed, Sept 17 - 9AM ET | 2PM BST #xspace #roundtable #confidentialai 4 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏4‏חודש דיווח על פוסט הזה Join our workshop on confidential payments this Friday, September 12 at the Crecimiento Argentina Aleph Hub in Buenos Aires, Argentina! 🎟️ Register: https://lnkd.in/etUg-MG3 Cámara Argentina Fintech ‏‏43,907‏ עוקבים‏ ‏4‏חודש נערך 🚀 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 𝗦𝘂𝗺𝗺𝗶𝘁 | 𝗦𝘂𝗺𝗮𝘁𝗲 𝗮 𝘂𝗻 𝗲𝘃𝗲𝗻𝘁𝗼 𝗽𝗮𝗿𝗮 𝗰𝗼𝗻𝗼𝗰𝗲𝗿 𝗺á𝘀 𝘀𝗼𝗯𝗿𝗲 𝗹𝗮 𝘂𝗻𝗶ó𝗻 𝗲𝗻𝘁𝗿𝗲 𝗹𝗮𝘀 𝗳𝗶𝗻𝗮𝗻𝘇𝗮𝘀 𝘁𝗿𝗮𝗱𝗶𝗰𝗶𝗼𝗻𝗮𝗹𝗲𝘀, 𝗯𝗹𝗼𝗰𝗸𝗰𝗵𝗮𝗶𝗻 𝘆 𝗪𝗲𝗯𝟯 Desde la 𝗖á𝗺𝗮𝗿𝗮 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 los invitamos a ser parte del 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 𝗦𝘂𝗺𝗺𝗶𝘁, un encuentro diseñado para explorar la intersección entre finanzas tradicionales, fintechs y proyectos Web3. Junto a Crecimiento, reuniremos a bancos, fintechs y startups para debatir sobre el 𝗳𝘂𝘁𝘂𝗿𝗼 𝗱𝗲𝗹 𝗱𝗶𝗻𝗲𝗿𝗼 𝗲𝗻 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮: desde open finance y pagos transfronterizos hasta IA, infraestructura y seguridad financiera. 📅 𝗩𝗶𝗲𝗿𝗻𝗲𝘀 𝟭𝟮 𝗱𝗲 𝘀𝗲𝗽𝘁𝗶𝗲𝗺𝗯𝗿𝗲 - 𝟵 𝗮 𝟭𝟳:𝟯𝟬 𝗵𝘀 📍 𝗖𝗼𝗻𝗰𝗲𝗽𝗰𝗶𝗼́𝗻 𝗔𝗿𝗲𝗻𝗮𝗹 𝟮𝟵𝟴𝟵, 𝗖𝗔𝗕𝗔 🗓️ 𝗔𝗴𝗲𝗻𝗱𝗮 𝗰𝗼𝗻𝗳𝗶𝗿𝗺𝗮𝗱𝗮 🔹𝟵:𝟯𝟬 - 𝗥𝗲𝗴𝗶𝘀𝘁𝗿𝗼 𝘆 𝗱𝗲𝘀𝗮𝘆𝘂𝗻𝗼 🔹𝟭𝟬:𝟭𝟱 - 𝗞𝗲𝘆𝗻𝗼𝘁𝗲 𝗱𝗲 𝗮𝗽𝗲𝗿𝘁𝘂𝗿𝗮 ( Crecimiento Argentina ) 🔹 𝟭𝟬:𝟯𝟬 - 𝗣𝗮𝗻𝗲𝗹: 𝗢𝗽𝗲𝗻 𝗙𝗶𝗻𝗮𝗻𝗰𝗲 Moderador: Teófilo Beato (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Eduardo Segovia Mattos (Open Finance) y Nicolas Ballare Scopel, CIIA ( MODO ). 🔹𝟭𝟭:𝟮𝟬 - 𝗣𝗮𝗻𝗲𝗹: 𝗣𝗮𝗴𝗼𝘀 𝘁𝗿𝗮𝗻𝘀𝗳𝗿𝗼𝗻𝘁𝗲𝗿𝗶𝘇𝗼𝘀 Moderador: Diego Martin Diaz (CAF). Speakers: Federico Martin Testoni (Depay), Henri de Royere ( Due ) y Catalina Ramallo ( DolarApp ). 🔹𝟭𝟮:𝟭𝟬 - 𝗪𝗼𝗿𝗸𝘀𝗵𝗼𝗽: 𝗖𝗼𝗻𝗳𝗶𝗱𝗲𝗻𝘁𝗶𝗮𝗹 𝗣𝗮𝘆𝗺𝗲𝗻𝘁𝘀 Moderador: Simon Puebla (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Fhenix . 🔹𝟭𝟰:𝟭𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗜𝗻𝗳𝗿𝗮𝗲𝘀𝘁𝗿𝘂𝗰𝘁𝘂𝗿𝗮 Moderador: Teófilo Beato (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Edwin A. Rodriguez ( Rain ) y Santiago Witis  ( Pomelo ). 🔹𝟭𝟱:𝟬𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗜𝗔 𝗲𝗻 𝘀𝗲𝗿𝘃𝗶𝗰𝗶𝗼𝘀 𝗳𝗶𝗻𝗮𝗻𝗰𝗶𝗲𝗿𝗼𝘀 Moderadora: Natalia Diaz (CAF). Speakers: Lucila Destefanis ( Satoshi Tango ), Juan Ignacio Segura ( PwC Argentina ) y Paula Barnes, CAMS (𝗣𝗼𝗺𝗲𝗹𝗼). 🔹𝟭𝟱:𝟱𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗧𝗮𝗿𝗷𝗲𝘁𝗮𝘀 Moderador: Julian Morla ( LI.FI ). Speakers: Patricio Molina ( belo ), Tobias Frieder ( Offramp Labs ) y Juan Manuel Brunetti ( Prex ). 🔹𝟭𝟲:𝟰𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗦𝗲𝗴𝘂𝗿𝗶𝗱𝗮𝗱 𝗰𝗼𝗺𝗼 𝘃𝗲𝗻𝘁𝗮𝗷𝗮 𝗰𝗼𝗺𝗽𝗲𝘁𝗶𝘁𝗶𝘃𝗮 Moderador: Matias Fermin (CAF). Speakers: Ivan Lendner ( Whalemate ), Marcelo Triviño ( Ualá ), Lucas Paus ( BIND | Soluciones financieras ) y Santiago Agustín Fernandez (Klar). 🔹𝟭𝟳:𝟯𝟬 - 𝗖𝗶𝗲𝗿𝗿𝗲 🔗 Registrate acá: https://lnkd.in/etUg-MG3 Será una jornada para aprender, compartir y conectar con quienes están construyendo el próximo capítulo del ecosistema financiero argentino. ¡Los esperamos! לייק תגובה שיתוף הצטרף עכשיו כדי לראות מה חסר לך מצא אנשים שאתה מכיר בהם Fhenix עיין מקומות תעסוקה מומלצים עבורך הצג את כל העדכונים, חֲדָשׁוֹת והמאמרים הצטרף עכשיו דפים דומים SCRT Labs Software Development ‏.‏, ‏.‏ Zama Software Development Xai Games Blockchain Services medflyt Software Development ‏Brooklyn‏, ‏New York‏ StarkWare Software Development ‏Netanya‏, ‏Hasharon‏ Optalysys Computer Hardware Manufacturing ‏Leeds‏, ‏England‏ Arbitrum Technology, Information and Internet ‏George Town‏, ‏Grand Cayman‏ Utila Software Development ‏New York City‏, ‏New York‏ Zengo Wallet Financial Services ‏Tel Aviv‏, ‏NA‏ Polygon Labs Technology, Information and Internet הצג עוד דפים דומים הצג פחות דפים דומים מימון Fhenix 4 total rounds סיבוב אחרון סדרה לא ידועה 22 בנוב׳ 2025 External Crunchbase Link לסבב הגיוס האחרון משקיעים Biprogy Translink Capital ראה מידע נוסף על crunchbase ‏LinkedIn © 2026 על אודות נגישות הסכם המשתמש מדיניות פרטיות מדיניות עוגיות מדיניות זכויות יוצרים מדיניות המותג בקרת אורחים הנחיות הקהילה العربية (ערבית) ‏বাংলা (בנגלה) ‏Čeština (צ’כית) ‏Dansk (דנית) ‏Deutsch (גרמנית) ‏Ελληνικά (יוונית) ‏English (אנגלית) ‏Español (ספרדית) فارسی (פרסית) ‏Suomi (פינית) ‏Français (צרפתית) ‏हिंदी (הינדי) ‏Magyar (הונגרית) ‏Bahasa Indonesia (אינדונזית) ‏Italiano (איטלקית) עברית (Hebrew) ‏日本語 (יפנית) ‏한국어 (קוריאנית) ‏मराठी (מראטהית) ‏Bahasa Malaysia (מלאית) ‏Nederlands (הולנדית) ‏Norsk (נורווגית) ‏ਪੰਜਾਬੀ (פנג’אבי) ‏Polski (פולנית) ‏Português (פורטוגזית) ‏Română (רומנית) ‏Русский (רוסית) ‏Svenska (שוודית) ‏తెలుగు (טלוגו) ‏ภาษาไทย (תאילנדית) ‏Tagalog (טאגאלוג) ‏Türkçe (טורקית) ‏Українська (אוקראינית) ‏Tiếng Việt (וייטנאמית) ‏简体中文 (סינית (מפושטת)) ‏正體中文 (סינית (מסורתית)) שפה מסכים ומצטרף LinkedIn הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn. היכנס כדי לראות את מי שאתה כבר מכיר בכתובת Fhenix התחברות טוב לראות אותך שוב דוא”ל או טלפון סיסמה הצגה שכחת סיסמא? התחברות או הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn. פעם ראשונה שלך ב-LinkedIn? ‏ להצטרפות ‏ או פעם ראשונה שלך ב-LinkedIn? ‏ להצטרפות ‏ הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn.
2026-01-13T09:29:19
https://www.linkedin.com/products/categories/mainframe-management-software?trk=organization_guest_product_card_category_link
Best Mainframe Management Software | Products | LinkedIn Skip to main content LinkedIn Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Clear text Used by Used by Mainframe Developer (10) Mainframe Programmer (10) Mainframe Operator (9) Mainframe Consultant (9) Mainframe System Programmer (9) See all products Find top products in Mainframe Management Software category Software used to oversee the operations of large mainframe systems. - Analyze performance in real time - Automatically identify, repair, and prevent errors - Maintain database security - Perform maintenance without service interruption 51 results HCL Mainframe Solutions Mainframe Management Software by HCLSoftware Essential solutions for mainframe users to optimize, modernize, and enable Innovation with their mainframe investments. • 𝐙 𝐀𝐬𝐬𝐞𝐭 𝐎𝐩𝐭𝐢𝐦𝐢𝐳𝐞𝐫 (𝐙𝐀𝐎) 𝐚𝐧𝐝 𝐙 𝐋𝐢𝐜𝐞𝐧𝐬𝐞 𝐕𝐞𝐫𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧 (𝐙𝐋𝐕): Discover, manage, and optimize mainframe software assets, and simplify mainframe software compliance. • 𝐙 𝐚𝐧𝐝 𝐈 𝐄𝐦𝐮𝐥𝐚𝐭𝐨𝐫 (𝐙𝐈𝐄) 𝐅𝐚𝐦𝐢𝐥𝐲: Emulation and Transformation solution for mainframe customers. • 𝐙 𝐃𝐚𝐭𝐚 𝐓𝐨𝐨𝐥𝐬 (𝐙𝐃𝐓) & 𝐙 𝐀𝐛𝐞𝐧𝐝 𝐈𝐧𝐯𝐞𝐬𝐭𝐢𝐠𝐚𝐭𝐨𝐫 (𝐙𝐀𝐈): Z Data tools for Data Access Modernization and Automated Abend Diagnostic. View product BMC AMI (Mainframe Transformation) Mainframe Management Software by BMC Software BMC Automated Mainframe Intelligence (AMI) solutions leverage AI, machine learning, and predictive analytics to achieve a self-managing mainframe. Take advantage of: - AMI Cloud - AMI Data for DB2 - AMI Data for IMS - AMI DevX - AMI Ops - AMI Security View product Vertali ZTRUST for SMP/e Mainframe Management Software by Vertali To help you to better protect your systems and data, ZTRUST® for SMP/e is a browser-based security and compliance solution that eases the installation and maintenance of mainframe software and subsystems. - Browser-based user interface – ease of use, enabling greater visibility - Automatic download of Authorized Program Analysis Reports (APAR) - Automatic detection of Consolidated Software Inventories (CSIs) - Centralised repository provided for all products and product features - Automatic product detection - Reports and alerts ZTRUST® for SMP/e means you can interrogate SMP/e and gain a centralised view of the status and applicability of all critical APARs across your entire mainframe estate. View product Vertali ZTRUST Password Synchronisation Mainframe Management Software by Vertali Enabling password/pass phrase synchronisation for multiple ESMs ZTRUST® Password Synchronisation is a security utility that eases the process of synchronizing password changes across multiple Enterprise Security Management (ESM) systems in real time. This software is particularly useful for sites with hybrid ESM products such as RACF, Top Secret and/or ACF2. - Links incompatible ESMs - Automatically detects password/pass phrase changes in real-time - Automatically forwards changes to remote ESM systems - Industry-standard TLS encryption in addition to internal ZTRUST® encryption - Enables single and bi-directional synchronisation - Agent and Master architecture to reduce system definitions and cross-sysplex connections - Secure repository to temporarily queue changes - Whitelist/blacklist configuration options View product Tectia® SSH Server for IBM z/OS Mainframe Management Software by SSH Communications Security QUANTUM-SAFE MAINFRAME SECURITY & DATA COMMUNICATIONS SOFTWARE Tectia SSH Server for IBM z/OS is the most trusted remote access and secure FTP (SFTP) server software in the industry. It's easy to set up, combines enterprise-grade reliability with high performance and a light toll on cryptographic processing, and comes with quantum-safe capabilities. Key benefits: > Quantum-safe encryption used during all file transfers > Eliminates the need to create temporary staging files between processes > Transfer output from the first process to the next process via Inter-process communication > Preserving and restoring dataset metadata when transferring between mainframes or distributed systems > Selecting desired parts of a dataset rather than dumping the whole dataset > Automatic options for codepage translations and dataset definitions, avoiding costly staging and pre/post-processing steps Note: Supported on IBM z/OS versions 2.2, 2.3, 2.4, 2.5 and 3.1 View product Find products trusted by professionals in your network See which products are used by connections in your network and those that share similar job titles Sign in to view full insights Lozen™ Connector for MuleSoft Mainframe Management Software by VirtualZ Computing Part of our mission is to ensure mainframe data is easy to access from as many types of applications as possible. While the Lozen Connector for MuleSoft can leverage Lozen OpenAPI capabilities, Lozen provides a specific connector for Mule that goes even further. Using Lozen’s Connector for MuleSoft: • Access to mainframe data is simplified, and data access can be designed directly within MuleSoft’s AnyPoint Studio. • You never leave the Mule development environment, making it easier to add Lozen data objects to your Mule workflows. • You can create complex workflows by dragging and dropping connectors from a palette of Mule connectors. • You have access to all of Lozen’s functionality by simply dragging the Lozen connector into your workspace. • Like our OpenAPI implementation, the Lozen Mule connector gives you ways to securely read or update any data managed by the Lozen server running on your z/OS platform. View product iCAN Mainframe Management Software by ISC International Software Company iCAN collects the definitions of your Production components in order to provide fast access to the information they contain. By identifying cross-references, links, and dependencies among the collected components, iCAN enables your teams to audit, analyze, and reverse-document your environments. The product’s query engine provides simplified access to information related to batch jobs, scripts, applications, JCL, scheduling definitions, data flows, file transfers, and any other information you may add to the collection sources. Each available display mode includes hyperlinks that allow users to navigate within the repository and quickly access the desired information. iCAN enables users—even non-technical ones—to easily take ownership of operational or functional information system knowledge. View product Beta JobZ - Enterprise Job Manager Mainframe Management Software by Beta Systems Software With Beta JobZ, our powerful job scheduling software for small to large businesses, you can conveniently connect all your distributed servers and the applications hosted there (e. g. SAP) with the mainframe. Our job scheduler tool lets you control your automated processes across all relevant platforms while saving costs at the same time. We support any mainframe scheduler and all common operating system platforms, such as Linux, Unix, Windows, IBM System i and IBM Z. Discover our job scheduling programs now! View product zTrust for Networks Mainframe Management Software by Vertali Vertali’s unique zTrust for Networks dramatically reduces the time, effort and costs involved in detecting, monitoring and optionally enforcing network access and provides a valuable tool for adhering to compliance standards and regulations. Network Discovery - Differentiating between clear and encrypted network connections is the key. zTrust software identifies: - Applications that are permanently or temporarily accepting inbound non-encrypted and/or inbound encrypted connections - Applications that are making outbound non encrypted and/or outbound encrypted connections - Network segments that are accessing specific applications without encryption and/or with encryption Network Micro-segmentation - A key mechanism to control network access and can often be a requirement for regulatory compliance such as PCI/DSS. e. View product E-GEN Mainframe Management Software by ISC International Software Company E-GEN effectively eliminates the difficulties associated with Production deployments by leveraging functional descriptions and knowledge of your rules and standards. E-GEN automatically generates all types of scripts, JCL, scheduling definitions, and other components to streamline application changes. Documentation production is also fully supported, from change request files to operational documentation. With E-GEN, your organization can achieve maximum efficiency in its Production deployment processes. Automatically generate and deploy your batch jobs and Operations components. View product See more How it works Explore Discover the best product for your need from a growing catalog of 25,000 products and categories trusted by LinkedIn professionals Learn Evaluate new tools, explore trending products in your industry and see who in your network is skilled in the product Grow Join communities of product users to learn best practices, celebrate your progress and accelerate your career LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English Language
2026-01-13T09:29:19
https://www.linkedin.com/posts/bmc-software_2025-highlights-wins-proof-points-and-activity-7411401863360741376-m7E8
2025 Highlights: Wins, Proof Points, and Momentum | BMC Software Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Skip to main content LinkedIn Top Content People Learning Jobs Games Sign in Join now for free BMC Software’s Post BMC Software 329,940 followers 2w Report this post 3 things worth stealing for your 2026 planning: - A mainframe modernization ROI blueprint - A customer orchestration playbook (Raymond James) - Two demo libraries you can explore in minutes Get it in The BMC Briefing: Edition 4 🔽 2025 Highlights: Wins, Proof Points, and Momentum BMC Software on LinkedIn 113 2 Comments Like Comment Share Copy LinkedIn Facebook X Noel Watson 2w Report this comment Thanks for sharing Like Reply 1 Reaction Bernard Denke 2w Report this comment Congrats! Like Reply 1 Reaction See more comments To view or add a comment, sign in 329,940 followers View Profile Connect More from this author 2025 Highlights: Wins, Proof Points, and Momentum BMC Software 2w Control-M Ranks #1 for the 8th Consecutive Year. What EMA and Gartner® Say About Orchestration BMC Software 1mo BMC Named a Leader in the 2025 Gartner® Magic Quadrant™ for SOAP BMC Software 2mo Explore content categories Career Productivity Finance Soft Skills & Emotional Intelligence Project Management Education Technology Leadership Ecommerce User Experience Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language Sign in to view more content Create your free account or sign in to continue your search Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy .
2026-01-13T09:29:19
https://web3.okx.com/sv/dex-swap
DEX Swap: Handla krypto med låga avgifter och hög likviditet | OKX Wallet Hoppa till huvudinnehåll Markn. Meme Pump Spårare Signal Topplista Swap Portfölj Värvning Mer Prod. DeFi Marknadsplats Säkerhet Version X Layer Läs mer om X Layer X Layer-utforskare Bro Utvecklardokumentation Testnätkran GitHub Wallet API Upptäck plånboks-API API-dokument Hantera API-nycklar Blockkedje-utforskare Anslut plånboken i DApps Boost X Launch Delta och tjäna tokens X-kampanj Gå med för att vinna spännande belöningar Belöningscenter Hämta bonusar och airdrops Avisering Språk Valuta Färginställningar Ladda ner OKX Wallet Allt om Web3 Tillkännagivanden Tillbaka Svenska 简体中文 繁體中文 English Tiếng Việt Русский Español (Latinoamérica) Bahasa Indonesia Français Deutsch Italiano Polski Čeština Română Português (Portugal) Português (Brasil) Українська Español (España) Nederlands العربية 日本語 Norsk (bokmål) Suomi Türkçe Tillbaka Tillbaka Lär dig Supportcenter Byte Bro Byte Vanliga frågor What is OKX DEX Swap? OKX DEX Swap is a decentralized exchange (DEX) aggregator that allows you to trade crypto seamlessly across multiple blockchains. Our X Routing algorithm finds the best prices by comparing liquidity pools, splitting orders, and optimizing for fees and slippage. Hur säkerställer OKX DEX de bästa swap-priserna? OKX DEX använder avancerade algoritmer för att samla likviditet från flera DEX och likviditetsleverantörer (LP). Genom att dela upp ordrar och ta hänsyn till faktorer som pris, prisförskjutning och nätverksavgifter ser vi till att du får bästa möjliga priser. Why do trades fail? 1. Insufficient network fee. During times of network congestion, network fees may skyrocket. Miners prioritize transactions with higher network fees, so if your transaction isn’t picked up by a miner for an extended period, it can result in transaction failure. Therefore, when initiating a transaction, it’s recommended to use the default values recommended by the platform. 2. Exceeded slippage. In volatile markets, liquidity depths can change. If the trading amount falls below the required minimum, the contract may halt the transaction, leading to a failed transaction due to exceeded slippage. 3.Duplicate transactions. If a user initiates multiple identical transactions simultaneously, but their balance is only sufficient to cover the first transaction, subsequent transactions will fail. What to do when I don’t have enough balance for the network fee? You can buy the corresponding tokens on the OKX Exchange and withdraw them to your wallet address, or swap tokens on OKX DEX to cover the network fees. What is slippage? Slippage refers to the price difference between the expected price of a trade and its actual execution price. Slippage is an implicit cost of trading and is inevitable in most trades, especially in a DEX. Slippage becomes more prominent during market volatility or when there is insufficient liquidity on the trading platform. How can slippage be avoided? Slippage cannot be completely avoided, but its impact can be minimized in three ways: 1. The best way is to choose assets with high trading volume and liquidity. 2. You can opt for transactions with smaller amounts, which are more likely to be executed, instead of placing large orders directly. 3. Consider adjusting the slippage tolerance during price fluctuations, where slippage may exceed the preset limit when your transaction is executed. In this case, increasing slippage could avoid transaction failure, but it also means you can't get a good price. Please proceed with caution. Why trading some tokens requires a higher slippage setting? Some tokens have a higher slippage tolerance due to the presence of transaction taxes, such as SAFEMOON, which are used for liquidity pool buy-back, token burning, and rewarding token holders. To help you trade these tokens, OKX DEX has launched automated slippage to improve the success rate of these trades at the current optimal slippage level. Växla till bästa pris Välj själv handelsvägar från alla stora DEX-aggregatorer, medan X Routing fortsätter att hitta de bästa anbuden från 400+ DEX:er och 30+ nätverk. Snabba anbud Anti-MEV ©2017 - 2026 WEB3.OKX.COM Svenska 简体中文 繁體中文 English Tiếng Việt Русский Español (Latinoamérica) Bahasa Indonesia Français Deutsch Italiano Polski Čeština Română Português (Portugal) Português (Brasil) Українська Español (España) Nederlands العربية 日本語 Norsk (bokmål) Suomi Türkçe More about OKX Wallet Ladda ned Learn Om oss Karriär Kontakta oss Användarvillkor Integritetsmeddelande X (tidigare Twitter) Product Tokens Swap Marketplace Earn Discover Build Explorer Säkerhet Support Supportcenter Kanalverifiering Tillkännagivanden DEX-avgiftsstruktur Anslut med OKX Bitcoin-plånbok Ethereum-plånbok Solana-plånbok
2026-01-13T09:29:19
https://www.linkedin.com/products/dataport-a%C3%B6r-ddatabox/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_image-click#main-content
dDataBox | LinkedIn Skip to main content LinkedIn Dataport AöR in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in dDataBox Managed File Transfer (MFT) Software by Dataport AöR See who's skilled in this Add as skill Request demo Report this product About Bei der Kommunikation mit unseren Kolleg*innen und Externen greifen wir täglich auf E-Mails zurück. Für den Versand sehr großer Dateien eignen sie sich allerdings nicht. Zudem bieten E-Mails nicht immer die erforderliche Vertraulichkeit. Mit der dDataBox können Sie Ihren Adressat*innen nun große Dateien sicher bereitstellen. Auf einen Blick: ✅ Ende-zu-Ende Verschlüsselung ✅ Smartphone-App (Android, iOS) ✅ Temporärer Down- und Uploadlink für Externe ✅ Benutzer*innen und Speichervolumen eigenständig verwalten ✅ Betrieb im hochsicheren Twin Data Center ✅ Günstiger Speicher This product is intended for Referent Head of Section Department Manager Clerical Officer Project Manager Information Technology Application Manager Process Manager Chief Information Security Officer Territory Manager Media Products media viewer No more previous content Kinderleicht und sicher Daten austauschen mit dDataBox Moderne Zusammenarbeit kennt keine Grenzen. Teams arbeiten heute organisationsübergreifend und unabhängig von Zeit und Ort. Dabei sind der schnelle und sichere Austausch aller relevanten Daten besonders wichtig. dDataBox kombiniert geräteunabhängige Zugriffsmöglichkeiten mit intuitiver, einfacher Handhabung. Vollverschlüsselung vom Endgerät über den Transport bis zum Datenspeicher gewährleistet die Sicherheit und Authentizität Ihrer Daten. dDataBox No more next content Similar products Axway Managed File Transfer Axway Managed File Transfer Managed File Transfer (MFT) Software Progress MOVEit Progress MOVEit Managed File Transfer (MFT) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software JSCAPE by Redwood JSCAPE by Redwood Managed File Transfer (MFT) Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software MLADU MLADU Managed File Transfer (MFT) Software Sign in to see more Show more Show less Dataport AöR products data[port]ai data[port]ai Data Science & Machine Learning Platforms Dataport Consulting Dataport Consulting Strategic Planning Software dMessenger dMessenger Enterprise Messaging Software dProjectTracking dProjectTracking Project Management Software dWebService dWebService Web Hosting dWorkflow dWorkflow Workflow Management Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/commands/cargo-publish.html#synopsis
cargo publish - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book cargo-publish(1) NAME cargo-publish — Upload a package to the registry SYNOPSIS cargo publish [ options ] DESCRIPTION This command will create a distributable, compressed .crate file with the source code of the package in the current directory and upload it to a registry. The default registry is https://crates.io . This performs the following steps: Performs a few checks, including: Checks the package.publish key in the manifest for restrictions on which registries you are allowed to publish to. Create a .crate file by following the steps in cargo-package(1) . Upload the crate to the registry. The server will perform additional checks on the crate. The client will poll waiting for the package to appear in the index, and may timeout. In that case, you will need to check for completion manually. This timeout does not affect the upload. This command requires you to be authenticated using cargo-login(1) or environment variables of the registry.token and registries.<name>.token config fields. See the reference for more details about packaging and publishing. OPTIONS Publish Options --dry-run Perform all checks without uploading. --no-verify Don’t verify the contents by building them. --allow-dirty Allow working directories with uncommitted VCS changes to be packaged. --index index The URL of the registry index to use. --registry registry Name of the registry to publish to. Registry names are defined in Cargo config files . If not specified, and there is a package.publish field in Cargo.toml with a single registry, then it will publish to that registry. Otherwise it will use the default registry, which is defined by the registry.default config key which defaults to crates-io . Package Selection By default, when no package selection options are given, the packages selected depend on the selected manifest file (based on the current working directory if --manifest-path is not given). If the manifest is the root of a workspace then the workspaces default members are selected, otherwise only the package defined by the manifest will be selected. The default members of a workspace can be set explicitly with the workspace.default-members key in the root manifest. If this is not set, a virtual workspace will include all workspace members (equivalent to passing --workspace ), and a non-virtual workspace will include only the root crate itself. -p spec … --package spec … Publish only the specified packages. See cargo-pkgid(1) for the SPEC format. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. --workspace Publish all members in the workspace. --all Deprecated alias for --workspace . --exclude SPEC … Exclude the specified packages. Must be used in conjunction with the --workspace flag. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. Compilation Options --target triple Publish for the specified target architecture. Flag may be specified multiple times. The default is the host architecture. The general format of the triple is <arch><sub>-<vendor>-<sys>-<abi> . Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. This may also be specified with the build.target config value . Note that specifying this flag makes Cargo run in a different mode where the target artifacts are placed in a separate directory. See the build cache documentation for more details. --target-dir directory Directory for all generated artifacts and intermediate files. May also be specified with the CARGO_TARGET_DIR environment variable, or the build.target-dir config value . Defaults to target in the root of the workspace. Feature Selection The feature flags allow you to control which features are enabled. When no feature options are given, the default feature is activated for every selected package. See the features documentation for more details. -F features --features features Space or comma separated list of features to activate. Features of workspace members may be enabled with package-name/feature-name syntax. This flag may be specified multiple times, which enables all specified features. --all-features Activate all available features of all selected packages. --no-default-features Do not activate the default feature of the selected packages. Manifest Options --manifest-path path Path to the Cargo.toml file. By default, Cargo searches for the Cargo.toml file in the current directory or any parent directory. --locked Asserts that the exact same dependencies and versions are used as when the existing Cargo.lock file was originally generated. Cargo will exit with an error when either of the following scenarios arises: The lock file is missing. Cargo attempted to change the lock file due to a different dependency resolution. It may be used in environments where deterministic builds are desired, such as in CI pipelines. --offline Prevents Cargo from accessing the network for any reason. Without this flag, Cargo will stop with an error if it needs to access the network and the network is not available. With this flag, Cargo will attempt to proceed without the network if possible. Beware that this may result in different dependency resolution than online mode. Cargo will restrict itself to crates that are downloaded locally, even if there might be a newer version as indicated in the local copy of the index. See the cargo-fetch(1) command to download dependencies before going offline. May also be specified with the net.offline config value . --frozen Equivalent to specifying both --locked and --offline . --lockfile-path PATH Changes the path of the lockfile from the default ( <workspace_root>/Cargo.lock ) to PATH . PATH must end with Cargo.lock (e.g. --lockfile-path /tmp/temporary-lockfile/Cargo.lock ). Note that providing --lockfile-path will ignore existing lockfile at the default path, and instead will either use the lockfile from PATH , or write a new lockfile into the provided PATH if it doesn’t exist. This flag can be used to run most commands in read-only directories, writing lockfile into the provided PATH . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #14421 ). Miscellaneous Options -j N --jobs N Number of parallel jobs to run. May also be specified with the build.jobs config value . Defaults to the number of logical CPUs. If negative, it sets the maximum number of parallel jobs to the number of logical CPUs plus provided value. If a string default is provided, it sets the value back to defaults. Should not be 0. --keep-going Build as many crates in the dependency graph as possible, rather than aborting the build on the first one that fails to build. For example if the current package depends on dependencies fails and works , one of which fails to build, cargo publish -j1 may or may not build the one that succeeds (depending on which one of the two builds Cargo picked to run first), whereas cargo publish -j1 --keep-going would definitely run both builds, even if the one run first fails. Display Options -v --verbose Use verbose output. May be specified twice for “very verbose” output which includes extra output such as dependency warnings and build script output. May also be specified with the term.verbose config value . -q --quiet Do not print cargo log messages. May also be specified with the term.quiet config value . --color when Control when colored output is used. Valid values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. May also be specified with the term.color config value . Common Options + toolchain If Cargo has been installed with rustup, and the first argument to cargo begins with + , it will be interpreted as a rustup toolchain name (such as +stable or +nightly ). See the rustup documentation for more information about how toolchain overrides work. --config KEY=VALUE or PATH Overrides a Cargo configuration value. The argument should be in TOML syntax of KEY=VALUE , or provided as a path to an extra configuration file. This flag may be specified multiple times. See the command-line overrides section for more information. -C PATH Changes the current working directory before executing any specified operations. This affects things like where cargo looks by default for the project manifest ( Cargo.toml ), as well as the directories searched for discovering .cargo/config.toml , for example. This option must appear before the command name, for example cargo -C path/to/my-project build . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #10098 ). -h --help Prints help information. -Z flag Unstable (nightly-only) flags to Cargo. Run cargo -Z help for details. ENVIRONMENT See the reference for details on environment variables that Cargo reads. EXIT STATUS 0 : Cargo succeeded. 101 : Cargo failed to complete. EXAMPLES Publish the current package: cargo publish SEE ALSO cargo(1) , cargo-package(1) , cargo-login(1)
2026-01-13T09:29:19
https://git-scm.com/book/it/v2/Git-Basics-Tagging
Git - Tagging About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Per Iniziare 1.1 Il Controllo di Versione 1.2 Una Breve Storia di Git 1.3 Cos’é Git? 1.4 La riga di comando 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Chiedere aiuto 1.8 Sommario 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Sommario 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendice A: Git in altri contesti A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Riassunto A2. Appendice B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendice C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.6 Git Basics - Tagging Tagging Like most VCSs, Git has the ability to tag specific points in history as being important. Typically people use this functionality to mark release points (v1.0, and so on). In this section, you’ll learn how to list the available tags, how to create new tags, and what the different types of tags are. Listing Your Tags Listing the available tags in Git is straightforward. Just type git tag : $ git tag v0.1 v1.3 This command lists the tags in alphabetical order; the order in which they appear has no real importance. You can also search for tags with a particular pattern. The Git source repo, for instance, contains more than 500 tags. If you’re only interested in looking at the 1.8.5 series, you can run this: $ git tag -l 'v1.8.5*' v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Creating Tags Git uses two main types of tags: lightweight and annotated. A lightweight tag is very much like a branch that doesn’t change – it’s just a pointer to a specific commit. Annotated tags, however, are stored as full objects in the Git database. They’re checksummed; contain the tagger name, e-mail, and date; have a tagging message; and can be signed and verified with GNU Privacy Guard (GPG). It’s generally recommended that you create annotated tags so you can have all this information; but if you want a temporary tag or for some reason don’t want to keep the other information, lightweight tags are available too. Annotated Tags Creating an annotated tag in Git is simple. The easiest way is to specify -a when you run the tag command: $ git tag -a v1.4 -m 'my version 1.4' $ git tag v0.1 v1.3 v1.4 The -m specifies a tagging message, which is stored with the tag. If you don’t specify a message for an annotated tag, Git launches your editor so you can type it in. You can see the tag data along with the commit that was tagged by using the git show command: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number That shows the tagger information, the date the commit was tagged, and the annotation message before showing the commit information. Lightweight Tags Another way to tag commits is with a lightweight tag. This is basically the commit checksum stored in a file – no other information is kept. To create a lightweight tag, don’t supply the -a , -s , or -m option: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 This time, if you run git show on the tag, you don’t see the extra tag information. The command just shows the commit: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Tagging Later You can also tag commits after you’ve moved past them. Suppose your commit history looks like this: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 beginning write support 0d52aaab4479697da7686c15f77a3d64d9165190 one more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc added a commit function 4682c3261057305bdd616e23b64b0857d832627b added a todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a started write support 9fceb02d0ae598e95dc970b74767f19372d61af8 updated rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a updated readme Now, suppose you forgot to tag the project at v1.2, which was at the “updated rakefile” commit. You can add it after the fact. To tag that commit, you specify the commit checksum (or part of it) at the end of the command: $ git tag -a v1.2 9fceb02 You can see that you’ve tagged the commit: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... Sharing Tags By default, the git push command doesn’t transfer tags to remote servers. You will have to explicitly push tags to a shared server after you have created them. This process is just like sharing remote branches – you can run git push origin [tagname] . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 If you have a lot of tags that you want to push up at once, you can also use the --tags option to the git push command. This will transfer all of your tags to the remote server that are not already there. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Now, when someone else clones or pulls from your repository, they will get all your tags as well. Checking out Tags You can’t really check out a tag in Git, since they can’t be moved around. If you want to put a version of your repository in your working directory that looks like a specific tag, you can create a new branch at a specific tag: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Of course if you do this and do a commit, your version2 branch will be slightly different than your v2.0.0 tag since it will move forward with your new changes, so do be careful. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://zh-cn.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT2jynv_wZlhhh0Kc6i7-3C09sWVWSFDLQsCv-oL323p5C3FDdb4_-W1Vj3y34PJEC89ywL62_PxmBRaMAr9S-DSfPKk_u5S67oaU3b74ETfMROUGkfvPlPCP9NB_XMRkVyiX1SVZSj_jsq4
Facebook Facebook 邮箱或手机号 密码 忘记账户了? 创建新账户 你暂时被禁止使用此功能 你暂时被禁止使用此功能 似乎你过度使用了此功能,因此暂时被阻止,不能继续使用。 Back 中文(简体) 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 日本語 Português (Brasil) Français (France) Deutsch 注册 登录 Messenger Facebook Lite 视频 Meta Pay Meta 商店 Meta Quest Ray-Ban Meta Meta AI Meta AI 更多内容 Instagram Threads 选民信息中心 隐私政策 隐私中心 关于 创建广告 创建公共主页 开发者 招聘信息 Cookie Ad Choices 条款 帮助 联系人上传和非用户 设置 动态记录 Meta © 2026
2026-01-13T09:29:19
https://www.netscout.com/product/arbor-cloud
Cloud DDoS Protection Service - On Demand DDoS Protection Skip to main content myNETSCOUT Login Solutions ← Back Enterprise Communication Service Provider By Industry Technology Partners Enterprise Observability nGenius Solutions for Observability Network Performance Management Application Performance Management Remote Site Observability Cloud Performance Monitoring VPN/VDI SaaS AIOps AIOps for Enterprises Collaboration UC&C and UCaaS Contact Centers Cybersecurity Network Detection and Response NetOps and SecOps Collaboration DDoS Protection Adaptive DDoS Protection Cybersecurity Stack Optimization DNS Infrastructure Protection 5G Transformation 5G for Enterprise Private 4G/5G Communication Service Provider AIOps AIOps for Communication Service Providers GeoMarketing Mobile Networks 5G 5G Fixed Wireless Access LTE/VoLTE/VoWi-Fi RAN Planning, Optimization, and Automation Cybersecurity & DDoS Protection Arbor DDoS Protection Carpet Bombing DDoS Protection DNS Infrastructure Protection Mobile Network Security Network Detection and Response Cloud and Digital Transformation Cloud Visibility Internet of Things (IoT) Network Function Virtualization (NFV) OTT Visibility Cable/MSO and Fixed Networks Cable/MSO and Fixed Networks Fiber (FTTx) and DAA Monitoring Analytics Automated Analytics Business Analytics By Industry Banking Capital Markets Education Government Healthcare Insurance Manufacturing Pharmaceutical Retail Telecommunications Transportation Utilities Featured Industry: Healthcare Optimize productivity, enhance the patient care experience, and improve bottom-line with powerful healthcare IT solutions. Learn more Technology Partners AWS Cisco Systems Citrix Dell Technologies F5 Google Cloud Microsoft Oracle Palo Alto Networks Red Hat ServiceNow Splunk VMware Products ← Back Enterprise Network Security Arbor DDoS Protection Communication Service Provider AIOps Enterprise nGeniusONE for Enterprise Enable real-time visibility into any enterprise network infrastructure environment to improve performance and user experience. InfiniStreamNG Capture, store, and analyze packet-level data in real-time to help troubleshoot issues and optimize performance across IT infrastructures. nGenius Edge Sensor Help improve user experience at remote sites and other network edges to achieve holistically strong network and application performance. vSTREAM Improve visibility into virtualized and cloud environments with deep packet inspection to further performance management and security. nGeniusPULSE Continuously test and track the performance of applications, network services, and user experience across all IT infrastructures. nGenius Packet Switches Packet flow switch architectures can combine with other NETSCOUT solutions to enable pervasive visibility across network environments. Network Security Omnis Cyber Intelligence NDR Leverage packet-level insights to power network detection and response (NDR) solutions and cyber threat hunting. Omnis CyberStream Detect, investigate, and respond to cyberthreats in real-time by leveraging packet-level data to identify suspicious activity across IT environments. nGenius Decryption Appliance Decrypt SSL and TLS-encrypted traffic to enable visibility into hidden threats within network traffic. Arbor DDoS Protection Arbor Sightline Leverage network intelligence to detect, analyze, and respond to DDoS threats across the most complex service provider and enterprise networks. Arbor Sightline Mobile Safeguard mobile services and ensure network performance and security with a DDoS protection solution specifically designed for mobile networks. Arbor Threat Mitigation System Detect, mitigate, and prevent DDoS attacks with real-time traffic analysis and automated attack mitigation. Arbor Edge Defense (AED) Powerful AI & ML powered DDoS protection backed by unmatched actionable threat intelligence maintains availability to critical services and applications. Arbor Global DDoS Threat Intelligence NETSCOUT global DDoS threat intelligence provides insights into the most current cyberthreats and their tactics, enhancing products and solutions. Arbor Cloud Cloud DDoS protection that defends against large-scale, volumetric attacks. Arbor Managed Services Round-the-clock support and assistance from Arbor product experts is available with managed services. Communication Service Provider nGenius Business Analytics Drive better business decisions at scale with efficient data collection, enrichment, and third-party application feeds. nGenius Session Analyzer Isolate and resolve congestion and interference problems contributing to 70% of all subscriber issues. nGeniusONE for Carrier Assure a premium user experience with end-to-end visibility across network domains, service enablers, and devices. InfiniStreamNG Access real-time views of RAN control plane for end-to-end visibility across physical and virtual network environments. ISNG-RAN Prioritize the mobile subscriber experience with scalable, cost-effective and cloud-ready RAN network monitoring. Omnis RAN Stay ahead of capacity and mobility targets with distinct automation modules to address growing RAN complexities. TrueCall Improve your understanding of network capacity and complexity with true user performance insights. AIOps Omnis AI Sensor Unparalleled telemetry insights at enterprise speed and scale across your entire IT ecosystem. Omnis AI Streamer Build data pipelines that sift through the noise to improve the efficiency, reliability, and scalability of IT operations and AI/ML workflows. Support & Resources ← Back Blogs, Resources, & Webinars Intelligence & Reports Support & Training Blogs, Resources, & Webinars Blog Stay updated with the latest insights, tips, and news on our blog. Resource Library Explore the latest case studies, whitepapers, eBooks, and more to discover real-world applications of our products and solutions. Learning Center Quickly learn the definitions of key phrases in networking, security, and DDoS. Webinars Gain insights from product and industry experts through our live & on-demand webinars. NETSCOUT Blog Whether you're a seasoned network performance professional or a novice in the industry, you're guaranteed to learn something new from the NETSCOUT blog. Read the blog Intelligence & Reports ASERT Threat Intelligence Team Meet the ASERT team, the leading experts on DDoS and cyber threat intelligence. DDoS Threat Intelligence Report Learn about the latest DDoS trends and statistics powered by NETSCOUT’s industry-leading global internet traffic visibility. Cyber Threat Horizon Look at live DDoS attack maps to discover what attacks are occurring in real-time. DDoS Threat Intelligence Report NETSCOUT’s latest DDoS Threat Intelligence dives into the most recent trends in DDoS adversarial tactics. Get the report Support & Training Overview Get support for your products and solutions to maximize the value of your investment. My.NETSCOUT Existing customers, log in to MyNETSCOUT to access additional assets, resources, and support channels. NETSCOUT University Our mission is to help you and your team to successfully implement NETSCOUT’s solutions so that your network functions seamlessly and effectively. NO LATENCY Explore tech tips, use cases, and more in our NO LATENCY newsletter. MasterCare Support NETSCOUT's award-winning global support service. Learn more Company ← Back About Us News & Events Partners About Us About Us Board of Directors Careers Customer Reviews Environmental, Social, and Governance Executive Team Industry Analyst Perspectives Investor Relations The Future of Intelligence The quality of your AI outcomes hinges on the data available. With NETSCOUT's Smart Data, ensure you have the network visibility you need to overcome the challenges of tomorrow. NETSCOUT provides the Data that Drives You. Watch the video News & Events Events Newsroom Press Releases Read the latest NETSCOUT news to stay informed on the latest updates, press, and more. Read the latest news Partners Partner Portal Become a Partner Learn about our NETSCOUT Connect 360 partner program. Search Under Attack? Contact Us Under Attack? Home Products Arbor Cloud DDoS Protection Services Arbor Cloud DDoS Protection Services Fully Managed, Intelligently Automated On-Demand DDoS Protection Overview Overview Benefits Features Resources Related Pages Proven Cloud DDoS Mitigation Provider NETSCOUT’s Arbor hybrid DDoS protection solution using Arbor Cloud employs an adaptive DDoS attack mitigation mechanism that operates both from data centers and the cloud. This approach is designed to ensure comprehensive protection against various DDoS threats. These services specifically counter high-volume attacks targeting bandwidth, "low and slow" attacks aimed at applications and infrastructure, and simultaneous multi-vector attacks. With Arbor Cloud, inbound traffic is routed via BGP or DNS through 16 DDoS scrubbing centers with over 15Tbps capacity. Even the largest attacks are well mitigated, and only clean traffic is returned to the internet access links and servers. Benefits Purpose Built Global DDoS Protection Arbor Cloud provides over 15 Tbps of DDoS attack mitigation capacity via 16 worldwide scrubbing centers located in Asia, Europe and The Americas. Seamless Hybrid Protection to Industry Best Practices On-premise protection guards against state-exhausting and application attacks aimed at your infrastructure in tandem with Arbor Cloud. It protects against attacks that target Firewalls, Intrusion Prevention Systems (IPS) and business-critical applications. Expertise The Arbor Cloud DDoS protection service is supported by NETSCOUT’s 24x7 ASERT Team. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Sub Minute Mitigation SLA With the intelligent automation capabilities of Arbor Cloud, we can provide mitigation initiation within 60-seconds of detecting an attack via cloud signaling from your AED, via flow-detection or as part of your always-on DDoS solution. Click to enlarge image Arbor Cloud Managed DDoS Protection Service for Enterprises Arbor Cloud combines on-premise DDoS defense with cloud-based traffic scrubbing services that are tightly integrated via an automated cloud signal. This multi-layered, adaptive DDoS protection approach is a proven industry best practice and is the only way to mitigate today's full spectrum of DDoS threats for both Service Providers and Enterprises, all from a single cloud DDoS protection provider. DDoS Scrubbing Capabilities NETSCOUT's Arbor Cloud is fueled by 16 DDoS scrubbing centers worldwide, boasting a capacity of more than 15 Tbps. These scrubbing centers are located across the globe to best protect against attacks originating from any corner of the world. This industry-leading capability helps monitor and maintain availability in the face of large-scale volumetric DDoS attacks . Solution Brief Why NETSCOUT's Arbor DDoS Attack Protection Solution is Better NETSCOUT Arbor has the industry’s broadest portfolio of DDoS attack protection products and services that enable organizations of any size to customize a solution to meet their technical and financial requirements of today – and the future. Learn more Features Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Using stateless packet-processing technology and/or cloud-based IP flow analysis, DDoS attacks can be automatically detected and routed to Arbor Cloud global DDoS scrubbing centers for mitigation. Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Click to enlarge image Arbor Cloud brings the availability of your network, services, and applications back under your control.  Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Click to enlarge image It can also be employed in an intelligent combination of in-cloud and on-premise hybrid DDoS protection, providing you the flexibility to design a comprehensive DDoS mitigation solution that fits your environment. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Click to enlarge image ATLAS Security Engineering & Response Team (ASERT) delivers world-class network security research and analysis for the benefit of today’s enterprise and network operators. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Using automated malware analysis pipelines, sinkholes, scanners, honeypots, open-source intelligence data sets and ASERT analysis, we can provide a unique view in the threat landscape.  Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Data Sheet NETSCOUT Arbor Cloud for Service Providers Augment your local mitigation capacity and staff with Arbor Cloud for Service Providers. Learn more Data Sheet NETSCOUT Arbor Cloud DDoS Protection for Enterprises NETSCOUT Arbor Cloud combines on-premise AED DDoS defense with global cloud-based traffic scrubbing services for best-of-breed, hybrid, DDoS attack protection. Read More Featured Resources Case Study Globe Telecom Protects Critical Infrastructure and Customer Experience with NETSCOUT Arbor DDoS Attack Protection Globe Telecom struggled with DDoS attacks due to a lack of real-time systems for detection and prevention. This caused multiple issues for the company. Report SPARK Matrix for Distributed Denial of Service (DDoS) Mitigation, Q3 2025 NETSCOUT Named both a Technology Leader and Ace Performer in DDoS Mitigation for 2025 White Paper IDC - Security Requirements Driving Comprehensive DDoS Protection Security buyers must maintain a vigilant approach toward assessing the state of their defenses—an incomplete strategy can be as costly of a mistake as a reactive strategy. White Paper EMA PRISM Report – DDoS Mitigation Solutions NETSCOUT Solution Profile Quick Look Healthcare Organizations are Increasingly Targets of DDoS Attacks Healthcare Organization Uses NETSCOUT Solution To Solve Cybersecurity Challenges by Stopping DDoS Attacks Impacting Availability of Services Solution Brief On-Premise Protection is the Best First Step Against DDoS and Cyberattacks for Academic Institutions Due to the complexity of DDoS attacks, Educational Institutions need to understand the risk reduction that on-premise edge protection can provide to their on line learning... Use Case Using Arbor Cloud to Protect Multi-Cloud and Hybrid Environments Against DDoS Attacks Arbor Cloud provides a superior means for protecting critical assets hosted in a single or multi-cloud environment, offering several advantages over less capable protection... Use Case Defending Government Agencies Against High-Volume Multi-Vector DDoS Attacks Attacks are often waged against national security infrastructure. NETSCOUT has observed a massive increase from the previous half year in attacks against the US national security... View more resources Related Pages Product ATLAS Intelligence Feed (AIF) ATLAS Intelligence Feed (AIF) provides up-to-date threat intelligence on the latest DDoS threats to our DDoS protection products. Automate defenses against trending attacks with AIF. Learn more Product Arbor Edge Defense (AED) Arbor Edge Defense (AED) is an inline security appliance deployed at the network perimeter, between the internet router and firewall. This AI and ML-powered solution provides stateless, on-premise DDoS protection. Learn more Product Arbor Sightline DDoS Attack Detection Solution Whether you are a service provider or have a complex enterprise network, Arbor Sightline helps monitor and identify networking and security issues at any network scale with AI and ML-powered insights. Learn more Product Arbor Threat Mitigation System for DDoS Attacks Filter out malicious traffic while allowing legitimate traffic to pass through, at any scale, with Arbor Threat Mitigation System (TMS). Learn more Contact NETSCOUT Today To learn more about how NETSCOUT can help your company, speak to one of our highly experienced subject matter experts. Contact Us Guardians of the Connected World ® Discover Our Data Observability Solutions Service Provider Service Assurance AIOps Omnis Network Security Arbor DDoS Protection All Solutions All Products Learn Blog Resource Library Learning Center MyNETSCOUT Events Webinars About About Us Careers Environmental, Social, and Governance For Investors Newsroom Partners Connect 1-888-357-7667 Contact Us Under DDoS Attack? © 2026 NETSCOUT Terms & Conditions Website Terms of Use Data Privacy and Trust Center Transparency in Supply Chains Privacy Policy Vendor Portal Do Not Sell or Share My Personal Information California Privacy Notice UK Disclosure
2026-01-13T09:29:19
https://th-th.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT2jynv_wZlhhh0Kc6i7-3C09sWVWSFDLQsCv-oL323p5C3FDdb4_-W1Vj3y34PJEC89ywL62_PxmBRaMAr9S-DSfPKk_u5S67oaU3b74ETfMROUGkfvPlPCP9NB_XMRkVyiX1SVZSj_jsq4
Facebook Facebook อีเมลหรือโทรศัพท์ รหัสผ่าน ลืมบัญชีใช่หรือไม่ สร้างบัญชีใหม่ คุณถูกบล็อกชั่วคราว คุณถูกบล็อกชั่วคราว ดูเหมือนว่าคุณจะใช้คุณสมบัตินี้ในทางที่ผิดโดยการใช้เร็วเกินไป คุณถูกบล็อกจากการใช้โดยชั่วคราว Back ภาษาไทย 한국어 English (US) Tiếng Việt Bahasa Indonesia Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch สมัคร เข้าสู่ระบบ Messenger Facebook Lite วิดีโอ Meta Pay Meta Store Meta Quest Ray-Ban Meta Meta AI เนื้อหาเพิ่มเติมจาก Meta AI Instagram Threads ศูนย์ข้อมูลการลงคะแนนเสียง นโยบายความเป็นส่วนตัว ศูนย์ความเป็นส่วนตัว เกี่ยวกับ สร้างโฆษณา สร้างเพจ ผู้พัฒนา ร่วมงานกับ Facebook คุกกี้ ตัวเลือกโฆษณา เงื่อนไข ความช่วยเหลือ การอัพโหลดผู้ติดต่อและผู้ที่ไม่ได้ใช้บริการ การตั้งค่า บันทึกกิจกรรม Meta © 2026
2026-01-13T09:29:19
https://es-la.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT2jynv_wZlhhh0Kc6i7-3C09sWVWSFDLQsCv-oL323p5C3FDdb4_-W1Vj3y34PJEC89ywL62_PxmBRaMAr9S-DSfPKk_u5S67oaU3b74ETfMROUGkfvPlPCP9NB_XMRkVyiX1SVZSj_jsq4
Facebook Facebook Correo o teléfono Contraseña ¿Olvidaste tu cuenta? Crear cuenta nueva Se te bloqueó temporalmente Se te bloqueó temporalmente Parece que hiciste un uso indebido de esta función al ir muy rápido. Se te bloqueó su uso temporalmente. Back Español 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch Registrarte Iniciar sesión Messenger Facebook Lite Video Meta Pay Tienda de Meta Meta Quest Ray-Ban Meta Meta AI Más contenido de Meta AI Instagram Threads Centro de información de votación Política de privacidad Centro de privacidad Información Crear anuncio Crear página Desarrolladores Empleo Cookies Opciones de anuncios Condiciones Ayuda Importación de contactos y no usuarios Configuración Registro de actividad Meta © 2026
2026-01-13T09:29:19
https://git-scm.com/book/be/v2/Git-Basics-Getting-a-Git-Repository
Git - Getting a Git Repository About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Першыя крокі 1.1 About Version Control 1.2 A Short History of Git 1.3 What is Git? 1.4 The Command Line 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Getting Help 1.8 Падсумаваньне 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Дадатак A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Summary A2. Дадатак B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Дадатак C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.1 Git Basics - Getting a Git Repository If you can read only one chapter to get going with Git, this is it. This chapter covers every basic command you need to do the vast majority of the things you’ll eventually spend your time doing with Git. By the end of the chapter, you should be able to configure and initialize a repository, begin and stop tracking files, and stage and commit changes. We’ll also show you how to set up Git to ignore certain files and file patterns, how to undo mistakes quickly and easily, how to browse the history of your project and view changes between commits, and how to push and pull from remote repositories. Getting a Git Repository You typically obtain a Git repository in one of two ways: You can take a local directory that is currently not under version control, and turn it into a Git repository, or You can clone an existing Git repository from elsewhere. In either case, you end up with a Git repository on your local machine, ready for work. Initializing a Repository in an Existing Directory If you have a project directory that is currently not under version control and you want to start controlling it with Git, you first need to go to that project’s directory. If you’ve never done this, it looks a little different depending on which system you’re running: for Linux: $ cd /home/user/my_project for macOS: $ cd /Users/user/my_project for Windows: $ cd C:/Users/user/my_project and type: $ git init This creates a new subdirectory named .git that contains all of your necessary repository files — a Git repository skeleton. At this point, nothing in your project is tracked yet. See Git Internals for more information about exactly what files are contained in the .git directory you just created. If you want to start version-controlling existing files (as opposed to an empty directory), you should probably begin tracking those files and do an initial commit. You can accomplish that with a few git add commands that specify the files you want to track, followed by a git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' We’ll go over what these commands do in just a minute. At this point, you have a Git repository with tracked files and an initial commit. Cloning an Existing Repository If you want to get a copy of an existing Git repository — for example, a project you’d like to contribute to — the command you need is git clone . If you’re familiar with other VCSs such as Subversion, you’ll notice that the command is "clone" and not "checkout". This is an important distinction — instead of getting just a working copy, Git receives a full copy of nearly all data that the server has. Every version of every file for the history of the project is pulled down by default when you run git clone . In fact, if your server disk gets corrupted, you can often use nearly any of the clones on any client to set the server back to the state it was in when it was cloned (you may lose some server-side hooks and such, but all the versioned data would be there — see Getting Git on a Server for more details). You clone a repository with git clone <url> . For example, if you want to clone the Git linkable library called libgit2 , you can do so like this: $ git clone https://github.com/libgit2/libgit2 That creates a directory named libgit2 , initializes a .git directory inside it, pulls down all the data for that repository, and checks out a working copy of the latest version. If you go into the new libgit2 directory that was just created, you’ll see the project files in there, ready to be worked on or used. If you want to clone the repository into a directory named something other than libgit2 , you can specify the new directory name as an additional argument: $ git clone https://github.com/libgit2/libgit2 mylibgit That command does the same thing as the previous one, but the target directory is called mylibgit . Git has a number of different transfer protocols you can use. The previous example uses the https:// protocol, but you may also see git:// or user@server:path/to/repo.git , which uses the SSH transfer protocol. Getting Git on a Server will introduce all of the available options the server can set up to access your Git repository and the pros and cons of each. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/commands/cargo-publish.html#option-cargo-publish---index
cargo publish - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book cargo-publish(1) NAME cargo-publish — Upload a package to the registry SYNOPSIS cargo publish [ options ] DESCRIPTION This command will create a distributable, compressed .crate file with the source code of the package in the current directory and upload it to a registry. The default registry is https://crates.io . This performs the following steps: Performs a few checks, including: Checks the package.publish key in the manifest for restrictions on which registries you are allowed to publish to. Create a .crate file by following the steps in cargo-package(1) . Upload the crate to the registry. The server will perform additional checks on the crate. The client will poll waiting for the package to appear in the index, and may timeout. In that case, you will need to check for completion manually. This timeout does not affect the upload. This command requires you to be authenticated using cargo-login(1) or environment variables of the registry.token and registries.<name>.token config fields. See the reference for more details about packaging and publishing. OPTIONS Publish Options --dry-run Perform all checks without uploading. --no-verify Don’t verify the contents by building them. --allow-dirty Allow working directories with uncommitted VCS changes to be packaged. --index index The URL of the registry index to use. --registry registry Name of the registry to publish to. Registry names are defined in Cargo config files . If not specified, and there is a package.publish field in Cargo.toml with a single registry, then it will publish to that registry. Otherwise it will use the default registry, which is defined by the registry.default config key which defaults to crates-io . Package Selection By default, when no package selection options are given, the packages selected depend on the selected manifest file (based on the current working directory if --manifest-path is not given). If the manifest is the root of a workspace then the workspaces default members are selected, otherwise only the package defined by the manifest will be selected. The default members of a workspace can be set explicitly with the workspace.default-members key in the root manifest. If this is not set, a virtual workspace will include all workspace members (equivalent to passing --workspace ), and a non-virtual workspace will include only the root crate itself. -p spec … --package spec … Publish only the specified packages. See cargo-pkgid(1) for the SPEC format. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. --workspace Publish all members in the workspace. --all Deprecated alias for --workspace . --exclude SPEC … Exclude the specified packages. Must be used in conjunction with the --workspace flag. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. Compilation Options --target triple Publish for the specified target architecture. Flag may be specified multiple times. The default is the host architecture. The general format of the triple is <arch><sub>-<vendor>-<sys>-<abi> . Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. This may also be specified with the build.target config value . Note that specifying this flag makes Cargo run in a different mode where the target artifacts are placed in a separate directory. See the build cache documentation for more details. --target-dir directory Directory for all generated artifacts and intermediate files. May also be specified with the CARGO_TARGET_DIR environment variable, or the build.target-dir config value . Defaults to target in the root of the workspace. Feature Selection The feature flags allow you to control which features are enabled. When no feature options are given, the default feature is activated for every selected package. See the features documentation for more details. -F features --features features Space or comma separated list of features to activate. Features of workspace members may be enabled with package-name/feature-name syntax. This flag may be specified multiple times, which enables all specified features. --all-features Activate all available features of all selected packages. --no-default-features Do not activate the default feature of the selected packages. Manifest Options --manifest-path path Path to the Cargo.toml file. By default, Cargo searches for the Cargo.toml file in the current directory or any parent directory. --locked Asserts that the exact same dependencies and versions are used as when the existing Cargo.lock file was originally generated. Cargo will exit with an error when either of the following scenarios arises: The lock file is missing. Cargo attempted to change the lock file due to a different dependency resolution. It may be used in environments where deterministic builds are desired, such as in CI pipelines. --offline Prevents Cargo from accessing the network for any reason. Without this flag, Cargo will stop with an error if it needs to access the network and the network is not available. With this flag, Cargo will attempt to proceed without the network if possible. Beware that this may result in different dependency resolution than online mode. Cargo will restrict itself to crates that are downloaded locally, even if there might be a newer version as indicated in the local copy of the index. See the cargo-fetch(1) command to download dependencies before going offline. May also be specified with the net.offline config value . --frozen Equivalent to specifying both --locked and --offline . --lockfile-path PATH Changes the path of the lockfile from the default ( <workspace_root>/Cargo.lock ) to PATH . PATH must end with Cargo.lock (e.g. --lockfile-path /tmp/temporary-lockfile/Cargo.lock ). Note that providing --lockfile-path will ignore existing lockfile at the default path, and instead will either use the lockfile from PATH , or write a new lockfile into the provided PATH if it doesn’t exist. This flag can be used to run most commands in read-only directories, writing lockfile into the provided PATH . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #14421 ). Miscellaneous Options -j N --jobs N Number of parallel jobs to run. May also be specified with the build.jobs config value . Defaults to the number of logical CPUs. If negative, it sets the maximum number of parallel jobs to the number of logical CPUs plus provided value. If a string default is provided, it sets the value back to defaults. Should not be 0. --keep-going Build as many crates in the dependency graph as possible, rather than aborting the build on the first one that fails to build. For example if the current package depends on dependencies fails and works , one of which fails to build, cargo publish -j1 may or may not build the one that succeeds (depending on which one of the two builds Cargo picked to run first), whereas cargo publish -j1 --keep-going would definitely run both builds, even if the one run first fails. Display Options -v --verbose Use verbose output. May be specified twice for “very verbose” output which includes extra output such as dependency warnings and build script output. May also be specified with the term.verbose config value . -q --quiet Do not print cargo log messages. May also be specified with the term.quiet config value . --color when Control when colored output is used. Valid values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. May also be specified with the term.color config value . Common Options + toolchain If Cargo has been installed with rustup, and the first argument to cargo begins with + , it will be interpreted as a rustup toolchain name (such as +stable or +nightly ). See the rustup documentation for more information about how toolchain overrides work. --config KEY=VALUE or PATH Overrides a Cargo configuration value. The argument should be in TOML syntax of KEY=VALUE , or provided as a path to an extra configuration file. This flag may be specified multiple times. See the command-line overrides section for more information. -C PATH Changes the current working directory before executing any specified operations. This affects things like where cargo looks by default for the project manifest ( Cargo.toml ), as well as the directories searched for discovering .cargo/config.toml , for example. This option must appear before the command name, for example cargo -C path/to/my-project build . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #10098 ). -h --help Prints help information. -Z flag Unstable (nightly-only) flags to Cargo. Run cargo -Z help for details. ENVIRONMENT See the reference for details on environment variables that Cargo reads. EXIT STATUS 0 : Cargo succeeded. 101 : Cargo failed to complete. EXAMPLES Publish the current package: cargo publish SEE ALSO cargo(1) , cargo-package(1) , cargo-login(1)
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-Git-op-een-server-krijgen
Git - Git op een server krijgen About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.2 Git op de server - Git op een server krijgen Git op een server krijgen We gaan nu het inrichten van een Git service op je eigen server behandelen waarin deze protocols worden gebuikt. Noot We zullen hier de commando’s en stappen laten zien om een eenvoudige, versimpelde installatie op een op Linux gebaseerde server op te zetten, alhoewel het ook mogelijk is deze services op een macOS of Windows server te draaien. Het daadwerkelijk opzetten van een productie server binnen jouw infrastructuur zal vrijwel zeker verschillen in de manier waarop de veiligheidsmaatregelen zijn ingericht of de gebruikte tooling van het besturingssysteem, maar hopelijk zal dit je een indruk geven van wat er allemaal bij komt kijken. Om initieel een Git server op te zetten, moet je een bestaande repository naar een nieuwe kale repository exporteren — een repository die geen werk directory bevat. Dit is over het algemeen eenvoudig te doen. Om je repository clonen om daarmee een nieuwe kale repository te maken, draai je het clone commando met de --bare optie. De conventie is om directories waar kale repositories in staan te laten eindigen met .git , zoals hier: $ git clone --bare my_project my_project.git Cloning into bare repository 'my_project.git'... done. Je zou nu een kopie van de Git directory gegevens in je my_project.git directory moeten hebben Dit is grofweg gelijk aan $ cp -Rf my_project/.git my_project.git Er zijn een paar kleine verschillen in het configuratie bestand, maar het komt op hetzelfde neer. Het neemt de Git repository zelf, zonder een werkdirectory, en maakt een directory specifiek hiervoor aan. De kale repository op een server zetten Nu je een kale kopie van je repository hebt, is het enige dat je moet doen het op een server zetten en je protocollen instellen. Laten we aannemen dat je een server ingericht hebt die git.example.com heet, waar je SSH toegang op hebt, en waar je al je Git repositories wilt opslaan onder de /srv/git directory. Aangenomen dat /srv/git bestaat op die server, kan je deze nieuwe repository beschikbaar stellen door je kale repository ernaartoe te kopiëren: $ scp -r my_project.git user@git.example.com:/srv/git Vanaf dat moment kunnen andere gebruikers, die SSH toegang hebben tot dezelfde server en lees-toegang hebben tot de /srv/git directory, jouw repository clonen door dit uit te voeren: $ git clone user@git.example.com:/srv/git/my_project.git Als een gebruiker met SSH op een server inlogt en schrijftoegang heeft tot de /srv/git/my_project.git directory, dan hebben ze automatisch ook push toegang. Git zal automatisch de correcte groep schrijfrechten aan een repository toekennen als je het git init commando met de --shared optie uitvoert. Merk op dat met het aanroepen van dit commando je geen enkele commit, refs, etc. zult verwijderen. $ ssh user@git.example.com $ cd /opt/git/my_project.git $ git init --bare --shared Je ziet hoe eenvoudig het is om een Git repository te nemen, een kale versie aan te maken, en het op een server plaatsen waar jij en je medewerkers SSH toegang tot hebben. Nu zijn jullie klaar om aan hetzelfde project samen te werken. Het is belangrijk om op te merken dat dit letterlijk alles is wat je moet doen om een bruikbare Git server te draaien waarop meerdere mensen toegang hebben: maak alleen een paar accounts met SSH toegang aan op een server, en stop een kale repository ergens waar al die gebruikers lees- en schrijftoegang toe hebben. Je bent er klaar voor — je hebt niets anders nodig. In de volgende paragrafen zul je zien hoe je meer ingewikkelde opstellingen kunt maken. Deze bespreking zal het niet hoeven aanmaken van gebruikers accounts voor elke gebruiker, publieke leestoegang tot repositories, grafische web interfaces en meer omvatten. Maar, hou in gedachten dat om samen te kunnen werken met mensen op een privé project, alles wat je nodig hebt een SSH server is en een kale repository. Kleine opstellingen Als je met een kleine groep bent of net begint met Git in je organisatie en slechts een paar ontwikkelaars hebt, dan kunnen de dingen eenvoudig voor je zijn. Een van de meest gecompliceerde aspecten van een Git server instellen is het beheren van gebruikers. Als je sommige repositories alleen-lezen voor bepaalde gebruikers wilt hebben, en lees/schrijf voor anderen, dan kunnen toegang en permissies een beetje lastiger te regelen zijn. SSH toegang Als je al een server hebt waar al je ontwikkelaars SSH toegang op hebben, dan is het over het algemeen het eenvoudigste om je eerste repository daar op te zetten, omdat je dan bijna niets hoeft te doen (zoals beschreven in de vorige paragraaf). Als je meer complexe toegangscontrole wilt op je repositories, dan kun je ze instellen met de normale bestandssysteem permissies van het operating systeem dat op je server draait. Als je je repositories op een server wilt zetten, die geen accounts heeft voor iedereen in je team die je schrijftoegang wilt geven, dan moet je SSH toegang voor ze opzetten. We gaan er vanuit dat je een server hebt waarmee je dit kunt doen, dat je reeds een SSH server geïnstalleerd hebt, en dat de manier is waarop je toegang hebt tot de server. Er zijn een paar manieren waarop je iedereen in je team toegang kunt geven. De eerste is voor iedereen accounts aanmaken, wat rechttoe rechtaan is maar bewerkelijk kan zijn. Je wilt vermoedelijk niet adduser uitvoeren en tijdelijke wachtwoorden instellen voor iedere gebruiker. Een tweede methode is een enkele git gebruiker aan te maken op de machine, aan iedere gebruiker die schijftoegang moet hebben vragen of ze je een publieke SSH sleutel sturen, en die sleutel toevoegen aan het ~/.ssh/authorized_keys bestand van die nieuwe gebruiker. Vanaf dat moment zal iedereen toegang hebben op die machine via de git gebruiker. Dit tast de commit data op geen enkele manier aan — de SSH gebruiker waarmee je inlogt zal de commits die je opgeslagen hebt niet beïnvloeden. Een andere manier waarop je het kunt doen is je SSH server laten authenticeren middels een LDAP server of een andere gecentraliseerde authenticatie bron, die misschien al ingericht is. Zolang iedere gebruiker shell toegang kan krijgen op de machine, zou ieder SSH authenticatie mechanisme dat je kunt bedenken moeten werken. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/bg/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d0%b8-%d0%bd%d0%b0-Git-%d0%a0%d0%b0%d0%b1%d0%be%d1%82%d0%b0-%d1%81-%d0%be%d1%82%d0%b4%d0%b0%d0%bb%d0%b5%d1%87%d0%b5%d0%bd%d0%b8-%d1%85%d1%80%d0%b0%d0%bd%d0%b8%d0%bb%d0%b8%d1%89%d0%b0
Git - Работа с отдалечени хранилища About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Начало 1.1 За Version Control системите 1.2 Кратка история на Git 1.3 Какво е Git 1.4 Конзолата на Git 1.5 Инсталиране на Git 1.6 Първоначална настройка на Git 1.7 Помощна информация в Git 1.8 Обобщение 2. Основи на Git 2.1 Създаване на Git хранилище 2.2 Запис на промени в хранилището 2.3 Преглед на историята на действията 2.4 Възстановяване на направени действия 2.5 Работа с отдалечени хранилища 2.6 Тагове в Git 2.7 Псевдоними в Git 2.8 Обобщение 3. Клонове в Git 3.1 Накратко за разклоненията 3.2 Основи на клоновете код и сливането 3.3 Управление на клонове 3.4 Стратегии за работа с клонове код 3.5 Отдалечени клонове 3.6 Управление на проект 3.7 Обобщение 4. GitHub 4.1 Създаване и настройка на акаунт 4.2 Как да сътрудничим в проект 4.3 Управление на проект 4.4 Управление на организация 4.5 Автоматизиране с GitHub 4.6 Обобщение 5. Git инструменти 5.1 Избор на къмити 5.2 Интерактивно индексиране 5.3 Stashing и Cleaning 5.4 Подписване на вашата работа 5.5 Търсене 5.6 Манипулация на историята 5.7 Мистерията на командата Reset 5.8 Сливане за напреднали 5.9 Rerere 5.10 Дебъгване с Git 5.11 Подмодули 5.12 Пакети в Git (Bundling) 5.13 Заместване 5.14 Credential Storage система 5.15 Обобщение 6. Настройване на Git 6.1 Git конфигурации 6.2 Git атрибути 6.3 Git Hooks 6.4 Примерна Git-Enforced политика 6.5 Обобщение 7. Git и други системи 7.1 Git като клиент 7.2 Миграция към Git 7.3 Обобщение 8. Git на ниско ниво 8.1 Plumbing и Porcelain команди 8.2 Git обекти 8.3 Git референции 8.4 Packfiles 8.5 Refspec спецификации 8.6 Транспортни протоколи 8.7 Поддръжка и възстановяване на данни 8.8 Environment променливи 8.9 Обобщение 9. Приложение A: Git в други среди 9.1 Графични интерфейси 9.2 Git във Visual Studio 9.3 Git във Visual Studio Code 9.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 9.5 Git в Sublime Text 9.6 Git в Bash 9.7 Git в Zsh 9.8 Git в PowerShell 9.9 Обобщение 10. Приложение B: Вграждане на Git в приложения 10.1 Git от команден ред 10.2 Libgit2 10.3 JGit 10.4 go-git 10.5 Dulwich A1. Приложение C: Git команди A1.1 Настройки и конфигурация A1.2 Издърпване и създаване на проекти A1.3 Snapshotting A1.4 Клонове и сливане A1.5 Споделяне и обновяване на проекти A1.6 Инспекция и сравнение A1.7 Дебъгване A1.8 Patching A1.9 Email команди A1.10 Външни системи A1.11 Административни команди A1.12 Plumbing команди 2nd Edition 2.5 Основи на Git - Работа с отдалечени хранилища Работа с отдалечени хранилища За да можете да сътрудничите в Git проекти, трябва да знаете как да управлявате отдалечените си хранилища. Отдалечените хранилища са версии на вашия проект, които се съхраняват някъде в корпоративната мрежа или в Интернет. Може да имате много от тях, като всяко от хранилищата може да е с права за вас само за четене или за четене/писане. Сътрудничеството с други разработчици изисква управление на тези отдалечени хранилища и издърпване/изпращане на данни от/към тях когато трябва да споделите работата си. Управлението на отдалечени хранилища включва умения за това как да добавяте такива, да премахвате хранилища, които вече не са валидни, управление на разнообразни отдалечени клонове код и дефинирането им като следени (tracked) или не и т.н. В тази секция разглеждаме някои от тези неща. Забележка Отдалечени хранилища могат да присъстват на локалната ви машина. Напълно възможно е да работите с “отдалечено” хранилище, което в действителност се намира на същия хост. Думата “отдалечено” не означава непременно, че хранилището е някъде другаде в мрежата или в Интернет, тя означава че то просто е другаде. Работата с подобно хранилище все още си изисква всичките стандартни push/pull/fetch операции. Показване на отдалечените хранилища За да видите кои отдалечени сървъри имате конфигурирани, използвайте командата git remote . Тя отпечатва съкратените имена на всяко отдалечено хранилище, което сте указали. Ако сте клонирали отдалечено хранилище, трябва да видите в списъка поне елемента origin  — това е подразбиращото се име, което Git дава на сървъра, от който сте клонирали: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Можете да ползвате и флага -v , който показва пълния URL, който Git пази за съответното кратко име на отдалеченото хранилище: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Ако имате повече от едно отдалечено хранилище, командата показва всички тях. Например, хранилище с много отдалечени копия за работа с няколко сътрудника, би могло да изглежда така. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) Това означава, че можем да издърпваме работата от всеки от тези потребители доста лесно. Може освен това да имаме права да изпращаме наш код към едно или повече от копията, въпреки че това не става ясно от тук. Отбележете също, че тези отдалечени копия използват различни протоколи, ще обърнем повече внимание на това в [_getting_git_on_a_server] . Добавяне на отдалечени хранилища Вече показахме как командата git clone самичка създава origin отдалеченото хранилище за вас. Ето как да си добавите изрично отдалечено хранилище. Командата git remote add <shortname> <url> добавя отдалечено хранилище със съкратено име, което впоследствие да можете да ползвате лесно в обръщенията към него: $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Сега вече можете да ползвате краткото име pb, вместо целия отдалечен URL. Например, ако искате да издърпате цялата информация, която Paul има, но все още не е в локалното ви хранилище, можете да използвате командата git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Сега master клонът на Paul е достъпен локално за вас като pb/master  — можете да го слеете в някой от вашите клонове код или да превключите към него, ако желаете да го прегледате. Ще разгледаме подробно какво са клоновете код и как да ги използваме в Клонове в Git . Fetching и Pulling на данни от отдалечени хранилища Както току що видяхте, извличането на данни от отдалечен проект се прави с: $ git fetch <remote> Командата се свързва с отдалечения сървър и изтегля всички данни за него, които все още нямате локално. След като направите това, ще имате указатели към всички клонове код на това отдалечено хранилище, така че да можете да ги инспектирате и слеете с локалните си клонове по всяко време. Ако клонирате хранилище, командата git clone автоматично го добавя като отдалечено под съкратеното име “origin”. Така че, git fetch origin изтегля всички нови данни от това хранилище, които са били добавени след като сте го клонирали (или последно актуализирали). Важно е да се отбележи, че git fetch само изтегля информацията в локалната ви база данни — тя не прави автоматично сливане с каквато и да било част от локалната ви работа и по никакъв начин не модифицира локалните файлове. Вие сами трябва да направите това ръчно, когато сте готови. Ако текущият ви клон код (branch) е настроен да проследява отдалечен клон (вижте следващата секция и Клонове в Git за повече информация), можете да използвате командата git pull за автоматично изтегляне и сливане на данните от отдалечения клон в локалния. Това може да е по-лесно и по-удобно за вас като работна последователност и по подразбиране, git clone автоматично настройва локалния ви master да следи отдалечения master клон от сървъра, от който сте клонирали (или както се казва клонът по подразбиране на него). Така че git pull изтегля данните от сървъра, от който сте клонирали и автоматично се опитва да ги слее с кода, върху който работите в момента, спестявайки ви една ръчна стъпка по сливането. Забележка From git version 2.27 onward, git pull will give a warning if the pull.rebase variable is not set. Git will keep warning you until you set the variable. If you want the default behavior of git (fast-forward if possible, else create a merge commit): git config --global pull.rebase "false" If you want to rebase when pulling: git config --global pull.rebase "true" Изпращане на данни към отдалечено хранилище Когато работата ви стигне до статус, в който искате да я публикувате с другите разработчици от даден проект, можете да изпратите промените си към главното хранилище. Командата за това е проста: git push <remote> <branch> . Ако искате да публикувате промените си от локалния master клон към origin сървъра (да кажем отново, клонирането обикновено настройва и двете кратки имена за вас автоматично), можете да изпълните следното: $ git push origin master Тази команда работи само ако, сте клонирали хранилището от отдалечен сървър, ако имате права за писане в него и ако никой междувременно не е изпратил към него нещо преди вас. Ако вие и някой друг сте клонирали едно и също хранилище и ако другият е изпратил обратно данни преди вас, то вашето изпращане правилно ще бъде отказано. Ще трябва първо да изтеглите работата на другия колега, да я слеете в локалното си копие и след това ще можете да изпратите към отдалечения сървър. Вижте Клонове в Git за повече подробности относно това как да изпращате към отдалечен сървър. Преглед на отдалечено хранилище Ако желаете повече информация за отдалечено хранилище, ползвайте командата git remote show <remote> . Ако я изпълните с определено кратко име като origin например, ще видите следното: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Показва се адреса на отдалеченото хранилище, както и информация за проследяването на клоновете. Командата подсказва, че ако сте в локалния master клон и изпълните git pull , това автоматично ще го слее с промените в отдалеченoто хранилище след изтеглянето му. Тя също така отпечатва и всички отдалечени референции, които са издърпани. Това е прост пример, който е вероятно да срещнете. Когато ползвате Git по-интензивно обаче, може да видите и доста повече данни от git remote show : $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Тази команда показва към кой отдалечен клон ще се изпращат вашите промени, когато изпълните git push докато сте в определен локален клон. Тя още ви показва отдалечените клонове на сървъра, които вие все още нямате локално при вас, локално съхранените от преди клонове, които вече не съществуват на сървъра, и множество локални клонове, които могат да се слеят автоматично със съответните си отдалечени такива (които следят) при изпълнение на git pull . Премахване и преименуване на отдалечени хранилища Използвайте git remote rename за смяна на краткото име на отдалечено хранилище. Например, ако искате да смените името на pb с paul : $ git remote rename pb paul $ git remote origin paul Това също променя и имената, под които се показват отдалечените проследявани клонове. Този, който преди се казваше pb/master сега е paul/master . Ако по някаква причина искате да премахнете отдалечено хранилище, например сменили сте сървъра или пък някое от огледалата или пък определен сътрудник не участва повече в проекта — можете да използвате командите git remote remove или git remote rm : $ git remote remove paul $ git remote origin Веднъж след като премахнете референция към отдалечено хранилище по този начин, всички remote-tracking клонове и конфигурационни настройки асоциирани с него, също се премахват. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.dataport.de/dprojecttracking?mtm_campaign=dprojecttracking&mtm_source=linkedin&mtm_medium=pp&mtm_content=awareness&mtm_group=rv13&trk=products_details_guest_secondary_call_to_action
dProjectTracking: Projekte agil steuern und umsetzen | Dataport Ihre Cookie-Einstellungen für diese Webseite Treffen Sie hier Ihre persönlichen Einstellungen Erforderliche Cookies Erforderliche Cookies sind für die Nutzung der Webseite zwingend erforderlich. Die Webseite kann ohne diese Cookies nicht richtig funktionieren. Einzelne Cookies Cookiename: cookiehint; Anbieter: Dataport; Zweck: Dieser Cookie speichert die Einstellung zu Cookies; Laufzeit: 1 Jahr Statistik Diese Cookies helfen uns bei der Analyse des Nutzerverhaltens, um die Qualität unserer Webseite fortlaufend zu verbessern. Einzelne Cookies Cookiename: _pk_id.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, wie die Webseite von Besucher*innen genutzt wird; Laufzeit: 12 Monate Cookiename: _pk_ref.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, woher die Besucher*innen dieser Webseite gekommen sind; Laufzeit: 6 Monate Cookiename: _pk_ses.*, _pk_cvar.*; Anbieter: Dataport; Zweck: Erhebung statistischer Daten in Matomo darüber, wie die Webseite von Besucher*innen genutzt wird; Laufzeit: 30 Minuten Weitere Informationen zu Cookies finden Sie in der Datenschutzerklärung Abbrechen Speichern Ihre Daten sind bei uns sicher Als Datenschutzexpert*innen klären wir transparent darüber auf, welche Daten wie und warum verarbeitet werden ( zur Datenschutzerklärung zum Impressum ). Auf dieser Website verwenden wir Cookies. Das sind kleine Dateien, die lokal im Browser gespeichert werden. Einige dieser Cookies sind unbedingt erforderlich, damit Sie diese Website nutzen können und alles richtig funktioniert. Über die Verwendung anderer Cookies entscheiden Sie selbst. Ganz gleich wie Sie sich entscheiden – wir respektieren Ihre Privatsphäre und wir gehen stets verantwortungsvoll mit den Daten um, die Sie uns anvertrauen. Ich entscheide selbst Ich bin einverstanden Ich lehne ab Zur Suche Ansprechpartner*innen finden Störung melden Zu den Jobs Zum Inhalt Menü Unternehmen Dataport Nachhaltigkeit Diversity Digitale Souveränität Geschäftsbericht Organe Partner Services & Produkte Consulting Innovationsmanagement Projektmanagement Rechenzentrum Druckzentrum Software-Entwicklung Weiterbildung Client-Betriebsmodelle IT-Beschaffung SAP-Kompetenzzentrum Fachanwendungen IT-Produkte Unsere Themen Künstliche Intelligenz IT Sicherheit Cloud Onlinezugangsgesetz Open Source Netze Alle Themen Karriere Arbeiten bei Dataport Onboarding Benefits Rund ums Bewerben Qualifizierung Einstieg Aufgabengebiete Veranstaltungen Insights Podcast Jobs Newsroom Übersicht Stories News & Presse Downloads Datareport Support Leichte Sprache Gebärden-sprache English Summary Dialog schließen Ihr direkter Weg zum Ziel Ihr Suchbegriff Suchen Dialog schließen Ansprechpartner*innen Bitte wählen Sie Ihr Anliegen Bitte wählen Pressefragen Karrierefragen Produkte und Vertrieb Internetseite Britta Heinrich Pressesprecherin 040 42846-3047 0171 3342284 E-Mail schreiben Karen Hoffmann-Abraham Stellvertretende Pressesprecherin 040 42846-3736 0176 42860910 E-Mail schreiben Karriereteam E-Mail schreiben Michael Schulz-Heirich Bereichsleiter Vertrieb 040 42846-2881 E-Mail schreiben Online-Redaktion E-Mail schreiben Zentrale: 0431 3295-0 poststelle@dataport.de Mo–Fr: 06:30–18:00 Uhr Dialog schließen Störungsmeldung Region wählen Bitte wählen Hamburg Schleswig-Holstein Bremen Sachsen-Anhalt Kreis Dithmarschen Störungsart wählen Bitte wählen Bei Problemen mit dem BASIS-Standardarbeitsplatz Bei Problemen mit der Schulverwaltungsplattform TUVAS Bei Problemen mit LOGINEO Bei allen anderen IT-Störungen Bei Telefonstörung Bei Mobilfunkproblemen Störungsart wählen Bitte wählen Bei Problemen mit der Plattform ABS Bei Problemen mit dem +1-Arbeitsplatz Bei anderen IT-Störungen Störungsart wählen Bitte wählen Bei Problemen mit dem BASIS-Standardarbeitsplatz Bei Problemen mit dem Fachverfahren MESO Bei allen anderen IT-Störungen Bei Mobilfunkproblemen Störungsart wählen Bitte wählen Bei IT-Problemen mit den StApl-Arbeitsplätzen Bei IT-Problemen aller anderen Kunden Bei IT-Problemen der Polizei ST Störungsart wählen Bitte wählen Bei IT-Störungen Ihre Rufnummer 040 428-333 Ihre Rufnummer 040 42846-3990 Ihre Rufnummer 040 42846-4284 Ihre Rufnummer 040 42846-750 Ihre Rufnummer 040 428-22 Ihre Rufnummer 040 42846-1965 Ihre Rufnummer 0431 3295 7000 Ihre Rufnummer 0431 3295-888 Ihre Rufnummer 0431 3295-444 Ihre Rufnummer 0421 361-55555 Ihre Rufnummer 0421 83558-3559 Ihre Rufnummer 0421 83558-444 Ihre Rufnummer 040 42846-1965 Ihre Rufnummer 0391 6627-1500 Ihre Rufnummer 0391 6627-1545 Ihre Rufnummer 0391-6627-1501 Ihre Rufnummer 0431 3295-3295 Dialog schließen Jobs Zum Jobportal Jetzt Jobalert einrichten Startseite Services & Produkte IT-Produkte dProjectTracking dProjectTracking – Ein Projekt-Tickettool für alle Zwecke Kurz erklärt dProjectTracking basiert auf der Lösung JIRA von Atlassian und bietet Ihnen eine schlanke, webbasierte Ticketlösung. Sie wurde für die Projektsteuerung in agilen Softwareentwicklungsprojekten nach Scrum und KANBAN entwickelt. dProjectTracking eignet sich z. B. zum einfachen Delegieren von Projektaufgaben, im Qualitätsmanagement oder auch im kontinuierlichen Verbesserungsprozess. Sie können sich dieses YouTube-Video direkt auf unserer Internetseite anschauen. Wir weisen darauf hin, dass dabei personenbezogene Daten (z. B. IP-Adresse) an YouTube übermittelt werden. Datenschutz . Video anschauen Viele Projekte, viele Aufgaben Die digitale Transformation sorgt für permanente Veränderung. Veränderung organisiert sich über agile, flexible Aufbau- und Prozessorganisation und entsprechende Projekte. Unser Arbeitsalltag ist zunehmend von Projektaufgaben aus unterschiedlichen Projekten und Aufgabenstellungen bestimmt. Während Projektmanager umfassende Projektpläne, Zeitschienen und Ressourcen im Blick haben müssen, sind auf der Umsetzungsebene andere Aspekte des Projektmanagements entscheidend. Wer hat welche Aufgaben? Mit welchen Fristen? Wie sieht der nächste Scrum-Sprint aus? Und wie organisieren wir das Teamwork im Projekt? Es braucht den richtigen Fokus Agiles Projektmanagement gilt heute als state-of-the-art, und das aus gutem Grund. Die Trennung von Planungs- und Umsetzungsphasen sowie ein schlankes Taskmanagement mit Methoden wie dem Kanban-Board sorgen für Effizienzsteigerungen und für mehr Qualität in den drei zentralen Projektmanagementdisziplinen Geld, Zeit und Ziel. Dies gelingt nur mit einem passenden digitalen Werkzeug zur Zusammenarbeit. Mit dProjectTracking stellt Ihnen Dataport viel Erfahrung aus agilem Projektmanagement zur Verfügung. Auf Basis von Jira nutzen Sie vordefinierte Strukturen, die sich in der Praxis bewährt haben. Mit dem intuitiven Ticketing von Jira lassen sich Projektaufgaben schnell und ad hoc bearbeiten. dProjectTracking sorgt für Transparenz und effizientes Teamwork. Produktvorteile Flexibel Sie steuern Projekte auf der Basis von Scrum und KANBAN. Sie haben alles im Blick und verwalten Aufgaben intuitiv. Sie nutzen vordefinierte Felder, um Akten und Vorgänge zu verknüpfen. Sie können jederzeit den aktuellen Projektstatus nachverfolgen. Attraktiv dProjectTracking wird für Sie im sicheren Dataport Rechenzentrum bereitgestellt. Sie verwalten und steuern differenziert die Zugriffsrechte Ihrer Benutzer. Sie nutzen attraktive Projekt-, User- und Speicherplatzpakete. Sie profitieren von der flexiblen Laufzeit mit nur einem Monat Kündigungsfrist. Vielseitig Sie erfassen und dokumentieren Beschwerden, Innovationsideen, Compliance Vorfälle etc. dProjectTracking ist für jegliche Aufgabenverwaltung geeignet. Sie haben die volle Kontrolle über den gesamten Ticketlebenszyklus. Viel Leistung für wenig Geld dProjectTracking buchen Sie ganz flexibel passgenau zu Ihren Anforderungen. Speicherplatz, Useranzahl, Anzahl der Projekte – Sie bezahlen nur das, was Sie wirklich brauchen. dProjectTracking überzeugt: Über 35 Kunden Über 110 Projekte Über 1.000 Nutzer Probieren Sie es aus. Klingt interessant, aber Sie möchten sich das einmal in Ruhe näher anschauen? Dann sichern Sie sich hier Ihren kostenfreien Testzugang zu dProjectTracking. Kostenlosen Testzugang bestellen Überzeugt? Dann buchen. Sie möchten dProjectTracking zum erfolgreichen und effizienten Projektmanagement nutzen? Hier können Sie Ihr dProjectTracking direkt buchen. Jetzt bestellen Downloads dProjectTracking Auftragsformular (PDF) 2 MB dProjectTracking Produktblatt (PDF) 434 KB Änderungsauftrag (PDF) 945 KB Jetzt mehr Infos und/oder Testzugang anfordern * Pflichtfelder Vorname * Nachname * E-Mail-Adresse * Telefon Behörde/Institution/Firma * Ihre Nachricht Einwilligung für die elektronische Übermittlung * Ich willige in die elektronische Übermittlung meiner Daten gemäß Art. 6 Abs. 1 lit. a DSGVO ein (Opt-In). Eine Weiterverarbeitung der übermittelten Daten erfolgt ggf. auf Basis einer anderen Rechtsgrundlage. Absenden Weitere Informationen finden Sie in der Datenschutzerklärung . Bitte dieses Feld NICHT ausfüllen! Ihr Kontakt Michael Schulz-Heirich Bereichsleiter Vertrieb 040 42846-2881 E-Mail schreiben Team dProjectTracking E-Mail schreiben Weiterführende Informationen Bildnachweise Bildnachweise © Dataport © deagreezstock – stock.adobe.com © Elnur – stock.adobe.com © Nicolas Herrbach – stock.adobe.com © Dataport © Dataport Schwerpunktthemen Open Source IT Sicherheit Onlinezugangsgesetz Cloud Netze Services & Produkte Consulting Innovationsmanagement Projektmanagement Rechenzentrum Druckzentrum Software-Entwicklung Weiterbildung Client-Betriebsmodelle IT-Beschaffung SAP-Kompetenzzentrum Fachanwendungen IT-Produkte Wir machen digitale Zukunft Starten Sie mit uns durch! Karriere bei Dataport Zu den Stellenangeboten Kontakt Dataport Altenholz Standort (Zentrale) Altenholzer Straße 10-14 24161 Altenholz 0431 3295-0 poststelle@dataport.de Kontakt Zertifizierung Portale Kundenportal Jobportal Weiterbildungsportal Shop Vergabeplattform dataport.kommunal Folgen Sie uns Kontakt Impressum AGB Datenschutz Barrierefreiheit Vergabeverfahren Fördermittelprojekte Sitemap
2026-01-13T09:29:19
https://git-scm.com/book/zh-tw/v2/%e9%96%8b%e5%a7%8b-Git-%e5%9f%ba%e7%a4%8e%e8%a6%81%e9%bb%9e
Git - Git 基礎要點 About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. 開始 1.1 關於版本控制 1.2 Git 的簡史 1.3 Git 基礎要點 1.4 命令列 1.5 Git 安裝教學 1.6 初次設定 Git 1.7 取得說明文件 1.8 摘要 2. Git 基礎 2.1 取得一個 Git 倉儲 2.2 紀錄變更到版本庫中 2.3 檢視提交的歷史記錄 2.4 復原 2.5 與遠端協同工作 2.6 標籤 2.7 Git Aliases 2.8 總結 3. 使用 Git 分支 3.1 簡述分支 3.2 分支和合併的基本用法 3.3 分支管理 3.4 分支工作流程 3.5 遠端分支 3.6 衍合 3.7 總結 4. 伺服器上的 Git 4.1 通訊協定 4.2 在伺服器上佈署 Git 4.3 產生你的 SSH 公鑰 4.4 設定伺服器 4.5 Git 常駐程式 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 第3方 Git 託管方案 4.10 總結 5. 分散式的 Git 5.1 分散式工作流程 5.2 對專案進行貢獻 5.3 維護一個專案 5.4 Summary 6. GitHub 6.1 建立帳戶及設定 6.2 參與一個專案 6.3 維護專案 6.4 Managing an organization 6.5 Scripting GitHub 6.6 總結 7. Git 工具 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 總結 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. 附錄 A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. 附錄 B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. 附錄 C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 1.3 開始 - Git 基礎要點 Git 基礎要點 你要如何用幾句話形容 Git? 請仔細閱讀這個重要的章節,如果你瞭解 Git 的本質以及運作的基礎,那麼你將能夠輕鬆且有效率的使用 Git。 在學習之前,試著忘記以前所知道的其它版本控制系統,如:Subversion 及 Perforce。這將會幫助你使用此工具時發生不必要的誤解。 Git 儲存資料及對待資料的方式遠異於其它系統,即使它們的使用者介面是很相似的。 瞭解這些差異會幫助你更準確的使用此工具。 記錄檔案快照,而不是差異 Git 與其它版本控制系統(包含 Subversion 以及與它相關的)最主要的差別是如何處理資料的方式。 一般來說,其他大部分的系統是紀錄一連串檔案更改的資訊。 這些系統(CVS、Subversion、Perforce、Bazaar…等等)儲存一組基本的檔案以及這些檔案隨時間遞增的更動資料。 圖表 4. 將檔案存成版本與版本之間每個檔案的差異。 但是 Git 不是用這種方式儲存及看待這些資料, 而是將其視為小型檔案系統的一組快照(Snapshot)。 每當你提交(commit)(註:在 Git 儲存目前專案的狀態)時,Git 會紀錄下你所有目前檔案的樣子,並且參照到這次快照中。 為了講求效率,只要檔案沒有變更,Git 不會再度儲存該檔案,而是直接將上一次相同的檔案參照到這次快照中。 Git 把它的資料視為一連串的快照。 圖表 5. 將檔案存成許多次的快照。 這是 Git 和其他相似的版本管理系統之間一個重要的差異。 它使得 Git 從各方面重新考量被其他老一輩版本管理系統所採用的作法, 並使得 Git 更像是一個上面有一些強大工具的小型檔案系統,而不僅只是版本管理系統。 本書將會在 [ch03-git-branching] 裡面介紹 Git 分支時,帶領你探索採用這種做法所獲得的好處。 大部份的操作皆可在本地端完成 大部份 Git 的操作皆只需要本地端的檔案及資源即可完成 — 通常並不需要網路上其它電腦的資訊。 如果你以前使用過每項操作都需要網路延遲的集中式版本控制系統,在這方面 Git 將會讓你覺得速度快到有如神助。 因為專案所有的歷史資料都存在你的本機磁碟中,大多數的操作看起來都像是瞬間完成的。 例如:想要瀏覽專案的歷史時,Git 不需要到伺服器下載歷史再顯示 — 就只需要從本機的資料庫讀取。 這意味著你幾乎馬上就可以看到專案的歷史。 若讀者想瞭解某個檔案一個月前的版本與現在版本的差別,Git 可以找出一個月前的檔案並在本機比對差異,而不是要求遠端的伺服器執行這項工作,或者從伺服器取回舊版本的檔案之後才在本機比對。 這也代表你只有一點點操作沒辦法在你斷線或是中斷 VPN 後執行。 如果你在飛機或火車上想要做些小工作,你可以愉快的提交並且等到連上網路後再上傳。 如果你回家後沒辦法使 VPN 正常運作,你仍然可以進行你的工作。 在很多其他的系統上,這麼做通常是不可能或是非常痛苦的事。 以 Perforce 為例,當你連不上伺服器時,你幾乎沒事可做。在 Subversion 和 CVS 中,你可以修改檔案,但是你沒辦法提交版本(只因為你連不上資料庫)。 這看起來可能不是什麼大問題,但是你可能會驚訝於 Git 能做到的事情有這麼大的差異。 Git 能檢查完整性 在 Git 中所有的物件在儲存前都會被計算校驗碼(checksum)並以校驗碼參照物件。 這意味著你不可能瞞著 Git 對任何檔案或目錄進行修改。 此功能內建在 Git 底層並整合到它的設計哲學。 Git 更能夠馬上察覺傳輸時的遺失或是檔案的毀損。 Git 用來計算校驗碼的機制稱為 SHA-1 雜湊演算法。 一個校驗碼是由 40 個 16 進位的字母(0–9 和 a–f)所組成,Git 會根據檔案的內容和資料夾的結構來計算。 一個 SHA-1 校驗碼看起來如下所示: 24b9da6552252987aa493b52f8696cd6d3b00373 你會 Git 中到處都看到校驗碼,因為校驗碼被 Git 到處使用。 事實上在 Git 的資料庫內,每個檔案都是用其內容的校驗碼來儲存,而不是使用檔名。 Git 通常只增加資料 當你使用 Git,幾乎所有的動作都只是增加資料到Git的資料庫。 你很難藉此讓做出讓系統無法復原或者清除資料的動作。 在任何版本控制系統中,你尚未提交的修改都有可能會遺失或者搞亂。 但是只要你提交快照到 Git 後,很難會發生遺失的情況,特別是你定期將資料庫推送(push)到其它儲存庫時,就更不可能會弄丟資料。 這讓我們在使用 Git 的時候時可以像在玩玩具一樣,因為我們知道我們可以隨意操作而不會弄壞任何東西。 在 復原 中,我們會進一步討論 Git 如何儲存資料,以及你如何復原看似遺失的資料。 三種狀態 現在,請特別注意。 若你希望接下來的學習過程順利些,請務必記住以下這些關於 Git 的知識。 Git 會把你的檔案標記為三種主要的狀態:已提交(committed)、已修改(modified)及已預存(staged)。 已提交代表這檔案己安全地存在你的本地端資料庫。 己修改代表這檔案已被修改但尚未提交到本地端資料庫。 已預存代表這檔案將會被存到下次你提交的快照中。 這帶領我們到 Git 專案的三個主要區域:Git 資料夾、工作目錄(working directory)以及預存區(staging area)。 圖表 6. 工作目錄,預存區及 Git 資料夾。 Git 資料夾是 Git 用來儲存你專案的後設資料及物件資料庫的地方。 這是 Git 最重要的部份,而且當你克隆一個其他電腦的儲存庫時,這個資料夾也會被同時複製。 工作目錄是專案被檢出的某一個版本。 這些檔案從 Git 目錄內被壓縮過的資料庫中拉出來並放在硬碟供你使用或修改。 預存區是一個單一檔案,一般來說放在 Git 目錄下,儲存關於下次提交的資訊。 有時它會稱為索引「index」,但現在更常被稱呼為預存區。 基本 Git 工作流程大致如下: 你在你工作目錄修改檔案。 預存檔案,將檔案的快照新增到預存區。 做提交的動作,這會讓存在預存區的檔案快照永久地儲存在 Git 目錄中。 若檔案已被存於 Git 資料夾內,則稱為已提交。 若檔案先被修改,接著被增加到預存區域,則稱為已預存。 若檔案被檢出後有被修改,但未被預存,則稱為已修改。 在 [ch02-git-basics] 內你將會學到更多關於這些狀態的知識以及如何利用它們的優點或者直接略過預存步驟。 prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.rdstation.com/produtos/conversas/?trk=products_details_guest_secondary_call_to_action
Atendimento centralizado para canais digitais é no RD Station Conversas Produtos Segmentos Planos Recursos Contato Assista à demonstração Fale com vendas Entrar Assista à demonstração Fale com vendas Entrar Marketing Digital CRM de Vendas Atendimento e Chatbot Ecommerce RD Station Marketing Automatize e gerencie seu Marketing Digital Conheça mais o produto Mensagem de WhatsApp Aumente sua conversão em até 40% com disparos em massa via WhatsApp Email Marketing Comece do zero ou escolha mais de 100 modelos de email criados por especialistas Landing Page Crie Landing Pages em minutos de forma fácil com nosso editor Automação de Marketing Guie seus contatos em cada etapa da jornada, aumentando a conversão Para Lojas Virtuais Crie uma estrutura de vendas contínua com funcionalidades exclusivas de ecommerce Radar GEO Saiba como ChatGPT e Gemini interpretam sua marca e otimize sua estratégia digital Jornada ideal Explore soluções para suas dores e entenda como enriquecer a jornada da sua audiência Integre o RD Station Marketing Assista à Demonstração Integrações Veja todos os recursos Comece o teste grátis agora RD Station CRM Controle processos e aumente suas vendas Conheça mais o produto Funil de Vendas Organize seus processos em um ou vários Funis de Vendas personalizados CRM de Vendas por WhatsApp Seu CRM de vendas sempre atualizado, sem sair do WhatsApp Web Automação de Vendas Automatize a criação de tarefas, envio de e-mails e mudanças de etapa Gestão de Equipes Gerencie e melhore a performance do seu time em um só lugar Relatórios de vendas Acompanhe seus resultados em tempo real com apoio de IA Radar GEO Saiba como ChatGPT e Gemini interpretam sua marca e otimize sua estratégia digital Jornada ideal Explore soluções para suas dores e entenda como enriquecer a jornada da sua audiência Integre o RD Station CRM Assista a demonstração Integrações Veja todos os recursos Crie uma conta gratuita agora RD Station Conversas Automatize seu atendimento e venda mais Conheça mais o produto Agentes de IA Assistentes inteligentes que automatizam seu atendimento no WhatsApp Dashboards personalizados Relatórios sob medida para acompanhar a performance do seu time Chatbot para WhatsApp Automatize seus atendimentos e crie chatbots que convertem em poucos cliques Atendimento automatizado Aumente a satisfação dos seus clientes com um atendimento rápido e eficiente Campanhas de WhatsApp Programe disparos e conecte os leads aos vendedores com Campanhas de WhatsApp Radar GEO Saiba como ChatGPT e Gemini interpretam sua marca e otimize sua estratégia digital Jornada ideal Explore soluções para suas dores e entenda como enriquecer a jornada da sua audiência Integre com o RD Station Conversas Assista a demonstração Integrações Veja todos os recursos Conheça todas as funcionalidades do RD Station Conversas Expanda sua loja virtual com estrutura eficiente Acelere suas vendas, automatize sua operação e expanda com
todo suporte que precisar Saiba mais RD Station Marketing Crie uma estrutura de vendas contínua com funcionalidades exclusivas de ecommerce Mensagem de WhatsApp Ganhe agilidade com contato rápido e direcione o Lead para o lugar certo Email Marketing Comece do zero ou escolha mais de 100 modelos de email criados por especialistas Catálogo de produtos Personalize sua comunicação com o consumidor e se diferencie Radar GEO Saiba como ChatGPT e Gemini interpretam sua marca e otimize sua estratégia digital Jornada ideal Explore soluções para suas dores e entenda como enriquecer a jornada da sua audiência Fale com vendas Para o seu negócio Marketing + CRM Manifesto RD Station Prisma Conheça os lançamentos do ecossistema RD Station Saiba mais Por que a RD Station? Saiba como te apoiamos Cases de Sucesso Saiba mais Assista à demonstração Aprenda a dar os primeiros passos Assista à demonstração Para o seu negócio Marketing + CRM Manifesto RD Station Prisma Conheça os lançamentos do ecossistema RD Station Saiba mais Porque a RD Station Cases de Sucesso Saiba mais Assista à demonstração Aprenda a dar os primeiros passos Assista Agora Para o seu negócio Manifesto RD Station Prisma Conheça os lançamentos do ecossistema RD Station Saiba mais Porque a RD Station Cases de Sucesso Saiba mais Assista à demonstração Aprenda a dar os primeiros passos Assista Agora Para o seu negócio Manifesto RD Station Prisma Conheça os lançamentos do ecossistema RD Station Saiba mais Para o seu negócio Cases de Sucesso Marketing + CRM Prisma: Lançamentos RD Station Manifesto RD Station Educação Atraia e retenha alunos, aumentando matrículas e fortalecendo sua instituição. Ecommerce Escale seu e-commerce com automação, dados e campanhas que geram mais vendas. Serviços Mostre valor, conquiste clientes e transforme atendimento em fidelização. Saúde Atraia pacientes, fortaleça relacionamentos e otimize a gestão da sua clínica. Varejo Conecte sua marca ao público certo, aumente vendas e crie relações duradouras. Financeiro Aumente a confiança, conquiste clientes e escale resultados no setor financeiro. Distribuição Impulsione sua distribuidora com automação, previsibilidade e clientes fiéis. Estratégias da RD Station Qual o seu principal desafio hoje?
Descubra como podemos te ajudar. Solucionar meu problema RD Station Marketing Para quem quer melhorar sua presença digital Conhecer Planos RD Station CRM Para quem não quer perder nenhuma negociação Conhecer Planos RD Station Conversas Para quem quer centralizar e automatizar atendimentos Conhecer Planos Estratégias da RD Station Soluções da RD para conectar Marketing e Vendas com eficiência Saiba mais Parceiros Aprenda Integrações Parceiros Seja um parceiro Seja uma empresa parceira Conheça nosso programa de parceria e aceleração para agências de marketing e consultarias vendas Programa de afiliados Indique o RD Station Marketing e tenha acesso a benefícios exclusivos Encontre um parceiro Encontre uma empresa parceira As melhores agências e consultorias do Brasil estão no nosso Marketplace. Encontre a ideial para te atender. Aprenda Demonstrações RD Station Marketing Veja na prática como automatizar seu Marketing RD Station CRM Saiba como usar o CRM para fechar mais negócios RD Station Conversas Veja como automatizar e centralizar o atendimento Conteúdos Portal RD Station Insights e práticas para crescer em Marketing e Vendas Materiais Gratuitos Conteúdos práticos para avançar com sua estratégia Ferramentas Gratuitas Conquiste mais produtividade com nossas opções Cursos Universidade RD Station Capacitação para evoluir na carreira e nos negócios Curso de Marketing Digital Gratuito O curso gratuito mais completo e atualizado do mercado PEACE: Metodologia para Gestão de Vendas Aprenda gestão comercial com quem vive na prática Panorama de Marketing e Vendas Veja as principais tendências do mercado brasileiro Saiba mais Integrações Desenvolvedores Appstore RD Station Marketing Encontre aplicativos que se integram ao RD Station Marketing Appstore RD Station CRM Encontre aplicativos que se integram ao RD Station CRM API Center Documentação para apoiar desenvolvedores a construir soluções que vão ajudar sua empresa a crescer Parceiros de integração Integre sua solução ao RD Station, conquiste clientes e amplie negócios Inteligência artificial Agentes de IA Converta e qualifique leads e automatiza fluxos com IA para aumentar a eficiência da sua operação Saiba mais Fale com a gente Para conhecer nossos produtos e planos ou contatar áreas Suporte Está com problemas em algum dos produtos? Entre em contato com nosso suporte Transforme atendimento em resultado com mais eficiência Com o RD Station Conversas, você automatiza e organiza seus atendimentos em canais como WhatsApp, site e redes sociais. Ganhe escala, produtividade e visão total da operação sem perder o controle. Conheça nossa ferramenta na prática Fale com vendas ou assista à demonstração A base para um atendimento moderno, eficiente e escalável Conheça as principais funcionalidades do RD Station Conversas Chat Agentes de IA Chatbots Campanhas Dashboard Categorização de atendimento Qualificação
inteligente Fluxos de
automação LANÇAMENTO Agentes avançados de IA: aumente sua produtividade com bots mais inteligentes! Agora, você pode capacitar seus agentes de IA para atender com mais eficiência, transferindo conversas no momento certo, qualificando leads automaticamente, consultando outros sistemas da empresa e até realizando agendamentos. Saiba mais Empresas que confiam no 
RD Station Conversas Confira nossos cases de sucesso aqui Transforme atendimentos em resultado 80% das empresas que batem metas todos os meses utilizam o WhatsApp nas suas operações. Com o RD Station Conversas, você gerencia seus canais de atendimento, seja qual for a estratégia. Marketing: Para engajar seus leads Engaje clientes com campanhas segmentadas no WhatsApp, qualifique leads automaticamente e monitore os resultados em tempo real para um impacto contínuo. Saiba mais Vendas: Para aumentar conversões Feche mais negócios agilizando negociações, respondendo mais rápido e com todas as conversas em um só lugar para um processo de vendas mais simples e eficiente. Saiba mais Pós-vendas: Para ter clientes satisfeitos Ofereça um atendimento contínuo e de qualidade, reduzindo custos com automação e garantindo maior satisfação com suporte eficaz em todas as etapas do relacionamento. Saiba mais Tecnologia que libera seu time para fazer o que importa Ver todas as funcionalidades Converter Analisar Filtrando . . . Converter Relacionar Chatbot com IA Automatize atendimentos com fluxos naturais e personalizáveis. Converter Engajar Campanhas de WhatsApp Envie mensagens em massa com personalização e rastreabilidade. Analisar Relatórios de atendimento Tenha acesso a relatórios com histórico, TMA, TME, atendente e mais. Analisar Dashboards personalizados Monitore o que importa com dados sob medida. Converter Relacionar Agentes de Atendimento com IA IA treinada nos seus documentos para respostas mais inteligentes. Analisar Origem de contato Entenda de onde vêm seus contatos e otimize as estratégias de captação. Converter Enagajar Qualificação inteligente Configure fluxos automatizados para coletar dados-chave dos contatos e enviar para o funil de vendas. Converter Relacionar Carteiras de atendimento Distribua atendimentos entre seus operadores, definindo responsáveis e expedientes para cada carteira. Analisar Etiquetas Crie categorias para a IA organizar seu atendimento automaticamente. Plano Advanced: mais inteligência e controle na sua operação Conheça o novo Plano Advanced, mais autonomia e inteligência para sua operação com os novos recursos: Agentes de Atendimento com IA e Dashboards Personalizados . Conecte com o que já funciona no seu dia a dia O RD Station Conversas possui integrações nativas com RD Station Marketing, CRM, plataformas de automação e ferramentas de atendimento. Conheça nossas integrações Milhares de empresas já viram a diferença na prática 56% mais vendas em canais digitais 80% dos times que bateram meta usam WhatsApp diariamente 50.000+ gestores já usam o RD Station Conversas Fale com vendas Conquiste os melhores resultados com a melhor ferramenta E-commerce "A RD Station é fundamental para trabalhar em toda a jornada do cliente, isso nos ajudou a definir processos e se tornou um método de trabalho." Renan Nascimento | Gerente Executivo Comercial Assista uma demonstração +20% em vendas Produtos escolhidos RDSM RDSCRM RDSC Saúde "A RD Station entrega informações seguras e dados reais que possibilitam uma tomada de decisão estratégica e geram resultados efetivos." Claudio Bucco | Diretor Assista uma demonstração +20% em vendas Produtos escolhidos RDSM RDSC Educação "A RD Station é a base que nos ajuda a atingir os resultados com escala, controle e visibilidade para todos os indicadores importantes." Rafael Brito | Diretor de Marketing e Comercial Assista uma demonstração +de 20% de crescimento com a RD Station Produtos escolhidos RDSM RDSCRM RDSC E-commerce "A RD Station está presente em todo o nosso ecossistema de relacionamento, e já gerou resultados expressivos desde o primeiro ano." Flávio Aguiar | CEO Assista uma demonstração 85% de crescimento com a RD Station Produtos escolhidos RDSM RDSC E-commerce "A RD Station permite interagir com o cliente no momento certo, focando no momento de compra do cliente e não na capacidade da empresa." Paulo Marcos Pereira dos Santos | Líder de Experiência do Cliente Assista uma demonstração 110% de aumento em sua receita Produtos escolhidos RDSM RDSCRM RDSC Escolha o plano ideal para seu momento Basic Centralize seus canais de atendimento e gerencie todas as conversas. R$1.349 /mês undefined R$ Valor cobrado no Plano Anual Fale com vendas Disparos de mensagens em massa Dashboard de campanhas Dashboard de monitoramento em tempo real da operação Dashboard de atendimentos Chatbot com fluxos de automação básicos Nuvem de palavras Integração nativa com RD Station Marketing, RD Station CRM e Exact Sales, garantindo conexão total com o ecossistema. Até 1 número Franquia de até 1000 atendimentos de clientes únicos por mês Usuários ilimitados Suporte por e-mail e Chat Escolha como quer ser atendimento: chat, e-mail e/ou WhatsApp Pro Melhore a experiência dos clientes com atendimento automatizado. R$2.609 /mês undefined R$ Valor cobrado no Plano Anual Fale com vendas Todas as opções do Basic e: Distribuição de leads com regras personalizadas Origem dos contatos Análise de emoções Funil de vendas Chatbots com fluxos inteligentes, chamadas de API e envio de arquivos. Integrações nativas e via APl com os principais CRMs do mercado, casos de uso avançados com RD Station CRM e integração nativa com Shopify. Possibilidade de 2 ou mais números Franquia de até 3000 atendimentos de clientes únicos por mês Usuários ilimitados Suporte completo Atendimento por chat, e-mail e WhatsApp, além de apoio estratégico de especialistas. Advanced Escale seus resultados com automação e agentes de IA para uma estratégia conversacional avançada. undefined Preço sob consulta Fale com vendas Todas as opções do Pro e: Agentes Copiloto com Inteligência Artificial e habilidades avançadas NOVO Chatbots avançados com qualificação inteligente de leads Categorização automática dos atendimentos com IA, com aplicação de etiquetas inteligentes Dashboards personalizados Tela analítica da origem de contato Inteligência artificial nativa e APis avançadas para integrações exclusivas. Franquia de 5000 atendimentos de clientes únicos por mês Usuários ilimitados Suporte completo Atendimento por chat, e-mail e WhatsApp, além de apoio estratégico de especialistas. Pronto para transformar a forma como
sua empresa atende e vende? Fale com vendas Preencha os campos abaixo e aguarde um dos nossos consultores entrar em contato: Nome* Por favor, preencha seu nome Email corporativo* Por favor, preencha com o seu email corporativo Telefone (Whatsapp de preferência)* Por favor, preencha com o seu telefone Empresa* Por favor, preencha o nome da sua empresa Site da empresa* Por favor, preencha com o site da sua empresa Qual o seu cargo?* Selecione Sócio(a) / CEO / Proprietário(a) Diretor(a) de Vendas Diretor(a) de Marketing Diretor(a) Outras Áreas Gerente de Marketing Gerente de Vendas Coordenador(a)/Supervisor(a) de Marketing Coordenador(a)/Supervisor(a) de Vendas Analista/Assistente de Marketing Analista/Assistente de Vendas Vendedor(a) / Executivo(a) de Contas Estudante Outros Cargos Por favor, preencha seu cargo Quantos funcionários a sua empresa possui?* Selecione 1 2-3 4-10 11-50 51-200 Mais de 200 Por favor, informe o número de atendentes Ao preencher o formulário, concordo * em receber comunicações de acordo com meus interesses. Ao informar meus dados, eu concordo com a Política de privacidade * Você pode alterar suas permissões de comunicação a qualquer tempo. Nos comprometemos em guardar suas informações em segurança. Solicitar contato Pronto, recebemos suas informações Em breve alguém do nosso time fala com você 👋 Enquanto isso, conheça o RD Station Conversas. Centralize canais, automatize o atendimento com agentes de IA e ganhe eficiência. Saiba mais A RD é diferente de tudo. Conseguiu extrapolar a tecnologia. Paulo Sampaio CEO e Head Comercial Além de melhorar nossos resultados, a produtividade também ficou melhor. Thiago dos Santos Especialista em Marketing O meu objetivo profissional é bater as metas dentro da minha empresa. E a RD é a base. Samantha M. CEO na Azul Soluções Preencha os campos abaixo e aguarde um dos nossos consultores entrar em contato: Telefone +55 48 3877-2700 Horário de atendimento: segunda a sexta-feira, das 08h às 12h e das 13h às 18h. Endereço Rod. Virgílio Várzea, 587 - 3º piso, sala 302 - Monte Verde, Florianópolis - SC, 88032-000 © 2026 RD Station. 13.021.784/0001-86 RD Station Marketing Conheça o Produto Preços e Planos Apps e Integrações Cases de Sucesso Suporte - Central de Ajuda Assista à demonstração Funcionalidades Landing Pages Automação de Marketing Email Marketing Lead Scoring Agendamento de Posts Link da Bio RD Station CRM Conheça o Produto Preços e Planos CRM gratuito Integrações Assista à demonstração Funcionalidades Funil de Vendas Vendas por WhatsApp Automação de Vendas Gestão de Equipes Relatórios RD Station Conversas Conheça o produto Preços e planos Cases de Sucesso Assista a Demonstração Suporte - Central de Ajuda Funcionalidades Chatbot para WhatsApp Atendimento Omnichannel Relatório de Atendimento Campanhas de WhatsApp Automação para pós venda Aprenda RD Summit Universidade RD Station Blog RD Station Materiais Gratuitos Agências de Resultados Glossário Parceiros Sobre Agências Afiliados Marketplace Developers Institucional Sobre Nós Trabalhe Conosco Contato Informação Legal e Privacidade Inteligência Artificial Mapa do site Mapa do portal
2026-01-13T09:29:19
https://www.linkedin.com/products/freshworks-inc-freshservice/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_image-click#main-content
Freshservice | LinkedIn Skip to main content LinkedIn Freshworks in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Freshservice Service Desk Software by Freshworks See who's skilled in this Add as skill Get started Report this product About Freshservice is an intelligent, right-sized service management solution for modern businesses of all sizes. Freshservice does this by taking a fresh approach to building and delivering modern employee experiences and unified service management —empowering businesses to achieve efficiency, fast time-to-value, and improved employee satisfaction and productivity. This product is intended for Chief Information Officer Information Technology Manager Information Technology Asset Manager Change Manager Information Technology Operations Manager Chief Technology Officer Vice President Information Technology Head of Information Technology Media Products media viewer No more previous content Freshservice for ITSM Transform IT support into a single, easy-to-use service management solution. With AI-powered tools, automated ticketing, and simple setup, IT teams can finally get back to solving real problems. Freshservice for ITOM Safeguard your business from operational incidents. Track service health, eliminate noise with Freddy AI, automate escalation, & update customers and stakeholders – minus any tool hopping. Freshservice for ITAM Build a backbone for efficient service delivery with complete visibility into your on-premise and cloud infrastructure with IT Asset Management software. Freshservice for Business Teams IT, HR, Facilities, Legal - extending Freshservice across business teams eliminates the need for multiple tools, all while delivering your employees a consistent service environment. No more next content Featured customers of Freshservice EF Education 519,430 followers TaylorMade Golf Company Sporting Goods Manufacturing 172,325 followers HelloFresh Consumer Services 422,334 followers RingCentral Software Development 295,494 followers Chargebee Financial Services 520,368 followers INCAE Business School Higher Education 181,676 followers Payvision Financial Services 14,507 followers VICE Media Media Production 660,568 followers University of Aberdeen Higher Education 155,678 followers NHS Food and Beverage Services 712,120 followers Western Sussex Hospitals NHS Trust Hospitals and Health Care 2,739 followers Addison Lee Transportation, Logistics, Supply Chain and Storage 23,478 followers M+C Saatchi UK Advertising Services 84,163 followers Moneycorp Financial Services 25,699 followers PowerSchool E-Learning Providers 160,054 followers UNiDAYS Advertising Services 18,275 followers Radisson Hotel Group Hospitality 878,390 followers Swinerton Construction 133,048 followers Wake Forest University Higher Education 98,864 followers Trainline Technology, Information and Internet 55,507 followers Show more Show less Similar products Jira Service Management Jira Service Management Service Desk Software Intercom Intercom Service Desk Software Atomicwork Atomicwork Service Desk Software SolarWinds Service Desk SolarWinds Service Desk Service Desk Software Service Desk Service Desk Service Desk Software TOPdesk TOPdesk Service Desk Software Sign in to see more Show more Show less Freshworks products Freshchat Freshchat Live Chat Software Freshdesk Freshdesk Help Desk Software Freshdesk Omni Freshdesk Omni Help Desk Software Freshmarketer Freshmarketer Marketing Automation Software Freshsales Freshsales Customer Relationship Management (CRM) Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/ru/v2/%d0%92%d0%b2%d0%b5%d0%b4%d0%b5%d0%bd%d0%b8%d0%b5-%d0%a3%d1%81%d1%82%d0%b0%d0%bd%d0%be%d0%b2%d0%ba%d0%b0-Git
Git - Установка Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Введение 1.1 О системе контроля версий 1.2 Краткая история Git 1.3 Что такое Git? 1.4 Командная строка 1.5 Установка Git 1.6 Первоначальная настройка Git 1.7 Как получить помощь? 1.8 Заключение 2. Основы Git 2.1 Создание Git-репозитория 2.2 Запись изменений в репозиторий 2.3 Просмотр истории коммитов 2.4 Операции отмены 2.5 Работа с удалёнными репозиториями 2.6 Работа с тегами 2.7 Псевдонимы в Git 2.8 Заключение 3. Ветвление в Git 3.1 О ветвлении в двух словах 3.2 Основы ветвления и слияния 3.3 Управление ветками 3.4 Работа с ветками 3.5 Удалённые ветки 3.6 Перебазирование 3.7 Заключение 4. Git на сервере 4.1 Протоколы 4.2 Установка Git на сервер 4.3 Генерация открытого SSH ключа 4.4 Настраиваем сервер 4.5 Git-демон 4.6 Умный HTTP 4.7 GitWeb 4.8 GitLab 4.9 Git-хостинг 4.10 Заключение 5. Распределённый Git 5.1 Распределённый рабочий процесс 5.2 Участие в проекте 5.3 Сопровождение проекта 5.4 Заключение 6. GitHub 6.1 Настройка и конфигурация учётной записи 6.2 Внесение собственного вклада в проекты 6.3 Сопровождение проекта 6.4 Управление организацией 6.5 Создание сценариев GitHub 6.6 Заключение 7. Инструменты Git 7.1 Выбор ревизии 7.2 Интерактивное индексирование 7.3 Припрятывание и очистка 7.4 Подпись 7.5 Поиск 7.6 Перезапись истории 7.7 Раскрытие тайн reset 7.8 Продвинутое слияние 7.9 Rerere 7.10 Обнаружение ошибок с помощью Git 7.11 Подмодули 7.12 Создание пакетов 7.13 Замена 7.14 Хранилище учётных данных 7.15 Заключение 8. Настройка Git 8.1 Конфигурация Git 8.2 Атрибуты Git 8.3 Хуки в Git 8.4 Пример принудительной политики Git 8.5 Заключение 9. Git и другие системы контроля версий 9.1 Git как клиент 9.2 Переход на Git 9.3 Заключение 10. Git изнутри 10.1 Сантехника и Фарфор 10.2 Объекты Git 10.3 Ссылки в Git 10.4 Pack-файлы 10.5 Спецификации ссылок 10.6 Протоколы передачи данных 10.7 Обслуживание репозитория и восстановление данных 10.8 Переменные окружения 10.9 Заключение A1. Приложение A: Git в других окружениях A1.1 Графические интерфейсы A1.2 Git в Visual Studio A1.3 Git в Visual Studio Code A1.4 Git в Eclipse A1.5 Git в IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.6 Git в Sublime Text A1.7 Git в Bash A1.8 Git в Zsh A1.9 Git в PowerShell A1.10 Заключение A2. Приложение B: Встраивание Git в ваши приложения A2.1 Git из командной строки A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Приложение C: Команды Git A3.1 Настройка и конфигурация A3.2 Клонирование и создание репозиториев A3.3 Основные команды A3.4 Ветвление и слияния A3.5 Совместная работа и обновление проектов A3.6 Осмотр и сравнение A3.7 Отладка A3.8 Внесение исправлений A3.9 Работа с помощью электронной почты A3.10 Внешние системы A3.11 Администрирование A3.12 Низкоуровневые команды 2nd Edition 1.5 Введение - Установка Git Установка Git Прежде чем использовать Git, вы должны установить его на своём компьютере. Даже если он уже установлен, наверное, это хороший повод, чтобы обновиться до последней версии. Вы можете установить Git из собранного пакета или другого установщика, либо скачать исходный код и скомпилировать его самостоятельно. Примечание В этой книге используется Git версии 2.8.0 . Хотя большинство используемых нами команд должны работать даже в старых версиях Git, некоторые из них могут не работать или действовать немного иначе, если вы используете старую версию. Поскольку Git отлично справляется с сохранением обратной совместимости, любая версия после 2.8 должна работать нормально. Установка в Linux Если вы хотите установить Git под Linux как бинарный пакет, это можно сделать, используя обычный менеджер пакетов вашего дистрибутива. Если у вас Fedora (или другой похожий дистрибутив, такой как RHEL или CentOS), можно воспользоваться dnf : $ sudo dnf install git-all Если же у вас дистрибутив, основанный на Debian, например, Ubuntu, попробуйте apt : $ sudo apt install git Чтобы воспользоваться дополнительными возможностями, посмотрите инструкцию по установке для нескольких различных разновидностей Unix на сайте Git https://git-scm.com/download/linux . Установка на Mac Существует несколько способов установки Git на Mac. Самый простой — установить Xcode Command Line Tools. В версии Mavericks (10.9) и выше вы можете добиться этого просто первый раз выполнив 'git' в терминале. $ git --version Если Git не установлен, вам будет предложено его установить. Если Вы хотите получить более актуальную версию, то можете воспользоваться бинарным установщиком. Установщик Git для OS X доступен для скачивания с сайта Git https://git-scm.com/download/mac . Рисунок 7. OS X инсталлятор Git Установка в Windows Для установки Git в Windows также имеется несколько способов. Официальная сборка доступна для скачивания на официальном сайте Git. Просто перейдите на страницу https://git-scm.com/download/win , и загрузка запустится автоматически. Обратите внимание, что это отдельный проект, называемый Git для Windows; для получения дополнительной информации о нём перейдите на https://gitforwindows.org . Для автоматической установки вы можете использовать пакет Git Chocolatey . Обратите внимание, что пакет Chocolatey поддерживается сообществом. Установка из исходников Многие предпочитают устанавливать Git из исходников, поскольку такой способ позволяет получить самую свежую версию. Обновление бинарных инсталляторов, как правило, немного отстаёт, хотя в последнее время разница не столь существенна. Если вы действительно хотите установить Git из исходников, у вас должны быть установлены следующие библиотеки, от которых он зависит: autotools, curl, zlib, openssl, expat и libiconv. Например, если в вашей системе используется dnf (Fedora) или apt-get (системы на базе Debian), вы можете использовать одну из следующих команд для установки всех зависимостей, используемых для сборки и установки бинарных файлов Git: $ sudo dnf install dh-autoreconf curl-devel expat-devel gettext-devel \ openssl-devel perl-devel zlib-devel $ sudo apt-get install dh-autoreconf libcurl4-gnutls-dev libexpat1-dev \ gettext libz-dev libssl-dev Для того, чтобы собрать документацию в различных форматах (doc, html, info), понадобится установить дополнительные зависимости: $ sudo dnf install asciidoc xmlto docbook2X $ sudo apt-get install asciidoc xmlto docbook2x Примечание Пользователи RHEL и производных от неё (таких как CentOS или Scientific Linux) должны подключить репозиторий EPEL для корректной установки пакета docbook2X Если вы используете систему на базе Debian (Debian/Ubuntu/Ubuntu-производные), вам так же понадобится установить пакет install-info : $ sudo apt-get install install-info Если вы используете систему на базе RPM (Fedora/RHEL/RHEL-производные), вам так же понадобится установить пакет getopt , который уже установлен в системах на базе Debian: $ sudo dnf install getopt К тому же из-за различий имён бинарных файлов вам понадобится сделать следующее: $ sudo ln -s /usr/bin/db2x_docbook2texi /usr/bin/docbook2x-texi Когда все необходимые зависимости установлены, вы можете пойти дальше и скачать самый свежий архив с исходниками из следующих мест: с сайта Kernel.org https://www.kernel.org/pub/software/scm/git , или зеркала на сайте GitHub https://github.com/git/git/releases . Конечно, немного проще скачать последнюю версию с сайта GitHub, но на странице kernel.org релизы имеют подписи, если вы хотите проверить, что скачиваете. Затем скомпилируйте и установите: $ tar -zxf git-2.8.0.tar.gz $ cd git-2.8.0 $ make configure $ ./configure --prefix=/usr $ make all doc info $ sudo make install install-doc install-html install-info После этого вы можете получать обновления Git посредством самого Git: $ git clone git://git.kernel.org/pub/scm/git/git.git prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/commands/cargo-publish.html#description
cargo publish - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book cargo-publish(1) NAME cargo-publish — Upload a package to the registry SYNOPSIS cargo publish [ options ] DESCRIPTION This command will create a distributable, compressed .crate file with the source code of the package in the current directory and upload it to a registry. The default registry is https://crates.io . This performs the following steps: Performs a few checks, including: Checks the package.publish key in the manifest for restrictions on which registries you are allowed to publish to. Create a .crate file by following the steps in cargo-package(1) . Upload the crate to the registry. The server will perform additional checks on the crate. The client will poll waiting for the package to appear in the index, and may timeout. In that case, you will need to check for completion manually. This timeout does not affect the upload. This command requires you to be authenticated using cargo-login(1) or environment variables of the registry.token and registries.<name>.token config fields. See the reference for more details about packaging and publishing. OPTIONS Publish Options --dry-run Perform all checks without uploading. --no-verify Don’t verify the contents by building them. --allow-dirty Allow working directories with uncommitted VCS changes to be packaged. --index index The URL of the registry index to use. --registry registry Name of the registry to publish to. Registry names are defined in Cargo config files . If not specified, and there is a package.publish field in Cargo.toml with a single registry, then it will publish to that registry. Otherwise it will use the default registry, which is defined by the registry.default config key which defaults to crates-io . Package Selection By default, when no package selection options are given, the packages selected depend on the selected manifest file (based on the current working directory if --manifest-path is not given). If the manifest is the root of a workspace then the workspaces default members are selected, otherwise only the package defined by the manifest will be selected. The default members of a workspace can be set explicitly with the workspace.default-members key in the root manifest. If this is not set, a virtual workspace will include all workspace members (equivalent to passing --workspace ), and a non-virtual workspace will include only the root crate itself. -p spec … --package spec … Publish only the specified packages. See cargo-pkgid(1) for the SPEC format. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. --workspace Publish all members in the workspace. --all Deprecated alias for --workspace . --exclude SPEC … Exclude the specified packages. Must be used in conjunction with the --workspace flag. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. Compilation Options --target triple Publish for the specified target architecture. Flag may be specified multiple times. The default is the host architecture. The general format of the triple is <arch><sub>-<vendor>-<sys>-<abi> . Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. This may also be specified with the build.target config value . Note that specifying this flag makes Cargo run in a different mode where the target artifacts are placed in a separate directory. See the build cache documentation for more details. --target-dir directory Directory for all generated artifacts and intermediate files. May also be specified with the CARGO_TARGET_DIR environment variable, or the build.target-dir config value . Defaults to target in the root of the workspace. Feature Selection The feature flags allow you to control which features are enabled. When no feature options are given, the default feature is activated for every selected package. See the features documentation for more details. -F features --features features Space or comma separated list of features to activate. Features of workspace members may be enabled with package-name/feature-name syntax. This flag may be specified multiple times, which enables all specified features. --all-features Activate all available features of all selected packages. --no-default-features Do not activate the default feature of the selected packages. Manifest Options --manifest-path path Path to the Cargo.toml file. By default, Cargo searches for the Cargo.toml file in the current directory or any parent directory. --locked Asserts that the exact same dependencies and versions are used as when the existing Cargo.lock file was originally generated. Cargo will exit with an error when either of the following scenarios arises: The lock file is missing. Cargo attempted to change the lock file due to a different dependency resolution. It may be used in environments where deterministic builds are desired, such as in CI pipelines. --offline Prevents Cargo from accessing the network for any reason. Without this flag, Cargo will stop with an error if it needs to access the network and the network is not available. With this flag, Cargo will attempt to proceed without the network if possible. Beware that this may result in different dependency resolution than online mode. Cargo will restrict itself to crates that are downloaded locally, even if there might be a newer version as indicated in the local copy of the index. See the cargo-fetch(1) command to download dependencies before going offline. May also be specified with the net.offline config value . --frozen Equivalent to specifying both --locked and --offline . --lockfile-path PATH Changes the path of the lockfile from the default ( <workspace_root>/Cargo.lock ) to PATH . PATH must end with Cargo.lock (e.g. --lockfile-path /tmp/temporary-lockfile/Cargo.lock ). Note that providing --lockfile-path will ignore existing lockfile at the default path, and instead will either use the lockfile from PATH , or write a new lockfile into the provided PATH if it doesn’t exist. This flag can be used to run most commands in read-only directories, writing lockfile into the provided PATH . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #14421 ). Miscellaneous Options -j N --jobs N Number of parallel jobs to run. May also be specified with the build.jobs config value . Defaults to the number of logical CPUs. If negative, it sets the maximum number of parallel jobs to the number of logical CPUs plus provided value. If a string default is provided, it sets the value back to defaults. Should not be 0. --keep-going Build as many crates in the dependency graph as possible, rather than aborting the build on the first one that fails to build. For example if the current package depends on dependencies fails and works , one of which fails to build, cargo publish -j1 may or may not build the one that succeeds (depending on which one of the two builds Cargo picked to run first), whereas cargo publish -j1 --keep-going would definitely run both builds, even if the one run first fails. Display Options -v --verbose Use verbose output. May be specified twice for “very verbose” output which includes extra output such as dependency warnings and build script output. May also be specified with the term.verbose config value . -q --quiet Do not print cargo log messages. May also be specified with the term.quiet config value . --color when Control when colored output is used. Valid values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. May also be specified with the term.color config value . Common Options + toolchain If Cargo has been installed with rustup, and the first argument to cargo begins with + , it will be interpreted as a rustup toolchain name (such as +stable or +nightly ). See the rustup documentation for more information about how toolchain overrides work. --config KEY=VALUE or PATH Overrides a Cargo configuration value. The argument should be in TOML syntax of KEY=VALUE , or provided as a path to an extra configuration file. This flag may be specified multiple times. See the command-line overrides section for more information. -C PATH Changes the current working directory before executing any specified operations. This affects things like where cargo looks by default for the project manifest ( Cargo.toml ), as well as the directories searched for discovering .cargo/config.toml , for example. This option must appear before the command name, for example cargo -C path/to/my-project build . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #10098 ). -h --help Prints help information. -Z flag Unstable (nightly-only) flags to Cargo. Run cargo -Z help for details. ENVIRONMENT See the reference for details on environment variables that Cargo reads. EXIT STATUS 0 : Cargo succeeded. 101 : Cargo failed to complete. EXAMPLES Publish the current package: cargo publish SEE ALSO cargo(1) , cargo-package(1) , cargo-login(1)
2026-01-13T09:29:19
https://git-scm.com/book/pl/v2/Git-na-serwerze-Podsumowanie
Git - Podsumowanie About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pierwsze kroki 1.1 Wprowadzenie do kontroli wersji 1.2 Krótka historia Git 1.3 Podstawy Git 1.4 Linia poleceń 1.5 Instalacja Git 1.6 Wstępna konfiguracja Git 1.7 Uzyskiwanie pomocy 1.8 Podsumowanie 2. Podstawy Gita 2.1 Pierwsze repozytorium Gita 2.2 Rejestrowanie zmian w repozytorium 2.3 Podgląd historii rewizji 2.4 Cofanie zmian 2.5 Praca ze zdalnym repozytorium 2.6 Tagowanie 2.7 Aliasy 2.8 Podsumowanie 3. Gałęzie Gita 3.1 Czym jest gałąź 3.2 Podstawy rozgałęziania i scalania 3.3 Zarządzanie gałęziami 3.4 Sposoby pracy z gałęziami 3.5 Gałęzie zdalne 3.6 Zmiana bazy 3.7 Podsumowanie 4. Git na serwerze 4.1 Protokoły 4.2 Uruchomienie Git na serwerze 4.3 Generowanie Twojego publicznego klucza SSH 4.4 Konfigurowanie serwera 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Inne opcje hostowania przez podmioty zewnętrzne 4.10 Podsumowanie 5. Rozproszony Git 5.1 Rozproszone przepływy pracy 5.2 Wgrywanie zmian do projektu 5.3 Utrzymywanie projektu 5.4 Podsumowanie 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Narzędzia Gita 7.1 Wskazywanie rewizji 7.2 Interaktywne używanie przechowali 7.3 Schowek i czyszczenie 7.4 Signing Your Work 7.5 Searching 7.6 Przepisywanie historii 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugowanie z Gitem 7.11 Moduły zależne 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Podsumowanie 8. Dostosowywanie Gita 8.1 Konfiguracja Gita 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git i inne systemy 9.1 Git jako klient 9.2 Migracja do Gita 9.3 Podsumowanie 10. Mechanizmy wewnętrzne w Git 10.1 Komendy typu plumbing i porcelain 10.2 Obiekty Gita 10.3 Referencje w Git 10.4 Spakowane pliki (packfiles) 10.5 Refspec 10.6 Protokoły transferu 10.7 Konserwacja i odzyskiwanie danych 10.8 Environment Variables 10.9 Podsumowanie A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.10 Git na serwerze - Podsumowanie Podsumowanie Istnieje kilka sposobów na stworzenie repozytorium Gita, w celu kooperacji z innymi lub dzielenia się swoją pracą. Postawienie własnego serwera daje Ci sporą kontrolę i umożliwia działanie serwera za własnym firewallem, ale taki serwer na ogół wymaga sporo czasu na stworzenie i utrzymanie. Jeśli umieścisz swoje dane na gotowym hostingu, to jest to łatwe do skonfigurowania i utrzymania, ale musisz być w stanie utrzymać swój kod na cudzych serwerach, a niektóre organizacje na to nie pozwalają. Określenie, które rozwiązanie lub połączenie rozwiązań jest odpowiednie dla Ciebie i Twojej organizacji powinno być dość proste. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://uk.linkedin.com/company/optalysys?trk=organization_guest_main-feed-card_feed-actor-image
Optalysys | LinkedIn Skip to main content LinkedIn Articles People Learning Jobs Games Sign in Join now for free Optalysys Computer Hardware Manufacturing Leeds, England 3,604 followers Enabling a world of secure computing. Follow View all 61 employees Report this company About us 🔒 Enabling a world of secure computing. Harnessing the power of light, we push the boundaries of what is possible in data security by bringing speed to always-encrypted data technology. Through optical computing, we accelerate Fully Homomorphic Encryption (FHE) beyond the capabilities of electrical computing, making secure data sharing and collaboration a reality for everyone, everywhere. Website http://www.optalysys.com External link for Optalysys Industry Computer Hardware Manufacturing Company size 51-200 employees Headquarters Leeds, England Type Privately Held Founded 2013 Specialties deep learning, hardware, processors, optical, Quantum Resilient Encryption, FHE, Photonic computing, Full Homomorphic Encryption, and Cryptography Locations Primary New Station Street Platform Leeds, England LS1 4JB, GB Get directions Victoria Street Runway East Bristol, England BS1 6PU, GB Get directions Employees at Optalysys Peter White Phil Cheetham Tim Wilkinson Tim Whitfield See all employees Updates Optalysys 3,604 followers 2w Report this post ❄️ Season's greetings from the team here at Optalysys ❄️ 36 1 Comment Like Comment Share Optalysys 3,604 followers 1mo Edited Report this post 💂🏼♀️ Greetings from London: we're here at Economic Crime Congress 2025 with 600+ leaders of government, industry, and civil society focusing on real-world collaboration... 📍 Find us on stand 6 - come for the swag bag, stay for the privacy-enhancing technologies chat. And head to the Connect Zone at 1.20pm for Nic Lawrence 's talk on that exact topic! UK Finance | #ECC2025 | QEII Centre | Joseph Wilson | Robbyn Snow 30 2 Comments Like Comment Share Optalysys reposted this Nic Lawrence 1mo Report this post Tomorrow I will be talking at the Economic Crime Congress in London - outlining how Optalysys technology can help businesses and public sector groups collaborate securely. The setting at the QE2 centre, just round the corner from the Palace of Westminster is spectacular! Together with Robbyn Snow and Joseph Wilson we will also have a stand and look forward to some good conversations! #securecomputing #fhe 71 Like Comment Share Optalysys 3,604 followers 1mo Report this post 📍 Next stop: tomorrow's UK Finance Economic Crime Congress 2025 at QEII Centre 💷 Find us on stand 6 (floor 2) all day, and then in the Connect Zone at 1.20pm, where VP of Product, Nic Lawrence , will be giving a lightning talk on ‘Privacy-enhancing technologies (PETs) for real-world intelligence sharing’. The team in attendance is Joseph Wilson , Nic Lawrence and Robbyn Snow . Come and chat to us about your privacy problems 💬 #ecc2025 #economiccrime #PETs #privacy 26 Like Comment Share Optalysys 3,604 followers 1mo Report this post 🏆 We’ve been named as a TechRound Top 100 Startup in UK & Europe! Recognition in the TechRound100 2025 highlights our groundbreaking approach to secure, scalable computing across the UK and Europe 🌟 and we are incredibly proud to be included.      Read the full announcement: https://lnkd.in/ek3rBX7x   Read the full entry: https://lnkd.in/eJkUPFh6 View the full list: https://lnkd.in/eZf4i4ru Thank you to TechRound and esteemed judges Mary Pedler ( INPUT Global ), Andrey Kuznetsov ( Haqq Network ), SAMUEL OLAMIDE ( EdgeTrace ), Furkat Kasimov , Andrei Komissarov ( OpenWay and @DEVAR), Ayoola Samagbeyi , and David Soffer ( TechRound ). #TechRound100 #tech #leaders   Optalysys Named Among TechRound's Top 100 Startups in UK & Europe - Optalysys https://optalysys.com 53 1 Comment Like Comment Share Optalysys 3,604 followers 1mo Report this post This time last week we were wrapping up an incredible week in Buenos Aires for DevConnect ARG 2025 🇦🇷 It was an amazing opportunity to connect with builders, innovaters, friends and new faces from across the Ethereum ecosystem. Our key takeaways: - Privacy has landed and it’s here to stay  - Privacy & scalability are the bridge to enterprise adoption - Crypto 🤝 AI: it's happening! - That Buenos Aires was the perfect backdrop Read on for our full roundup ⬇️ DevConnect ARG 🇦🇷 2025 roundup: Privacy + scalability = Ethereum for the masses Optalysys on LinkedIn 21 2 Comments Like Comment Share Optalysys 3,604 followers 1mo Edited Report this post Happy Thanksgiving to our friends in North America! 🍂 🦃 💛 🧡 🤎 🙏 Thank you for all you do to help Optalysys' reach to the world 🌎 #gratitude #thanksgiving #colleagues 20 2 Comments Like Comment Share Optalysys 3,604 followers 1mo Report this post 💷 What does the 2025 Autumn Budget mean for UK businesses, investment, and growth? Stuart Mellis , CFO of Optalysys, comments on the future of Business Asset Disposal Relief 🔴 #ukbudget #leak #announcement FieldHouse Associates 2,145 followers 1mo 🚰 The mistaken publication of the Chancellor's plans an hour before her speech today solidified this Budget's position as the leakiest in recent memory... But after a tumultuous road to the announcement, what does the 2025 Autumn Budget mean for UK businesses, investment, and growth? A piece by Patricia Cullen in Entrepreneur UK sets out the initial response from business leaders and investors to the policies announced 🗣️ 📌 Martin Jacob , Professor of Accounting and Control at IESE Business School explores the increasing complexity of the UK tax system. 📌 Timothy Mills , Managing Partner at ACF Investors , delves into the positive decision to raise EIS limits. 📌 Duncan Johnson , CEO of Northern Gritstone and Stuart Mellis , CFO of Optalysys , comment on the future of Business Asset Disposal Relief. 📌 Tanya Suarez Ph.D. , CEO of IoT Tribe discusses the UK's competitive position in the global tech sector. 📌 And John Dunlop , Head of Tax at Marriott Harrison , explores the proposed reduction in EOT tax relief. Read the full article via the link: https://lnkd.in/eB7qDcXz Budget Meets Backlash | Entrepreneur entrepreneur.com 14 Like Comment Share Optalysys 3,604 followers 1mo Edited Report this post 🟨 We just popped in to Zama 's CoFHE Shop at #devconnect , in preparation for our session tomorrow. Come for the caffeine, stay for the talks ☕ 💬 Hardware Accelerated FHE in Practice: Talk + Technical Workshop 💬 Featuring Nic Lawrence (VP Product, Optalysys) and Joseph Wilson (Head of Strategic Innovation, Optalysys). 📍 Rosedal Café, Godoy Cruz 2905 ⏰ TUES 18th Nov | 14:00 - 15:30 🎟️ https://luma.com/9vfkncep In part one: Join Nic as he shares... - a detailed look at the first enterprise-grade FHE acceleration server - an analysis of some DeFi use cases and how SW tools can help developers and architects to predict system performance and hardware requirements (tools are available to try) - a look ahead to an exciting photonic future for confidential computing In part two: Joe will lead a technical workshop, including... - showcasing the Optalysys testnet, built on Zama’s fhEVM architecture + Optalysys’ LightLocker Node FHE co-processor - an encrypted online game, using a dApp running on the Optalysys testnet. Everyone is invited to join us for this interactive session. The team looks forward to sharing ideas and discussing real-world deployment of confidential blockchains over coffee and snacks. 🟡 About the Zama CoFHE Shop: Your favorite homebase is back for Ethereum Foundation 's DevConnect ARG week. The Zama CoFHE Shop is open from Monday, November 17th to Friday, November 21st, 7.30AM to 8.30PM (with evening afterwork events). Come enjoy good coffee, attend cool meetups and meet the Zama team. #FHE #coffee #talktalktalk #hardware 38 Like Comment Share Optalysys 3,604 followers 1mo Report this post 📍DevConnect is a go! 👋 from Marcella Arthur Aditi Vyakarnam Joseph Wilson Nic Lawrence and Robbyn Snow . 🆕 Track our trip on our new #telegram (🔎Optalysys): @TeamOptalysys. 24 Like Comment Share Join now to see what you are missing Find people you know at Optalysys Browse recommended jobs for you View all updates, news, and articles Join now Similar pages Salience Labs Computer Hardware Manufacturing Oxford, England Lumai Computer Hardware Manufacturing Lightelligence Computer Hardware Manufacturing Boston, MA Niobium Microsystems Computer Hardware Manufacturing Dayton, Ohio Wave Photonics Semiconductors Cambridge, UK Belfort Data Security Software Products Fractile Computer Hardware Manufacturing London, London Inpher Software Development New York, NY Heronic Technologies Software Development London, London Cornami, Inc. Computer Hardware Show more similar pages Show fewer similar pages Browse jobs Program Manager jobs 13,502 open jobs Analyst jobs 63,799 open jobs Engineer jobs 79,088 open jobs Customer Engineer jobs 27,091 open jobs Portfolio Manager jobs 6,625 open jobs Vice President jobs 12,303 open jobs Software Engineering Manager jobs 4,717 open jobs Director jobs 156,873 open jobs Engineering Manager jobs 15,782 open jobs Design Verification Engineer jobs 496 open jobs President jobs 2,134 open jobs Chief Executive Officer jobs 12,046 open jobs Physicist jobs 138 open jobs Optical Engineer jobs 1,960 open jobs Senior Project Manager jobs 12,345 open jobs Scientist jobs 4,117 open jobs Developer jobs 33,470 open jobs Communications Manager jobs 7,892 open jobs Head of Sales jobs 5,319 open jobs Editor jobs 2,723 open jobs Show more jobs like this Show fewer jobs like this Funding Optalysys 4 total rounds Last Round Series A Aug 17, 2023 External Crunchbase Link for last round of funding US$ 27.5M Investors Lingotto imec.xpand + 2 Other investors See more info on crunchbase More searches More searches Delivery Director jobs Technical Program Manager jobs Director jobs Delivery Manager jobs Engineering Manager jobs Chief Technology Officer jobs Revenue Officer jobs Chief Product Officer jobs Junior Scientist jobs Firmware Engineer jobs Head of Product jobs Account Manager jobs Technology Officer jobs Machine Learning Engineer jobs Embedded Software Engineer jobs Embedded System Engineer jobs Engineer jobs Vice President jobs Scientist jobs Senior Project Manager jobs Developer jobs Project Manager jobs Senior Product Manager jobs Design Verification Engineer jobs Vice President of Engineering jobs Software Engineering Manager jobs Chief Information Officer jobs Hardware Engineering Manager jobs Vice President Research And Development jobs Product Manager jobs Associate Product Manager jobs Marketing Executive jobs Architect jobs Campaign Manager jobs Marketing Manager jobs Optical Engineer jobs Director of Information Technology jobs Data Engineer jobs Robotics Engineer jobs Writer jobs Integration Engineer jobs Design Engineer jobs Senior Scientist jobs Researcher jobs Quantitative Researcher jobs Quantitative Analyst jobs Director of Software jobs Head of Software jobs Senior Recruitment Consultant jobs Software Test Engineer jobs Content Manager jobs Android Developer jobs Python Developer jobs Head of Content jobs Head of Information Technology jobs Product Designer jobs LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Sign in to see who you already know at Optalysys Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy .
2026-01-13T09:29:19
https://il.linkedin.com/company/fhenix
Fhenix | LinkedIn דילוג לתוכן הראשי ‏LinkedIn מאמרים אנשים למידה משרות משחקים התחברות להירשם עכשיו Fhenix Software Development Programmable. Scalable. Encrypted. מעקב הצגת ‏כל ‏54‏ העובדים‏ דווח על חברה זו עלינו Fhenix extends Ethereum’s capabilities with Fully Homomorphic Encryption (FHE) Coprocessors. Leveraging computation over encrypted data, Fhenix gives blockchain developers the power to build end-to-end encrypted applications for the first time אתר אינטרנט https://www.fhenix.io/ קישור חיצוני עבור Fhenix תעשייה Software Development גודל החברה 11-50 עובדים משרדים ראשיים Tel Aviv סוג בבעלות פרטית הקמה 2023 מיקומים הראשי Tel Aviv, IL קבלת הוראות הגעה עובדים ב- Fhenix Guy Itzhaki Doron Zarchy Tovi Levis Giancarlo Di Bella צפייה בכל עובדים עדכונים ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Optalysys ‏‏3,604‏ עוקבים‏ ‏2‏חודש דיווח על פוסט הזה DevConnect ARG is loading... ↻ and we're proud to be partnering with Fhenix for Encryption Day. Buenos Aires, here we come #privacybydesign Our sponsor cohort: Arbitrum | Ethereum Foundation | Horizen | Artifact Systems | Hyli | Silhouette Eyes on for more details coming soon... X 🔗 https://lnkd.in/e4uRuzHh #devconnect #encryptionday #privacybydesign 15 ‏1‏ ‏תגובה‏ לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש נערך דיווח על פוסט הזה Encryption Day Returns November 19! 👩💻 Live coding sessions and demos from the Ethereum Foundation , Fhenix , and more from 10:00–12:00 🎤 Panels and presentations from some of the brightest minds in the Ethereum ecosystem all day! ( RAILGUN DAO , MetaCartel (metacartel.org) 1kx , + more) 🍪 Argentinian treats and good company! 🔗 Get your ticket and share with friends! https://luma.com/e9ces3lr 🙌 Brought to you by Optalysys , Artifact Staking , Hyli , zkVerify , Arbitrum , & Silhouette Exchange ! Encryption Day · Luma luma.com 5 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש דיווח על פוסט הזה Every ecosystem has a moment where privacy stops being optional. For Japan, that moment starts now... Japan Joins the Encrypted Frontier: Our Strategic Raise with BIPROGY & TransLink Capital ‏Fhenix‏ ב-LinkedIn 24 ‏2‏ ‏תגובות‏ לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏2‏חודש נערך דיווח על פוסט הזה Come hack at ETHSamba 's ETHLatam Hackathon, with a total bounty of 25k+ USD in prizes! 📅 November 6 - 8 🔗 https://lnkd.in/gq9rJrkR ETHSamba ‏‏466‏ עוקבים‏ ‏2‏חודש נערך Privacidade não morreu — está sendo reprogramada. Com computação criptografada para Ethereum, a Fhenix traz lógica privada por padrão. 20 000× mais throughput, 37× menos latência — performance digna do playground fintech que o Brasil se tornou. Bora pra ETH LATAM 4 לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Optalysys ‏‏3,604‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה 🗞️ In the news: FHE research from the Fhenix camp! Guy Itzhaki CEO, Fhenix; Bringing Encryption to Ethereum; @GuyItzh ‏3‏חודש Today marks a significant milestone for both Fhenix and the broader FHE community — our newest research on highly efficient FHE decryption is now live. Why does this matter? Because it represents a breakthrough in how FHE decryption can scale securely and efficiently, paving the way for real-world, high-throughput confidential computation. The challenge we solved: For years, delivering a secure and efficient FHE decryption method has been one of the hardest open problems in the field. Existing approaches were slow, costly, or required significant computational resources, making them impractical for DeFi-scale workloads. Our new Threshold FHE Decryption protocol eliminates these limitations — delivering both security and speed at levels previously thought unattainable. We achieved: • 20,000× higher throughput • 37× lower latency • Real-time encrypted computation — unlocking Private DeFi Next Week, we will present our paper, “High-Throughput Universally Composable Threshold FHE Decryption,” at ACM CCS 2025 — one of the world’s top cryptography conferences. If you're attending, be sure to hear Guy Zyskind talk. Link to the paper: https://lnkd.in/gPkM3RAd High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 9 לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Guy Itzhaki ‏3‏חודש דיווח על פוסט הזה Today marks a significant milestone for both Fhenix and the broader FHE community — our newest research on highly efficient FHE decryption is now live. Why does this matter? Because it represents a breakthrough in how FHE decryption can scale securely and efficiently, paving the way for real-world, high-throughput confidential computation. The challenge we solved: For years, delivering a secure and efficient FHE decryption method has been one of the hardest open problems in the field. Existing approaches were slow, costly, or required significant computational resources, making them impractical for DeFi-scale workloads. Our new Threshold FHE Decryption protocol eliminates these limitations — delivering both security and speed at levels previously thought unattainable. We achieved: • 20,000× higher throughput • 37× lower latency • Real-time encrypted computation — unlocking Private DeFi Next Week, we will present our paper, “High-Throughput Universally Composable Threshold FHE Decryption,” at ACM CCS 2025 — one of the world’s top cryptography conferences. If you're attending, be sure to hear Guy Zyskind talk. Link to the paper: https://lnkd.in/gPkM3RAd High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 33 ‏4‏ ‏תגובות‏ לייק תגובה שיתוף ‏ ‏Fhenix‏ ‏ פרסם מחדש את זה Blockchain Futurist Conference ‏‏6,284‏ עוקבים‏ ‏3‏חודש נערך דיווח על פוסט הזה ⚡ Your newly added speakers are here! ⚡ 👇 Just 5 weeks until we hear from these Web3 & AI legends 👇 ◾ Iggy Azalea, Rapper & Songwriter ◾ Jake Antifaev , Co-Founder & CEO, Thrust ◾ Webster Barnaby , Representative for HD-29, Florida House of Representatives ◾ Abhay M., SVP of Corporate Development and Investment, MoonPay ◾ André Beganski , Staff Reporter, Decrypt ◾ AnewbiZ 0x , Chief of Operations, Secret Network Foundation ◾ Jan Hans Georg Pachner , Secretary General, One Ocean Foundation ◾ Lauren Dutton , Devrel, Fhenix ◾ Richard Ells , CEO, Electroneum ◾ Lucía Gallardo , Founder, Emerge , Co-Founder, AeraTech ◾ Elina Litvak Cohen, Esq. , General Counsel, QuickNode ◾ Kyle K. , Co-Founder, Blockstreet™ Let's go! Web3, AI, Crypto, DeFi, RWAs, NFTs, Privacy, and more under one roof. November 5-6 at the Hard Rock Guitar and DAER. More announcements coming... see you all soon! 🎟️ https://lnkd.in/dhuz8gu 53 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה Three things just changed: 1/ FHE decryption got 20,000× faster 2/ Private DeFi is unlocked 3/ Fhenix made it happen Guy Zyskind 's paper detailing Fhenix's advancements in Threshold FHE Decryption has been featured ACM Special Interest Group on Security, Audit and Control (SIGSAC) 's 2025 conference. This acknowledgment places Fhenix’s research alongside the work of Microsoft, Google, Meta, Stanford, and MIT — validating that our progress isn’t just fast… it’s foundational. View the paper: https://lnkd.in/gbWuFihV Recap: https://lnkd.in/gtAzuPa7 High-Throughput Universally Composable Threshold FHE Decryption eprint.iacr.org 6 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏3‏חודש דיווח על פוסט הזה Join Guy Itzhaki amongst others tomorrow 9am ET / 13:00 UTC for a roundtable discussion on confidential AI! 🔗: https://lnkd.in/gVNgBFeR Optalysys ‏‏3,604‏ עוקבים‏ ‏3‏חודש This week: 𝗖𝗘𝗢𝘀 𝗗𝗲𝗰𝗼𝗱𝗲 𝗔𝗜 𝗖𝗼𝗻𝗳𝗶𝗱𝗲𝗻𝘁𝗶𝗮𝗹: 𝗭𝗞, 𝗧𝗘𝗘, 𝗙𝗛𝗘 | an X Space Roundtable Matt Garrow-Fisher will moderate this unmissable CEO roundtable with: Nick New – CEO, Optalysys Guy Itzhaki – CEO, Fhenix Robert Viglione – CEO, @HorizenLabs Marvin Tong – CEO, @PhalaNetwork 📌 Be there: https://lnkd.in/eBXugGXc 🗓️ Add to your calendar now: Wed, Sept 17 - 9AM ET | 2PM BST #xspace #roundtable #confidentialai 4 לייק תגובה שיתוף Fhenix ‏‏1,113‏ עוקבים‏ ‏4‏חודש דיווח על פוסט הזה Join our workshop on confidential payments this Friday, September 12 at the Crecimiento Argentina Aleph Hub in Buenos Aires, Argentina! 🎟️ Register: https://lnkd.in/etUg-MG3 Cámara Argentina Fintech ‏‏43,907‏ עוקבים‏ ‏4‏חודש נערך 🚀 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 𝗦𝘂𝗺𝗺𝗶𝘁 | 𝗦𝘂𝗺𝗮𝘁𝗲 𝗮 𝘂𝗻 𝗲𝘃𝗲𝗻𝘁𝗼 𝗽𝗮𝗿𝗮 𝗰𝗼𝗻𝗼𝗰𝗲𝗿 𝗺á𝘀 𝘀𝗼𝗯𝗿𝗲 𝗹𝗮 𝘂𝗻𝗶ó𝗻 𝗲𝗻𝘁𝗿𝗲 𝗹𝗮𝘀 𝗳𝗶𝗻𝗮𝗻𝘇𝗮𝘀 𝘁𝗿𝗮𝗱𝗶𝗰𝗶𝗼𝗻𝗮𝗹𝗲𝘀, 𝗯𝗹𝗼𝗰𝗸𝗰𝗵𝗮𝗶𝗻 𝘆 𝗪𝗲𝗯𝟯 Desde la 𝗖á𝗺𝗮𝗿𝗮 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 los invitamos a ser parte del 𝗙𝗶𝗻𝘁𝗲𝗰𝗵 𝗦𝘂𝗺𝗺𝗶𝘁, un encuentro diseñado para explorar la intersección entre finanzas tradicionales, fintechs y proyectos Web3. Junto a Crecimiento, reuniremos a bancos, fintechs y startups para debatir sobre el 𝗳𝘂𝘁𝘂𝗿𝗼 𝗱𝗲𝗹 𝗱𝗶𝗻𝗲𝗿𝗼 𝗲𝗻 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮: desde open finance y pagos transfronterizos hasta IA, infraestructura y seguridad financiera. 📅 𝗩𝗶𝗲𝗿𝗻𝗲𝘀 𝟭𝟮 𝗱𝗲 𝘀𝗲𝗽𝘁𝗶𝗲𝗺𝗯𝗿𝗲 - 𝟵 𝗮 𝟭𝟳:𝟯𝟬 𝗵𝘀 📍 𝗖𝗼𝗻𝗰𝗲𝗽𝗰𝗶𝗼́𝗻 𝗔𝗿𝗲𝗻𝗮𝗹 𝟮𝟵𝟴𝟵, 𝗖𝗔𝗕𝗔 🗓️ 𝗔𝗴𝗲𝗻𝗱𝗮 𝗰𝗼𝗻𝗳𝗶𝗿𝗺𝗮𝗱𝗮 🔹𝟵:𝟯𝟬 - 𝗥𝗲𝗴𝗶𝘀𝘁𝗿𝗼 𝘆 𝗱𝗲𝘀𝗮𝘆𝘂𝗻𝗼 🔹𝟭𝟬:𝟭𝟱 - 𝗞𝗲𝘆𝗻𝗼𝘁𝗲 𝗱𝗲 𝗮𝗽𝗲𝗿𝘁𝘂𝗿𝗮 ( Crecimiento Argentina ) 🔹 𝟭𝟬:𝟯𝟬 - 𝗣𝗮𝗻𝗲𝗹: 𝗢𝗽𝗲𝗻 𝗙𝗶𝗻𝗮𝗻𝗰𝗲 Moderador: Teófilo Beato (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Eduardo Segovia Mattos (Open Finance) y Nicolas Ballare Scopel, CIIA ( MODO ). 🔹𝟭𝟭:𝟮𝟬 - 𝗣𝗮𝗻𝗲𝗹: 𝗣𝗮𝗴𝗼𝘀 𝘁𝗿𝗮𝗻𝘀𝗳𝗿𝗼𝗻𝘁𝗲𝗿𝗶𝘇𝗼𝘀 Moderador: Diego Martin Diaz (CAF). Speakers: Federico Martin Testoni (Depay), Henri de Royere ( Due ) y Catalina Ramallo ( DolarApp ). 🔹𝟭𝟮:𝟭𝟬 - 𝗪𝗼𝗿𝗸𝘀𝗵𝗼𝗽: 𝗖𝗼𝗻𝗳𝗶𝗱𝗲𝗻𝘁𝗶𝗮𝗹 𝗣𝗮𝘆𝗺𝗲𝗻𝘁𝘀 Moderador: Simon Puebla (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Fhenix . 🔹𝟭𝟰:𝟭𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗜𝗻𝗳𝗿𝗮𝗲𝘀𝘁𝗿𝘂𝗰𝘁𝘂𝗿𝗮 Moderador: Teófilo Beato (𝗖𝗿𝗲𝗰𝗶𝗺𝗶𝗲𝗻𝘁𝗼 𝗔𝗿𝗴𝗲𝗻𝘁𝗶𝗻𝗮). Speakers: Edwin A. Rodriguez ( Rain ) y Santiago Witis  ( Pomelo ). 🔹𝟭𝟱:𝟬𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗜𝗔 𝗲𝗻 𝘀𝗲𝗿𝘃𝗶𝗰𝗶𝗼𝘀 𝗳𝗶𝗻𝗮𝗻𝗰𝗶𝗲𝗿𝗼𝘀 Moderadora: Natalia Diaz (CAF). Speakers: Lucila Destefanis ( Satoshi Tango ), Juan Ignacio Segura ( PwC Argentina ) y Paula Barnes, CAMS (𝗣𝗼𝗺𝗲𝗹𝗼). 🔹𝟭𝟱:𝟱𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗧𝗮𝗿𝗷𝗲𝘁𝗮𝘀 Moderador: Julian Morla ( LI.FI ). Speakers: Patricio Molina ( belo ), Tobias Frieder ( Offramp Labs ) y Juan Manuel Brunetti ( Prex ). 🔹𝟭𝟲:𝟰𝟱 - 𝗣𝗮𝗻𝗲𝗹: 𝗦𝗲𝗴𝘂𝗿𝗶𝗱𝗮𝗱 𝗰𝗼𝗺𝗼 𝘃𝗲𝗻𝘁𝗮𝗷𝗮 𝗰𝗼𝗺𝗽𝗲𝘁𝗶𝘁𝗶𝘃𝗮 Moderador: Matias Fermin (CAF). Speakers: Ivan Lendner ( Whalemate ), Marcelo Triviño ( Ualá ), Lucas Paus ( BIND | Soluciones financieras ) y Santiago Agustín Fernandez (Klar). 🔹𝟭𝟳:𝟯𝟬 - 𝗖𝗶𝗲𝗿𝗿𝗲 🔗 Registrate acá: https://lnkd.in/etUg-MG3 Será una jornada para aprender, compartir y conectar con quienes están construyendo el próximo capítulo del ecosistema financiero argentino. ¡Los esperamos! לייק תגובה שיתוף הצטרף עכשיו כדי לראות מה חסר לך מצא אנשים שאתה מכיר בהם Fhenix עיין מקומות תעסוקה מומלצים עבורך הצג את כל העדכונים, חֲדָשׁוֹת והמאמרים הצטרף עכשיו דפים דומים SCRT Labs Software Development ‏.‏, ‏.‏ Zama Software Development Xai Games Blockchain Services medflyt Software Development ‏Brooklyn‏, ‏New York‏ StarkWare Software Development ‏Netanya‏, ‏Hasharon‏ Optalysys Computer Hardware Manufacturing ‏Leeds‏, ‏England‏ Arbitrum Technology, Information and Internet ‏George Town‏, ‏Grand Cayman‏ Utila Software Development ‏New York City‏, ‏New York‏ Zengo Wallet Financial Services ‏Tel Aviv‏, ‏NA‏ Polygon Labs Technology, Information and Internet הצג עוד דפים דומים הצג פחות דפים דומים מימון Fhenix 4 total rounds סיבוב אחרון סדרה לא ידועה 22 בנוב׳ 2025 External Crunchbase Link לסבב הגיוס האחרון משקיעים Biprogy Translink Capital ראה מידע נוסף על crunchbase ‏LinkedIn © 2026 על אודות נגישות הסכם המשתמש מדיניות פרטיות מדיניות עוגיות מדיניות זכויות יוצרים מדיניות המותג בקרת אורחים הנחיות הקהילה العربية (ערבית) ‏বাংলা (בנגלה) ‏Čeština (צ’כית) ‏Dansk (דנית) ‏Deutsch (גרמנית) ‏Ελληνικά (יוונית) ‏English (אנגלית) ‏Español (ספרדית) فارسی (פרסית) ‏Suomi (פינית) ‏Français (צרפתית) ‏हिंदी (הינדי) ‏Magyar (הונגרית) ‏Bahasa Indonesia (אינדונזית) ‏Italiano (איטלקית) עברית (Hebrew) ‏日本語 (יפנית) ‏한국어 (קוריאנית) ‏मराठी (מראטהית) ‏Bahasa Malaysia (מלאית) ‏Nederlands (הולנדית) ‏Norsk (נורווגית) ‏ਪੰਜਾਬੀ (פנג’אבי) ‏Polski (פולנית) ‏Português (פורטוגזית) ‏Română (רומנית) ‏Русский (רוסית) ‏Svenska (שוודית) ‏తెలుగు (טלוגו) ‏ภาษาไทย (תאילנדית) ‏Tagalog (טאגאלוג) ‏Türkçe (טורקית) ‏Українська (אוקראינית) ‏Tiếng Việt (וייטנאמית) ‏简体中文 (סינית (מפושטת)) ‏正體中文 (סינית (מסורתית)) שפה מסכים ומצטרף LinkedIn הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn. היכנס כדי לראות את מי שאתה כבר מכיר בכתובת Fhenix התחברות טוב לראות אותך שוב דוא”ל או טלפון סיסמה הצגה שכחת סיסמא? התחברות או הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn. פעם ראשונה שלך ב-LinkedIn? ‏ להצטרפות ‏ או פעם ראשונה שלך ב-LinkedIn? ‏ להצטרפות ‏ הלחיצה על ’המשך‘ להצטרפות או להתחברות מהווה את הסמכתך ל‏ הסכם המשתמש ‏, ל‏ מדיניות הפרטיות ‏ ול‏ מדיניות קובצי ה-Cookie ‏ של LinkedIn.
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/reference/source-replacement.html#footnote-1
Source Replacement - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book Source Replacement This document is about redirecting communication with registries or repositories of git-based dependencies to another data source, such as a server mirroring the original registry or an exact local copy. If you want to patch individual dependencies, see overriding dependencies section of this documentation. If you want to control how Cargo makes network requests, see [http] and [net] configuration. A source is a provider that contains crates that may be included as dependencies for a package. Cargo supports the ability to replace one source with another to express strategies such as: Vendoring — custom sources can be defined which represent crates on the local filesystem. These sources are subsets of the source that they’re replacing and can be checked into packages if necessary. Mirroring — sources can be replaced with an equivalent version which acts as a cache for crates.io itself. Cargo has a core assumption about source replacement that the source code is exactly the same from both sources. Note that this also means that a replacement source is not allowed to have crates which are not present in the original source. As a consequence, source replacement is not appropriate for situations such as patching a dependency or a private registry. Cargo supports patching dependencies through the usage of the [patch] key , and private registry support is described in the Registries chapter . When using source replacement, running commands that need to contact the registry directly 1 requires passing the --registry option. This helps avoid any ambiguity about which registry to contact, and will use the authentication token for the specified registry. Configuration Configuration of replacement sources is done through .cargo/config.toml and the full set of available keys are: # The `source` table is where all keys related to source-replacement # are stored. [source] # Under the `source` table are a number of other tables whose keys are a # name for the relevant source. For example this section defines a new # source, called `my-vendor-source`, which comes from a directory # located at `vendor` relative to the directory containing this `.cargo/config.toml` # file [source.my-vendor-source] directory = "vendor" # The crates.io default source for crates is available under the name # "crates-io", and here we use the `replace-with` key to indicate that it's # replaced with our source above. # # The `replace-with` key can also reference an alternative registry name # defined in the `[registries]` table. [source.crates-io] replace-with = "my-vendor-source" # Each source has its own table where the key is the name of the source [source.the-source-name] # Indicate that `the-source-name` will be replaced with `another-source`, # defined elsewhere replace-with = "another-source" # Several kinds of sources can be specified (described in more detail below): registry = "https://example.com/path/to/index" local-registry = "path/to/registry" directory = "path/to/vendor" # Git sources can optionally specify a branch/tag/rev as well git = "https://example.com/path/to/repo" # branch = "master" # tag = "v1.0.1" # rev = "313f44e8" Registry Sources A “registry source” is one that works like crates.io itself. It’s an index that conforms to the specification at https://doc.rust-lang.org/cargo/reference/registry-index.html with a configuration file indicating where to download crates from. Registry sources can use either git or sparse HTTP protocol : # Git protocol registry = "ssh://git@example.com/path/to/index.git" # Sparse HTTP protocol registry = "sparse+https://example.com/path/to/index" # HTTPS git protocol registry = "https://example.com/path/to/index" Local Registry Sources A “local registry source” is intended to be a subset of another registry source, but available on the local filesystem (aka vendoring). Local registries are downloaded ahead of time, typically sync’d with a Cargo.lock , and are made up of a set of *.crate files and an index like the normal registry is. The primary way to manage and create local registry sources is through the cargo-local-registry subcommand, available on crates.io and can be installed with cargo install cargo-local-registry . Local registries are contained within one directory and contain a number of *.crate files downloaded from crates.io as well as an index directory with the same format as the crates.io-index project (populated with just entries for the crates that are present). Directory Sources A “directory source” is similar to a local registry source where it contains a number of crates available on the local filesystem, suitable for vendoring dependencies. Directory sources are primarily managed by the cargo vendor subcommand. Directory sources are distinct from local registries though in that they contain the unpacked version of *.crate files, making it more suitable in some situations to check everything into source control. A directory source is just a directory containing a number of other directories which contain the source code for crates (the unpacked version of *.crate files). Currently no restriction is placed on the name of each directory. Each crate in a directory source also has an associated metadata file indicating the checksum of each file in the crate to protect against accidental modifications. Git sources Git sources represent repositories used by git-based dependencies . They’re used to specify which git-based dependencies should be replaced with alternative sources. Git sources are not related to the git registries , and can’t be used to replace registry sources. Examples of such commands are in Publishing Commands . ↩
2026-01-13T09:29:19
https://www.linkedin.com/products/bmc-software-mainframe/
BMC AMI (Mainframe Transformation) | LinkedIn Skip to main content LinkedIn BMC Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in BMC AMI (Mainframe Transformation) Mainframe Management Software by BMC Software See who's skilled in this Add as skill Learn more Report this product About BMC Automated Mainframe Intelligence (AMI) solutions leverage AI, machine learning, and predictive analytics to achieve a self-managing mainframe. Take advantage of: - AMI Cloud - AMI Data for DB2 - AMI Data for IMS - AMI DevX - AMI Ops - AMI Security Featured customers of BMC AMI (Mainframe Transformation) CUA Financial Services 4,780 followers Sichuan Rural Credit Union & Cooperative Bank Accounting 127 followers Iberdrola Electric Power Generation 582,243 followers PT Bank Central Asia Tbk (BCA) Banking 709,265 followers ATPCO Airlines and Aviation 27,831 followers Dillard's Retail 77,946 followers HUK-COBURG Insurance 22,076 followers CPT Global Information Technology & Services 9,833 followers Show more Show less Similar products HCL Mainframe Solutions HCL Mainframe Solutions Mainframe Management Software Vertali ZTRUST for SMP/e Vertali ZTRUST for SMP/e Mainframe Management Software Vertali ZTRUST Password Synchronisation Vertali ZTRUST Password Synchronisation Mainframe Management Software Tectia® SSH Server for IBM z/OS Tectia® SSH Server for IBM z/OS Mainframe Management Software Lozen™ Connector for MuleSoft Lozen™ Connector for MuleSoft Mainframe Management Software iCAN iCAN Mainframe Management Software Sign in to see more Show more Show less BMC Software products Control-M Control-M Workload Automation Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://huggingface.co/datasets/LEMAS-Project/LEMAS-Dataset-train
LEMAS-Project/LEMAS-Dataset-train · Datasets at Hugging Face Hugging Face Models Datasets Spaces Community Docs Enterprise Pricing Log In Sign Up 1T"]},"cardExists":true,"createdAt":"2025-12-04T08:53:38.000Z","description":"\n\t\n\t\t\n\t\tOverview\n\t\n\nThis dataset is part of LEMAS-Project (lemas-project.github.io/LEMAS-Project).\nIt contains a large-scale training set (150k+ hours) and a curated evaluation set\n(500 utterances per language) covering 10 languages, all with word-level alignment.\n\n\t\n\t\t\n\t\tFields\n\t\n\n\nkey: unique utterance identifier; the first two characters indicate the language ID\naudio: relative path to the MP3 audio file (in the eval set, this key is renamed to \"file_name\" for compatibility with the viewer)… See the full description on the dataset page: https://huggingface.co/datasets/LEMAS-Project/LEMAS-Dataset-train.","downloads":3619,"downloadsAllTime":4067,"id":"LEMAS-Project/LEMAS-Dataset-train","isLikedByUser":false,"lastModified":"2026-01-09T03:33:49.000Z","likes":15,"datasetsServerInfo":{"viewer":"viewer-partial","numRows":125062262,"libraries":["datasets","dask","mlcroissant"],"formats":["json"],"modalities":["text"]},"discussionsDisabled":false,"discussionsSorting":"recently-created","repoType":"dataset","private":false,"gated":false,"tags":["task_categories:text-to-speech","task_categories:automatic-speech-recognition","language:it","language:pt","language:es","language:fr","language:de","language:vi","language:id","language:ru","language:en","language:zh","license:cc-by-nc-4.0","size_categories:100M Datasets: LEMAS-Project / LEMAS-Dataset-train like 15 Follow LEMAS 11 Tasks: Text-to-Speech Automatic Speech Recognition Modalities: Text Formats: json Languages: Italian Portuguese Spanish + 7 Size: 100M - 1B ArXiv: arxiv: 2601.04233 Libraries: Datasets Dask Croissant License: cc-by-nc-4.0 Dataset card Data Studio Files Files and versions xet Community 2 Dataset Viewer (First 5GB) Auto-converted to Parquet API Embed Duplicate Data Studio Subset (1) default · ~125M rows (showing the first 11.1M) default (~125M rows, showing the first 11.1M) Split (1) train · ~125M rows (showing the first 11.1M) train (~125M rows, showing the first 11.1M) SQL Console key string lengths 15 38 audio string lengths 22 45 dur float64 0.5 29.9 txt string lengths 4 561 align dict de__2zsNO2V9K4-00556-00150066-00150154 de000/_2zsNO2V9K4-00556-00150066-00150154.mp3 0.88 was da machen { "txt": "was da machen", "words": [ { "word": "was", "start": 0.082, "end": 0.246, "score": 0.663 }, { "word": "da", "start": 0.348, "end": 0.43, "score": 0.997 }, { "word": "machen", "start": 0.512, "end": 0.757, "score": 0.63 } ] } de_BMMAuLZglpy-00055-00014786-00015022 de000/BMMAuLZglpy-00055-00014786-00015022.mp3 2.37 motoraufkleber Fehler wurde behoben beim { "txt": "motoraufkleber Fehler wurde behoben beim", "words": [ { "word": "motoraufkleber", "start": 0, "end": 0.924, "score": 0.798 }, { "word": "Fehler", "start": 0.984, "end": 1.265, "score": 0.83 }, { "word": "wurde", "start": 1.306, "end": 1.547, "score": 0.993 }, { "word": "behoben", "start": 1.587, "end": 2.029, "score": 0.935 }, { "word": "beim", "start": 2.229, "end": 2.35, "score": 0.222 } ] } de__MJfoKau558-00113-00033631-00033868 de000/_MJfoKau558-00113-00033631-00033868.mp3 2.37 dann muss ich natürlich erst mal wieder { "txt": "dann muss ich natürlich erst mal wieder", "words": [ { "word": "dann", "start": 0.884, "end": 1.245, "score": 0.544 }, { "word": "muss", "start": 1.426, "end": 1.607, "score": 0.351 }, { "word": "ich", "start": 1.647, "end": 1.727, "score": 0.3 }, { "word": "natürlich", "start": 1.747, "end": 2.029, "score": 0.213 }, { "word": "erst", "start": 2.029, "end": 2.149, "score": 0.181 }, { "word": "mal", "start": 2.149, "end": 2.229, "score": 0.48 }, { "word": "wieder", "start": 2.249, "end": 2.37, "score": 0.123 } ] } de_oaRWyFJvKGg-00074-00017857-00018055 de000/oaRWyFJvKGg-00074-00017857-00018055.mp3 1.97 den einstieg haben wollen und da geht es { "txt": "den einstieg haben wollen und da geht es", "words": [ { "word": "den", "start": 0.02, "end": 0.121, "score": 0.942 }, { "word": "einstieg", "start": 0.181, "end": 0.523, "score": 0.972 }, { "word": "haben", "start": 0.583, "end": 0.804, "score": 0.96 }, { "word": "wollen", "start": 0.824, "end": 1.065, "score": 0.97 }, { "word": "und", "start": 1.447, "end": 1.568, "score": 0.623 }, { "word": "da", "start": 1.608, "end": 1.689, "score": 0.446 }, { "word": "geht", "start": 1.769, "end": 1.93, "score": 0.406 }, { "word": "es", "start": 1.93, "end": 1.97, "score": 0.073 } ] } de_6-4wCL-HjcY-00291-00121174-00121507 de000/6-4wCL-HjcY-00291-00121174-00121507.mp3 3.32 brauchen an den grünling bedient { "txt": "brauchen an den grünling bedient", "words": [ { "word": "brauchen", "start": 0.02, "end": 0.684, "score": 0.36 }, { "word": "an", "start": 1.328, "end": 1.429, "score": 0.496 }, { "word": "den", "start": 1.529, "end": 1.771, "score": 0.423 }, { "word": "grünling", "start": 1.831, "end": 2.314, "score": 0.838 }, { "word": "bedient", "start": 2.314, "end": 3.3, "score": 0.451 } ] } de_c2jcnRrEF5M-00303-00082363-00082481 de000/c2jcnRrEF5M-00303-00082363-00082481.mp3 1.18 eine regionalisierung der { "txt": "eine regionalisierung der", "words": [ { "word": "eine", "start": 0.102, "end": 0.224, "score": 0.206 }, { "word": "regionalisierung", "start": 0.244, "end": 1.119, "score": 0.487 }, { "word": "der", "start": 1.119, "end": 1.18, "score": 0.225 } ] } de_2R7gPvDwCTU-00360-00126595-00126981 de000/2R7gPvDwCTU-00360-00126595-00126981.mp3 3.86 einem kind ein vater zu seiner tochter { "txt": "einem kind ein vater zu seiner tochter", "words": [ { "word": "einem", "start": 0.201, "end": 0.382, "score": 0.92 }, { "word": "kind", "start": 0.462, "end": 0.784, "score": 0.967 }, { "word": "ein", "start": 1.97, "end": 2.091, "score": 0.967 }, { "word": "vater", "start": 2.151, "end": 2.553, "score": 0.98 }, { "word": "zu", "start": 3.036, "end": 3.136, "score": 0.956 }, { "word": "seiner", "start": 3.217, "end": 3.498, "score": 0.801 }, { "word": "tochter", "start": 3.659, "end": 3.86, "score": 0.281 } ] } de_Ykw3IKZK0qy-00320-00093241-00093486 de000/Ykw3IKZK0qy-00320-00093241-00093486.mp3 2.45 andere schriftart oder schriftgröße { "txt": "andere schriftart oder schriftgröße", "words": [ { "word": "andere", "start": 0.02, "end": 0.221, "score": 0.773 }, { "word": "schriftart", "start": 0.281, "end": 1.044, "score": 0.971 }, { "word": "oder", "start": 1.205, "end": 1.446, "score": 0.99 }, { "word": "schriftgröße", "start": 1.506, "end": 2.43, "score": 0.9 } ] } de_eKu59aCyt84-00005-00001055-00001481 de000/eKu59aCyt84-00005-00001055-00001481.mp3 4.26 die musik west ab über reist camelia { "txt": "die musik west ab über reist camelia", "words": [ { "word": "die", "start": 0.944, "end": 1.065, "score": 0.295 }, { "word": "musik", "start": 1.065, "end": 1.306, "score": 0.437 }, { "word": "west", "start": 1.346, "end": 1.608, "score": 0.53 }, { "word": "ab", "start": 1.708, "end": 1.808, "score": 0.664 }, { "word": "über", "start": 3.496, "end": 3.738, "score": 0.107 }, { "word": "reist", "start": 3.758, "end": 3.979, "score": 0.168 }, { "word": "camelia", "start": 4.039, "end": 4.26, "score": 0.118 } ] } de__4ClllQ5VLo-00173-00050712-00050995 de000/_4ClllQ5VLo-00173-00050712-00050995.mp3 2.83 Folge auch kombinieren also hat { "txt": "Folge auch kombinieren also hat", "words": [ { "word": "Folge", "start": 0.02, "end": 0.241, "score": 0.55 }, { "word": "auch", "start": 0.321, "end": 0.442, "score": 0.819 }, { "word": "kombinieren", "start": 0.482, "end": 1.064, "score": 0.878 }, { "word": "also", "start": 1.586, "end": 1.806, "score": 0.859 }, { "word": "hat", "start": 2.449, "end": 2.569, "score": 0.571 } ] } de_2422_2622_007126 de000/2422_2622_007126.mp3 12.37 freilich aber auch an feilheit und leere unter den dichtern dieser epoche keinem weicht { "txt": "freilich aber auch an feilheit und leere unter den dichtern dieser epoche keinem weicht", "words": [ { "word": "freilich", "start": 0.48, "end": 1.241, "score": 1 }, { "word": "aber", "start": 1.541, "end": 1.821, "score": 1 }, { "word": "auch", "start": 1.982, "end": 2.302, "score": 0.972 }, { "word": "an", "start": 2.422, "end": 2.562, "score": 0.999 }, { "word": "feilheit", "start": 2.982, "end": 4.063, "score": 0.984 }, { "word": "und", "start": 4.183, "end": 4.404, "score": 0.986 }, { "word": "leere", "start": 4.644, "end": 5.324, "score": 0.9 }, { "word": "unter", "start": 6.165, "end": 6.465, "score": 1 }, { "word": "den", "start": 6.545, "end": 6.725, "score": 0.999 }, { "word": "dichtern", "start": 6.886, "end": 7.726, "score": 0.991 }, { "word": "dieser", "start": 8.287, "end": 8.887, "score": 0.992 }, { "word": "epoche", "start": 9.047, "end": 9.728, "score": 0.963 }, { "word": "keinem", "start": 10.508, "end": 11.009, "score": 0.999 }, { "word": "weicht", "start": 11.409, "end": 12.19, "score": 0.989 } ] } de_vSwjN0IkWso-00054-00017773-00018066 de000/vSwjN0IkWso-00054-00017773-00018066.mp3 2.93 die saugen dann dass der spiritus { "txt": "die saugen dann dass der spiritus", "words": [ { "word": "die", "start": 0.843, "end": 0.943, "score": 0.957 }, { "word": "saugen", "start": 0.983, "end": 1.284, "score": 0.912 }, { "word": "dann", "start": 1.325, "end": 1.465, "score": 0.899 }, { "word": "dass", "start": 1.525, "end": 1.746, "score": 0.701 }, { "word": "der", "start": 2.488, "end": 2.589, "score": 0.599 }, { "word": "spiritus", "start": 2.629, "end": 2.93, "score": 0.668 } ] } de_ok5q438MhAE-00097-00065298-00066003 de000/ok5q438MhAE-00097-00065298-00066003.mp3 7.05 world compiler for optimizing execution { "txt": "world compiler for optimizing execution", "words": [ { "word": "world", "start": 0.821, "end": 2.684, "score": 0.604 }, { "word": "compiler", "start": 3.205, "end": 4.146, "score": 0.876 }, { "word": "for", "start": 5.608, "end": 5.728, "score": 0.752 }, { "word": "optimizing", "start": 5.808, "end": 6.429, "score": 0.923 }, { "word": "execution", "start": 6.489, "end": 7.05, "score": 0.86 } ] } de_VslTybTXsWY-00160-00034566-00034814 de000/VslTybTXsWY-00160-00034566-00034814.mp3 2.48 ungefähr zwei millionen aus die schwedt { "txt": "ungefähr zwei millionen aus die schwedt", "words": [ { "word": "ungefähr", "start": 0.464, "end": 0.807, "score": 0.883 }, { "word": "zwei", "start": 1.028, "end": 1.21, "score": 0.861 }, { "word": "millionen", "start": 1.27, "end": 1.553, "score": 0.743 }, { "word": "aus", "start": 1.613, "end": 1.774, "score": 0.985 }, { "word": "die", "start": 2.137, "end": 2.238, "score": 0.787 }, { "word": "schwedt", "start": 2.258, "end": 2.48, "score": 0.448 } ] } de_KFtEkiX6pko-00794-00239459-00239806 de000/KFtEkiX6pko-00794-00239459-00239806.mp3 3.47 video chats in beeren et also ihr habt { "txt": "video chats in beeren et also ihr habt", "words": [ { "word": "video", "start": 0.481, "end": 0.822, "score": 0.99 }, { "word": "chats", "start": 1.003, "end": 1.404, "score": 0.598 }, { "word": "in", "start": 1.946, "end": 2.046, "score": 0.842 }, { "word": "beeren", "start": 2.186, "end": 2.688, "score": 0.713 }, { "word": "et", "start": 2.728, "end": 2.908, "score": 0.59 }, { "word": "also", "start": 3.009, "end": 3.169, "score": 0.551 }, { "word": "ihr", "start": 3.209, "end": 3.289, "score": 0.247 }, { "word": "habt", "start": 3.289, "end": 3.47, "score": 0.252 } ] } de_9eKJtZ2bE0E-00977-00236940-00237266 de000/9eKJtZ2bE0E-00977-00236940-00237266.mp3 3.26 so eine art speed interview { "txt": "so eine art speed interview", "words": [ { "word": "so", "start": 0.302, "end": 0.382, "score": 0.714 }, { "word": "eine", "start": 0.423, "end": 0.543, "score": 0.449 }, { "word": "art", "start": 0.664, "end": 0.845, "score": 0.977 }, { "word": "speed", "start": 1.449, "end": 1.952, "score": 0.88 }, { "word": "interview", "start": 2.697, "end": 3.18, "score": 0.756 } ] } de_4J0R1oAf6_U-00112-00037820-00038126 de000/4J0R1oAf6_U-00112-00037820-00038126.mp3 3.06 das uns alle in größte Verantwortung ruft. { "txt": "das uns alle in größte Verantwortung ruft.", "words": [ { "word": "das", "start": 0.081, "end": 0.201, "score": 0.99 }, { "word": "uns", "start": 0.242, "end": 0.362, "score": 0.999 }, { "word": "alle", "start": 0.523, "end": 0.725, "score": 0.998 }, { "word": "in", "start": 0.846, "end": 0.926, "score": 0.997 }, { "word": "größte", "start": 0.986, "end": 1.309, "score": 0.997 }, { "word": "Verantwortung", "start": 1.349, "end": 1.973, "score": 0.979 }, { "word": "ruft.", "start": 2.033, "end": 2.376, "score": 0.751 } ] } de_9q1MlF1jZsE-00815-00212646-00212998 de000/9q1MlF1jZsE-00815-00212646-00212998.mp3 3.53 ist durch seinen Arbeitgeber analog { "txt": "ist durch seinen Arbeitgeber analog", "words": [ { "word": "ist", "start": 0.542, "end": 0.702, "score": 0.988 }, { "word": "durch", "start": 0.802, "end": 1.003, "score": 0.293 }, { "word": "seinen", "start": 1.063, "end": 1.384, "score": 0.58 }, { "word": "Arbeitgeber", "start": 1.464, "end": 2.367, "score": 0.707 }, { "word": "analog", "start": 2.948, "end": 3.41, "score": 0.765 } ] } de_9hcrNbBwzCA-00252-00060751-00061012 de000/9hcrNbBwzCA-00252-00060751-00061012.mp3 2.6 drei vier fünf oder zehn jahren das { "txt": "drei vier fünf oder zehn jahren das", "words": [ { "word": "drei", "start": 0.081, "end": 0.363, "score": 0.804 }, { "word": "vier", "start": 0.504, "end": 0.726, "score": 0.696 }, { "word": "fünf", "start": 0.826, "end": 1.109, "score": 0.8 }, { "word": "oder", "start": 1.189, "end": 1.35, "score": 0.953 }, { "word": "zehn", "start": 1.431, "end": 1.693, "score": 0.767 }, { "word": "jahren", "start": 1.774, "end": 2.177, "score": 0.883 }, { "word": "das", "start": 2.499, "end": 2.58, "score": 0.069 } ] } de_gRh8suav7EM-00064-00016393-00016596 de000/gRh8suav7EM-00064-00016393-00016596.mp3 2.03 gekriegt wie firmen arbeiten wie sie { "txt": "gekriegt wie firmen arbeiten wie sie", "words": [ { "word": "gekriegt", "start": 0.1, "end": 0.482, "score": 0.599 }, { "word": "wie", "start": 0.804, "end": 0.945, "score": 0.9 }, { "word": "firmen", "start": 1.025, "end": 1.327, "score": 0.773 }, { "word": "arbeiten", "start": 1.427, "end": 1.789, "score": 0.959 }, { "word": "wie", "start": 1.829, "end": 1.93, "score": 0.599 }, { "word": "sie", "start": 1.95, "end": 2.03, "score": 0.361 } ] } de_FUeCSQECuC0-00067-00032311-00032670 de000/FUeCSQECuC0-00067-00032311-00032670.mp3 3.6 sichern wir auch das erzeuge ein versprechen hält dass die gehälter der { "txt": "sichern wir auch das erzeuge ein versprechen hält dass die gehälter der", "words": [ { "word": "sichern", "start": 0.422, "end": 0.704, "score": 0.969 }, { "word": "wir", "start": 0.724, "end": 0.804, "score": 0.981 }, { "word": "auch", "start": 0.885, "end": 1.046, "score": 0.979 }, { "word": "das", "start": 1.086, "end": 1.207, "score": 0.984 }, { "word": "erzeuge", "start": 1.267, "end": 1.649, "score": 0.679 }, { "word": "ein", "start": 1.709, "end": 1.81, "score": 0.96 }, { "word": "versprechen", "start": 1.85, "end": 2.373, "score": 0.977 }, { "word": "hält", "start": 2.413, "end": 2.695, "score": 0.782 }, { "word": "dass", "start": 2.936, "end": 3.057, "score": 0.508 }, { "word": "die", "start": 3.097, "end": 3.198, "score": 0.761 }, { "word": "gehälter", "start": 3.218, "end": 3.54, "score": 0.472 }, { "word": "der", "start": 3.54, "end": 3.6, "score": 0.078 } ] } de_NKEKnYCiVTc-00231-00111027-00111308 de000/NKEKnYCiVTc-00231-00111027-00111308.mp3 2.82 stelle dir dafür zunächst eine kleine { "txt": "stelle dir dafür zunächst eine kleine", "words": [ { "word": "stelle", "start": 0.745, "end": 1.047, "score": 0.999 }, { "word": "dir", "start": 1.108, "end": 1.269, "score": 0.966 }, { "word": "dafür", "start": 1.329, "end": 1.632, "score": 0.997 }, { "word": "zunächst", "start": 1.712, "end": 2.175, "score": 0.981 }, { "word": "eine", "start": 2.276, "end": 2.437, "score": 0.999 }, { "word": "kleine", "start": 2.478, "end": 2.82, "score": 0.915 } ] } de_klWvdrWIqYM-00818-00216704-00216980 de000/klWvdrWIqYM-00818-00216704-00216980.mp3 2.76 da können sie hier nachlesen { "txt": "da können sie hier nachlesen", "words": [ { "word": "da", "start": 0.947, "end": 1.007, "score": 0.984 }, { "word": "können", "start": 1.088, "end": 1.35, "score": 0.922 }, { "word": "sie", "start": 1.37, "end": 1.471, "score": 0.996 }, { "word": "hier", "start": 1.491, "end": 1.874, "score": 0.139 }, { "word": "nachlesen", "start": 1.994, "end": 2.599, "score": 0.981 } ] } de_hMfzens7xf8-00068-00020422-00020558 de000/hMfzens7xf8-00068-00020422-00020558.mp3 1.37 berechnet und hier in der tabelle { "txt": "berechnet und hier in der tabelle", "words": [ { "word": "berechnet", "start": 0.101, "end": 0.524, "score": 0.705 }, { "word": "und", "start": 0.625, "end": 0.745, "score": 0.724 }, { "word": "hier", "start": 0.786, "end": 0.907, "score": 0.295 }, { "word": "in", "start": 0.927, "end": 0.987, "score": 0.379 }, { "word": "der", "start": 1.007, "end": 1.108, "score": 0.26 }, { "word": "tabelle", "start": 1.108, "end": 1.37, "score": 0.511 } ] } de_EhWv48hTOhQ-00096-00063373-00063942 de000/EhWv48hTOhQ-00096-00063373-00063942.mp3 5.7 gemeinsamen lebensunterhalt zu verdienen zu hause in fremdenzimmern becken städten die immer wieder { "txt": "gemeinsamen lebensunterhalt zu verdienen zu hause in fremdenzimmern becken städten die immer wieder", "words": [ { "word": "gemeinsamen", "start": 0.06, "end": 0.562, "score": 0.889 }, { "word": "lebensunterhalt", "start": 0.602, "end": 1.305, "score": 0.99 }, { "word": "zu", "start": 1.345, "end": 1.405, "score": 0.959 }, { "word": "verdienen", "start": 1.465, "end": 1.846, "score": 0.895 }, { "word": "zu", "start": 2.148, "end": 2.228, "score": 0.995 }, { "word": "hause", "start": 2.308, "end": 2.649, "score": 0.999 }, { "word": "in", "start": 2.71, "end": 2.77, "score": 0.996 }, { "word": "fremdenzimmern", "start": 2.83, "end": 3.613, "score": 0.994 }, { "word": "becken", "start": 3.733, "end": 4.054, "score": 0.648 }, { "word": "städten", "start": 4.195, "end": 4.616, "score": 0.852 }, { "word": "die", "start": 5.018, "end": 5.138, "score": 0.912 }, { "word": "immer", "start": 5.238, "end": 5.459, "score": 0.958 }, { "word": "wieder", "start": 5.539, "end": 5.7, "score": 0.485 } ] } de_pHfKGy9a5HM-00418-00122753-00123096 de000/pHfKGy9a5HM-00418-00122753-00123096.mp3 3.43 dass es den charakter der natürlichkeit { "txt": "dass es den charakter der natürlichkeit", "words": [ { "word": "dass", "start": 0.602, "end": 0.802, "score": 0.747 }, { "word": "es", "start": 0.802, "end": 0.883, "score": 0.5 }, { "word": "den", "start": 0.943, "end": 1.083, "score": 0.87 }, { "word": "charakter", "start": 1.123, "end": 1.785, "score": 0.993 }, { "word": "der", "start": 1.885, "end": 2.026, "score": 0.999 }, { "word": "natürlichkeit", "start": 2.066, "end": 3.089, "score": 0.997 } ] } de_oMefmzBD_rQ-01090-00242850-00243094 de000/oMefmzBD_rQ-01090-00242850-00243094.mp3 2.45 Thema generative KI wir haben natürlich { "txt": "Thema generative KI wir haben natürlich", "words": [ { "word": "Thema", "start": 0.02, "end": 0.181, "score": 0.65 }, { "word": "generative", "start": 0.241, "end": 0.823, "score": 0.942 }, { "word": "KI", "start": 1.165, "end": 1.466, "score": 0.71 }, { "word": "wir", "start": 1.968, "end": 2.089, "score": 0.662 }, { "word": "haben", "start": 2.129, "end": 2.229, "score": 0.378 }, { "word": "natürlich", "start": 2.249, "end": 2.45, "score": 0.087 } ] } de_5DowOTWL7Z0-01265-00377739-00378107 de000/5DowOTWL7Z0-01265-00377739-00378107.mp3 3.68 andere der aufklärung eine rolle spielen { "txt": "andere der aufklärung eine rolle spielen", "words": [ { "word": "andere", "start": 0.362, "end": 0.704, "score": 0.994 }, { "word": "der", "start": 0.804, "end": 0.965, "score": 0.929 }, { "word": "aufklärung", "start": 1.046, "end": 1.77, "score": 0.981 }, { "word": "eine", "start": 2.815, "end": 2.976, "score": 0.973 }, { "word": "rolle", "start": 3.016, "end": 3.217, "score": 0.999 }, { "word": "spielen", "start": 3.298, "end": 3.66, "score": 0.982 } ] } de_IqL1xdb-VnY-00117-00025601-00025852 de000/IqL1xdb-VnY-00117-00025601-00025852.mp3 2.51 gehen und sie noch ein letztes Mal sich { "txt": "gehen und sie noch ein letztes Mal sich", "words": [ { "word": "gehen", "start": 0.12, "end": 0.402, "score": 0.636 }, { "word": "und", "start": 0.643, "end": 0.763, "score": 0.921 }, { "word": "sie", "start": 0.823, "end": 0.924, "score": 0.817 }, { "word": "noch", "start": 0.964, "end": 1.104, "score": 0.922 }, { "word": "ein", "start": 1.185, "end": 1.305, "score": 0.935 }, { "word": "letztes", "start": 1.345, "end": 1.687, "score": 0.938 }, { "word": "Mal", "start": 1.767, "end": 2.068, "score": 0.773 }, { "word": "sich", "start": 2.309, "end": 2.47, "score": 0.207 } ] } de_PfjVIfOb62y-00803-00260253-00260792 de000/PfjVIfOb62y-00803-00260253-00260792.mp3 5.4 Der verbleibende Zuwachs zur Stärkung der Bildungssprache Deutsch { "txt": "Der verbleibende Zuwachs zur Stärkung der Bildungssprache Deutsch", "words": [ { "word": "Der", "start": 1.586, "end": 1.686, "score": 0.993 }, { "word": "verbleibende", "start": 1.706, "end": 2.188, "score": 0.995 }, { "word": "Zuwachs", "start": 2.248, "end": 2.71, "score": 0.992 }, { "word": "zur", "start": 3.272, "end": 3.413, "score": 0.997 }, { "word": "Stärkung", "start": 3.493, "end": 3.975, "score": 0.973 }, { "word": "der", "start": 4.015, "end": 4.095, "score": 0.997 }, { "word": "Bildungssprache", "start": 4.115, "end": 4.798, "score": 0.982 }, { "word": "Deutsch", "start": 4.858, "end": 5.28, "score": 0.985 } ] } de_IrNNyZBg4XU-02735-00635460-00635831 de000/IrNNyZBg4XU-02735-00635460-00635831.mp3 3.71 machen aber also vielleicht als { "txt": "machen aber also vielleicht als", "words": [ { "word": "machen", "start": 0.16, "end": 0.381, "score": 0.727 }, { "word": "aber", "start": 2.206, "end": 2.366, "score": 0.831 }, { "word": "also", "start": 2.567, "end": 2.687, "score": 0.522 }, { "word": "vielleicht", "start": 2.928, "end": 3.229, "score": 0.584 }, { "word": "als", "start": 3.59, "end": 3.69, "score": 0.237 } ] } de_F3bWpzyjxC8-00094-00054120-00054335 de000/F3bWpzyjxC8-00094-00054120-00054335.mp3 2.15 eine Art Trasse für den reinen { "txt": "eine Art Trasse für den reinen", "words": [ { "word": "eine", "start": 0.121, "end": 0.301, "score": 0.954 }, { "word": "Art", "start": 0.402, "end": 0.583, "score": 0.99 }, { "word": "Trasse", "start": 0.723, "end": 1.125, "score": 0.94 }, { "word": "für", "start": 1.266, "end": 1.427, "score": 0.963 }, { "word": "den", "start": 1.447, "end": 1.567, "score": 0.971 }, { "word": "reinen", "start": 1.668, "end": 2.05, "score": 0.931 } ] } de_3319Z5pYp_y-00023-00005355-00005585 de000/3319Z5pYp_y-00023-00005355-00005585.mp3 2.3 mit diesem gesetz schützen wir sensible { "txt": "mit diesem gesetz schützen wir sensible", "words": [ { "word": "mit", "start": 0.504, "end": 0.585, "score": 0.989 }, { "word": "diesem", "start": 0.625, "end": 0.827, "score": 0.998 }, { "word": "gesetz", "start": 0.888, "end": 1.291, "score": 0.991 }, { "word": "schützen", "start": 1.432, "end": 1.775, "score": 0.942 }, { "word": "wir", "start": 1.775, "end": 1.896, "score": 0.96 }, { "word": "sensible", "start": 1.937, "end": 2.3, "score": 0.543 } ] } de_UR-r5BeG8KY-00421-00109281-00109517 de000/UR-r5BeG8KY-00421-00109281-00109517.mp3 2.36 von mir auflaufen anhand seiner { "txt": "von mir auflaufen anhand seiner", "words": [ { "word": "von", "start": 0.141, "end": 0.262, "score": 0.406 }, { "word": "mir", "start": 0.303, "end": 0.403, "score": 0.485 }, { "word": "auflaufen", "start": 0.444, "end": 1.311, "score": 0.588 }, { "word": "anhand", "start": 1.916, "end": 2.239, "score": 0.693 }, { "word": "seiner", "start": 2.239, "end": 2.36, "score": 0.084 } ] } de_2422_2246_006728 de000/2422_2246_006728.mp3 12.7 doch wurde auch hier bestimmt dass den staedten ihre freibriefe bleiben und die abgabe nicht erhoeht werden solle ferner musste { "txt": "doch wurde auch hier bestimmt dass den staedten ihre freibriefe bleiben und die abgabe nicht erhoeht werden solle ferner musste", "words": [ { "word": "doch", "start": 0.481, "end": 0.721, "score": 0.991 }, { "word": "wurde", "start": 0.801, "end": 1.102, "score": 1 }, { "word": "auch", "start": 1.202, "end": 1.362, "score": 0.998 }, { "word": "hier", "start": 1.422, "end": 1.643, "score": 0.992 }, { "word": "bestimmt", "start": 1.723, "end": 2.524, "score": 0.993 }, { "word": "dass", "start": 2.744, "end": 2.945, "score": 0.984 }, { "word": "den", "start": 3.005, "end": 3.165, "score": 0.999 }, { "word": "staedten", "start": 3.265, "end": 3.966, "score": 0.893 }, { "word": "ihre", "start": 4.487, "end": 4.667, "score": 0.97 }, { "word": "freibriefe", "start": 4.788, "end": 5.729, "score": 0.994 }, { "word": "bleiben", "start": 5.969, "end": 6.61, "score": 0.999 }, { "word": "und", "start": 7.151, "end": 7.312, "score": 0.947 }, { "word": "die", "start": 7.472, "end": 7.592, "score": 0.97 }, { "word": "abgabe", "start": 7.712, "end": 8.353, "score": 1 }, { "word": "nicht", "start": 8.533, "end": 8.854, "score": 0.993 }, { "word": "erhoeht", "start": 9.034, "end": 9.515, "score": 0.875 }, { "word": "werden", "start": 9.635, "end": 9.976, "score": 0.999 }, { "word": "solle", "start": 10.076, "end": 10.396, "score": 1 }, { "word": "ferner", "start": 11.218, "end": 11.718, "score": 0.133 }, { "word": "musste", "start": 11.859, "end": 12.139, "score": 0.158 } ] } de_-SYT7UXDrjE-00191-00071182-00071395 de000/-SYT7UXDrjE-00191-00071182-00071395.mp3 2.13 ich finde für die projektplanung { "txt": "ich finde für die projektplanung", "words": [ { "word": "ich", "start": 0.904, "end": 1.005, "score": 0.466 }, { "word": "finde", "start": 1.025, "end": 1.226, "score": 0.309 }, { "word": "für", "start": 1.266, "end": 1.387, "score": 0.263 }, { "word": "die", "start": 1.387, "end": 1.507, "score": 0.529 }, { "word": "projektplanung", "start": 1.547, "end": 2.13, "score": 0.224 } ] } de_9tvvoBt3NKM-00378-00084647-00084853 de000/9tvvoBt3NKM-00378-00084647-00084853.mp3 2.06 dem begriff dream reactions traum { "txt": "dem begriff dream reactions traum", "words": [ { "word": "dem", "start": 0.04, "end": 0.141, "score": 0.271 }, { "word": "begriff", "start": 0.141, "end": 0.444, "score": 0.658 }, { "word": "dream", "start": 0.99, "end": 1.272, "score": 0.75 }, { "word": "reactions", "start": 1.353, "end": 1.919, "score": 0.612 }, { "word": "traum", "start": 1.959, "end": 2.06, "score": 0.069 } ] } de_Z6_mSFxSYvc-00482-00175426-00175748 de000/Z6_mSFxSYvc-00482-00175426-00175748.mp3 3.22 bewusstsein wie ein planet wenn die { "txt": "bewusstsein wie ein planet wenn die", "words": [ { "word": "bewusstsein", "start": 0.06, "end": 0.644, "score": 0.308 }, { "word": "wie", "start": 0.684, "end": 0.785, "score": 0.384 }, { "word": "ein", "start": 0.845, "end": 1.006, "score": 0.398 }, { "word": "planet", "start": 1.046, "end": 1.59, "score": 0.99 }, { "word": "wenn", "start": 2.656, "end": 2.838, "score": 0.123 }, { "word": "die", "start": 2.958, "end": 3.2, "score": 0.058 } ] } de_qtCPktFO38E-00164-00037938-00038162 de000/qtCPktFO38E-00164-00037938-00038162.mp3 2.23 wieder in überlegungen das heißt { "txt": "wieder in überlegungen das heißt", "words": [ { "word": "wieder", "start": 0.08, "end": 0.281, "score": 0.992 }, { "word": "in", "start": 0.342, "end": 0.442, "score": 0.885 }, { "word": "überlegungen", "start": 0.502, "end": 1.205, "score": 0.981 }, { "word": "das", "start": 1.246, "end": 1.386, "score": 0.992 }, { "word": "heißt", "start": 1.426, "end": 1.768, "score": 0.975 } ] } de_1kEKsyXmocu-00155-00039086-00039401 de000/1kEKsyXmocu-00155-00039086-00039401.mp3 3.15 ja das ist sozusagen auch zwei Ebenen { "txt": "ja das ist sozusagen auch zwei Ebenen", "words": [ { "word": "ja", "start": 0.843, "end": 0.923, "score": 0.866 }, { "word": "das", "start": 1.003, "end": 1.144, "score": 0.602 }, { "word": "ist", "start": 1.204, "end": 1.404, "score": 0.924 }, { "word": "sozusagen", "start": 1.505, "end": 1.906, "score": 0.228 }, { "word": "auch", "start": 2.267, "end": 2.388, "score": 0.663 }, { "word": "zwei", "start": 2.428, "end": 2.608, "score": 0.93 }, { "word": "Ebenen", "start": 2.729, "end": 3.03, "score": 0.586 } ] } de_5Ijca4_Uwi4-00175-00084863-00085349 de000/5Ijca4_Uwi4-00175-00084863-00085349.mp3 4.86 klicken oder man kann alles auf einer seite wählen dann werden unabhängig davon was für { "txt": "klicken oder man kann alles auf einer seite wählen dann werden unabhängig davon was für", "words": [ { "word": "klicken", "start": 0.12, "end": 0.382, "score": 0.414 }, { "word": "oder", "start": 0.984, "end": 1.145, "score": 0.996 }, { "word": "man", "start": 1.185, "end": 1.285, "score": 0.996 }, { "word": "kann", "start": 1.305, "end": 1.466, "score": 0.989 }, { "word": "alles", "start": 1.526, "end": 1.727, "score": 0.999 }, { "word": "auf", "start": 1.767, "end": 1.868, "score": 0.996 }, { "word": "einer", "start": 1.948, "end": 2.129, "score": 0.997 }, { "word": "seite", "start": 2.149, "end": 2.39, "score": 0.998 }, { "word": "wählen", "start": 2.45, "end": 2.731, "score": 0.779 }, { "word": "dann", "start": 3.173, "end": 3.314, "score": 0.942 }, { "word": "werden", "start": 3.374, "end": 3.575, "score": 0.976 }, { "word": "unabhängig", "start": 3.635, "end": 4.057, "score": 0.987 }, { "word": "davon", "start": 4.117, "end": 4.478, "score": 0.962 }, { "word": "was", "start": 4.639, "end": 4.76, "score": 0.546 }, { "word": "für", "start": 4.78, "end": 4.86, "score": 0.077 } ] } de_bBm_HDYTX3y-00124-00029425-00029727 de000/bBm_HDYTX3y-00124-00029425-00029727.mp3 3.02 das werdet ihr gleich sehen und ich muss { "txt": "das werdet ihr gleich sehen und ich muss", "words": [ { "word": "das", "start": 0.544, "end": 0.664, "score": 0.982 }, { "word": "werdet", "start": 0.725, "end": 0.946, "score": 0.868 }, { "word": "ihr", "start": 1.007, "end": 1.067, "score": 0.638 }, { "word": "gleich", "start": 1.127, "end": 1.329, "score": 0.991 }, { "word": "sehen", "start": 1.369, "end": 1.591, "score": 0.962 }, { "word": "und", "start": 2.154, "end": 2.396, "score": 0.437 }, { "word": "ich", "start": 2.859, "end": 2.919, "score": 0.095 }, { "word": "muss", "start": 2.919, "end": 3.02, "score": 0.054 } ] } de_8Mn96symhny-00043-00008865-00009067 de000/8Mn96symhny-00043-00008865-00009067.mp3 2.03 deshalb muss der gesetzgeber erneut { "txt": "deshalb muss der gesetzgeber erneut", "words": [ { "word": "deshalb", "start": 0.523, "end": 0.925, "score": 0.993 }, { "word": "muss", "start": 0.985, "end": 1.126, "score": 0.965 }, { "word": "der", "start": 1.166, "end": 1.266, "score": 0.977 }, { "word": "gesetzgeber", "start": 1.286, "end": 1.869, "score": 0.963 }, { "word": "erneut", "start": 1.909, "end": 2.03, "score": 0.182 } ] } de_Ybuns5Julyc-00334-00087798-00088034 de000/Ybuns5Julyc-00334-00087798-00088034.mp3 2.36 getränken wenn die leute sehen mit den { "txt": "getränken wenn die leute sehen mit den", "words": [ { "word": "getränken", "start": 0.061, "end": 1.432, "score": 0.303 }, { "word": "wenn", "start": 1.513, "end": 1.634, "score": 0.37 }, { "word": "die", "start": 1.634, "end": 1.694, "score": 0.35 }, { "word": "leute", "start": 1.735, "end": 1.916, "score": 0.611 }, { "word": "sehen", "start": 1.977, "end": 2.138, "score": 0.328 }, { "word": "mit", "start": 2.138, "end": 2.239, "score": 0.154 }, { "word": "den", "start": 2.239, "end": 2.34, "score": 0.373 } ] } de__M8L33hhReU-00038-00030832-00031683 de000/_M8L33hhReU-00038-00030832-00031683.mp3 8.52 Situation wieder ändert, zum Beispiel zurück zum normalen, dann adaptiert sich auch das motorische { "txt": "Situation wieder ändert, zum Beispiel zurück zum normalen, dann adaptiert sich auch das motorische", "words": [ { "word": "Situation", "start": 0.08, "end": 0.601, "score": 0.996 }, { "word": "wieder", "start": 0.682, "end": 0.942, "score": 0.986 }, { "word": "ändert,", "start": 1.123, "end": 1.604, "score": 0.852 }, { "word": "zum", "start": 2.085, "end": 2.225, "score": 0.878 }, { "word": "Beispiel", "start": 2.265, "end": 2.586, "score": 0.394 }, { "word": "zurück", "start": 2.626, "end": 2.867, "score": 0.953 }, { "word": "zum", "start": 2.927, "end": 3.067, "score": 0.934 }, { "word": "normalen,", "start": 3.127, "end": 3.629, "score": 0.942 }, { "word": "dann", "start": 4.07, "end": 4.27, "score": 0.782 }, { "word": "adaptiert", "start": 5.573, "end": 6.215, "score": 0.982 }, { "word": "sich", "start": 6.315, "end": 6.515, "score": 0.927 }, { "word": "auch", "start": 6.756, "end": 6.976, "score": 0.972 }, { "word": "das", "start": 7.277, "end": 7.538, "score": 0.81 }, { "word": "motorische", "start": 8.139, "end": 8.52, "score": 0.195 } ] } de_u4TWAKv-oOU-00552-00138847-00139056 de000/u4TWAKv-oOU-00552-00138847-00139056.mp3 2.09 kann und das geil ist sollte man das { "txt": "kann und das geil ist sollte man das", "words": [ { "word": "kann", "start": 0.04, "end": 0.221, "score": 0.352 }, { "word": "und", "start": 0.904, "end": 1.105, "score": 0.176 }, { "word": "das", "start": 1.186, "end": 1.326, "score": 0.585 }, { "word": "geil", "start": 1.367, "end": 1.588, "score": 0.458 }, { "word": "ist", "start": 1.648, "end": 1.748, "score": 0.397 }, { "word": "sollte", "start": 1.748, "end": 1.929, "score": 0.138 }, { "word": "man", "start": 1.969, "end": 2.03, "score": 0.241 }, { "word": "das", "start": 2.03, "end": 2.09, "score": 0.132 } ] } de_27_TYj2huaI-00536-00118391-00118571 de000/27_TYj2huaI-00536-00118391-00118571.mp3 1.79 optisch macht es auch noch was da ist { "txt": "optisch macht es auch noch was da ist", "words": [ { "word": "optisch", "start": 0.141, "end": 0.523, "score": 0.597 }, { "word": "macht", "start": 0.583, "end": 0.764, "score": 0.951 }, { "word": "es", "start": 0.804, "end": 0.865, "score": 0.88 }, { "word": "auch", "start": 0.965, "end": 1.066, "score": 0.96 }, { "word": "noch", "start": 1.066, "end": 1.187, "score": 0.565 }, { "word": "was", "start": 1.207, "end": 1.368, "score": 0.697 }, { "word": "da", "start": 1.408, "end": 1.468, "score": 0.207 }, { "word": "ist", "start": 1.589, "end": 1.75, "score": 0.458 } ] } de_0bNyd50mFmM-00339-00067562-00067753 de000/0bNyd50mFmM-00339-00067562-00067753.mp3 1.91 erste teil sprengung von einem der { "txt": "erste teil sprengung von einem der", "words": [ { "word": "erste", "start": 0.161, "end": 0.362, "score": 0.793 }, { "word": "teil", "start": 0.483, "end": 0.704, "score": 0.902 }, { "word": "sprengung", "start": 0.744, "end": 1.106, "score": 0.869 }, { "word": "von", "start": 1.146, "end": 1.367, "score": 0.795 }, { "word": "einem", "start": 1.568, "end": 1.769, "score": 0.351 }, { "word": "der", "start": 1.809, "end": 1.91, "score": 0.318 } ] } de_6labi1dXMxA-00315-00067052-00067223 de000/6labi1dXMxA-00315-00067052-00067223.mp3 1.7 der hat welche mieterhöhung bekommen und { "txt": "der hat welche mieterhöhung bekommen und", "words": [ { "word": "der", "start": 0.142, "end": 0.304, "score": 0.547 }, { "word": "hat", "start": 0.405, "end": 0.546, "score": 0.84 }, { "word": "welche", "start": 0.607, "end": 0.83, "score": 0.928 }, { "word": "mieterhöhung", "start": 0.87, "end": 1.255, "score": 0.757 }, { "word": "bekommen", "start": 1.275, "end": 1.619, "score": 0.767 }, { "word": "und", "start": 1.639, "end": 1.7, "score": 0.144 } ] } de_cBORQKH9GgI-00147-00034888-00035229 de000/cBORQKH9GgI-00147-00034888-00035229.mp3 3.41 der wohnprojekte tag in husum und der { "txt": "der wohnprojekte tag in husum und der", "words": [ { "word": "der", "start": 0.802, "end": 0.923, "score": 0.857 }, { "word": "wohnprojekte", "start": 0.943, "end": 1.685, "score": 0.767 }, { "word": "tag", "start": 1.765, "end": 2.026, "score": 0.81 }, { "word": "in", "start": 2.126, "end": 2.227, "score": 0.743 }, { "word": "husum", "start": 2.287, "end": 2.728, "score": 0.586 }, { "word": "und", "start": 3.229, "end": 3.31, "score": 0.171 }, { "word": "der", "start": 3.33, "end": 3.39, "score": 0.098 } ] } de_91O4zOhwvIQ-00371-00106052-00106248 de000/91O4zOhwvIQ-00371-00106052-00106248.mp3 1.96 instrumentalisiert. { "txt": "instrumentalisiert.", "words": [ { "word": "instrumentalisiert.", "start": 0.202, "end": 1.152, "score": 0.992 } ] } de_OWZCS-manaA-00145-00032130-00032351 de000/OWZCS-manaA-00145-00032130-00032351.mp3 2.21 mit speedfan außensohle und wie gesagt { "txt": "mit speedfan außensohle und wie gesagt", "words": [ { "word": "mit", "start": 0.121, "end": 0.241, "score": 0.466 }, { "word": "speedfan", "start": 0.321, "end": 0.743, "score": 0.62 }, { "word": "außensohle", "start": 0.864, "end": 1.507, "score": 0.635 }, { "word": "und", "start": 1.848, "end": 1.929, "score": 0.513 }, { "word": "wie", "start": 1.949, "end": 2.029, "score": 0.31 }, { "word": "gesagt", "start": 2.049, "end": 2.21, "score": 0.306 } ] } de_h4gAysYE6-8-01397-00319528-00319824 de000/h4gAysYE6-8-01397-00319528-00319824.mp3 2.96 mitarbeiter hat und das alles { "txt": "mitarbeiter hat und das alles", "words": [ { "word": "mitarbeiter", "start": 0.04, "end": 0.483, "score": 0.913 }, { "word": "hat", "start": 0.524, "end": 0.725, "score": 0.802 }, { "word": "und", "start": 2.457, "end": 2.557, "score": 0.51 }, { "word": "das", "start": 2.577, "end": 2.698, "score": 0.561 }, { "word": "alles", "start": 2.779, "end": 2.92, "score": 0.532 } ] } de_iHNat6jUTe4-00085-00060498-00060948 de000/iHNat6jUTe4-00085-00060498-00060948.mp3 4.5 ich ganz ehrlich bin. Soll ich sie erzählen wie ich sie kenneund du kannst jetzt sagen das ist totaler { "txt": "ich ganz ehrlich bin. Soll ich sie erzählen wie ich sie kenneund du kannst jetzt sagen das ist totaler", "words": [ { "word": "ich", "start": 0.04, "end": 0.1, "score": 0.792 }, { "word": "ganz", "start": 0.1, "end": 0.281, "score": 0.81 }, { "word": "ehrlich", "start": 0.342, "end": 0.603, "score": 0.826 }, { "word": "bin.", "start": 0.643, "end": 0.804, "score": 0.889 }, { "word": "Soll", "start": 1.728, "end": 1.848, "score": 0.492 }, { "word": "ich", "start": 1.848, "end": 1.929, "score": 0.297 }, { "word": "sie", "start": 1.929, "end": 2.009, "score": 0.426 }, { "word": "erzählen", "start": 2.069, "end": 2.33, "score": 0.37 }, { "word": "wie", "start": 2.33, "end": 2.431, "score": 0.573 }, { "word": "ich", "start": 2.451, "end": 2.551, "score": 0.981 }, { "word": "sie", "start": 2.571, "end": 2.672, "score": 0.648 }, { "word": "kenneund", "start": 2.712, "end": 3.134, "score": 0.61 }, { "word": "du", "start": 3.154, "end": 3.194, "score": 0.995 }, { "word": "kannst", "start": 3.254, "end": 3.455, "score": 0.677 }, { "word": "jetzt", "start": 3.455, "end": 3.556, "score": 0.075 }, { "word": "sagen", "start": 3.556, "end": 3.797, "score": 0.773 }, { "word": "das", "start": 3.837, "end": 3.917, "score": 0.64 }, { "word": "ist", "start": 3.917, "end": 3.978, "score": 0.317 }, { "word": "totaler", "start": 3.998, "end": 4.44, "score": 0.527 } ] } de_JoAKL3Z0xm8-00704-00194829-00195014 de000/JoAKL3Z0xm8-00704-00194829-00195014.mp3 1.85 beabsichtigt gott aber hatte { "txt": "beabsichtigt gott aber hatte", "words": [ { "word": "beabsichtigt", "start": 0.181, "end": 0.965, "score": 0.682 }, { "word": "gott", "start": 1.026, "end": 1.247, "score": 0.556 }, { "word": "aber", "start": 1.408, "end": 1.649, "score": 0.665 }, { "word": "hatte", "start": 1.689, "end": 1.85, "score": 0.383 } ] } de_LI6RrEQCkXk-00481-00193709-00193846 de000/LI6RrEQCkXk-00481-00193709-00193846.mp3 1.37 standardmäßig alle ausgehenden { "txt": "standardmäßig alle ausgehenden", "words": [ { "word": "standardmäßig", "start": 0.101, "end": 0.766, "score": 0.833 }, { "word": "alle", "start": 0.826, "end": 0.967, "score": 0.697 }, { "word": "ausgehenden", "start": 1.007, "end": 1.37, "score": 0.457 } ] } de_63eypVypFI0-00020-00004578-00004763 de000/63eypVypFI0-00020-00004578-00004763.mp3 1.85 wieder gefragt wie ich meine trainings { "txt": "wieder gefragt wie ich meine trainings", "words": [ { "word": "wieder", "start": 0.201, "end": 0.382, "score": 0.971 }, { "word": "gefragt", "start": 0.402, "end": 0.885, "score": 0.989 }, { "word": "wie", "start": 0.925, "end": 1.005, "score": 0.881 }, { "word": "ich", "start": 1.026, "end": 1.106, "score": 0.91 }, { "word": "meine", "start": 1.126, "end": 1.327, "score": 0.908 }, { "word": "trainings", "start": 1.408, "end": 1.85, "score": 0.657 } ] } de_pdRGd3tvN9y-00025-00013513-00013647 de000/pdRGd3tvN9y-00025-00013513-00013647.mp3 1.34 geben zu können { "txt": "geben zu können", "words": [ { "word": "geben", "start": 0.223, "end": 0.528, "score": 0.627 }, { "word": "zu", "start": 0.589, "end": 0.69, "score": 0.401 }, { "word": "können", "start": 0.792, "end": 1.198, "score": 0.623 } ] } de_Vra_BKtXd1M-00034-00006792-00007028 de000/Vra_BKtXd1M-00034-00006792-00007028.mp3 2.36 welt noch nicht so wirklich kennt es ist { "txt": "welt noch nicht so wirklich kennt es ist", "words": [ { "word": "welt", "start": 0.061, "end": 0.222, "score": 0.693 }, { "word": "noch", "start": 0.262, "end": 0.383, "score": 0.9 }, { "word": "nicht", "start": 0.403, "end": 0.605, "score": 0.845 }, { "word": "so", "start": 0.645, "end": 0.706, "score": 0.944 }, { "word": "wirklich", "start": 0.787, "end": 1.13, "score": 0.899 }, { "word": "kennt", "start": 1.19, "end": 1.472, "score": 0.659 }, { "word": "es", "start": 1.896, "end": 2.017, "score": 0.237 }, { "word": "ist", "start": 2.057, "end": 2.36, "score": 0.127 } ] } de_fkt-cg7r4RA-00051-00013587-00013712 de000/fkt-cg7r4RA-00051-00013587-00013712.mp3 1.25 menschlichen gesundheit { "txt": "menschlichen gesundheit", "words": [ { "word": "menschlichen", "start": 0.081, "end": 0.565, "score": 0.95 }, { "word": "gesundheit", "start": 0.605, "end": 1.19, "score": 0.993 } ] } de_XdEQI32GirE-00197-00035201-00035344 de000/XdEQI32GirE-00197-00035201-00035344.mp3 1.43 raketenantriebe die treibstoff { "txt": "raketenantriebe die treibstoff", "words": [ { "word": "raketenantriebe", "start": 0.081, "end": 0.826, "score": 0.927 }, { "word": "die", "start": 1.067, "end": 1.148, "score": 0.597 }, { "word": "treibstoff", "start": 1.168, "end": 1.43, "score": 0.151 } ] } de_R5imut2Lnpk-00087-00025060-00025397 de000/R5imut2Lnpk-00087-00025060-00025397.mp3 3.36 realisiert das holzgerüst dass sich dies { "txt": "realisiert das holzgerüst dass sich dies", "words": [ { "word": "realisiert", "start": 0.362, "end": 1.026, "score": 0.99 }, { "word": "das", "start": 1.066, "end": 1.207, "score": 0.995 }, { "word": "holzgerüst", "start": 1.288, "end": 2.012, "score": 0.976 }, { "word": "dass", "start": 2.756, "end": 2.897, "score": 0.666 }, { "word": "sich", "start": 2.937, "end": 3.098, "score": 0.453 }, { "word": "dies", "start": 3.139, "end": 3.34, "score": 0.677 } ] } de_-Z-09GB_P2u-00795-00266183-00266437 de000/-Z-09GB_P2u-00795-00266183-00266437.mp3 2.54 sodass alle Kinder in Hessen, egal ob in Gebieten { "txt": "sodass alle Kinder in Hessen, egal ob in Gebieten", "words": [ { "word": "sodass", "start": 0.04, "end": 0.262, "score": 0.988 }, { "word": "alle", "start": 0.383, "end": 0.564, "score": 0.958 }, { "word": "Kinder", "start": 0.605, "end": 0.887, "score": 0.999 }, { "word": "in", "start": 0.947, "end": 1.028, "score": 0.949 }, { "word": "Hessen,", "start": 1.068, "end": 1.411, "score": 0.964 }, { "word": "egal", "start": 1.532, "end": 1.875, "score": 0.999 }, { "word": "ob", "start": 1.935, "end": 2.036, "score": 0.999 }, { "word": "in", "start": 2.076, "end": 2.157, "score": 0.992 }, { "word": "Gebieten", "start": 2.197, "end": 2.54, "score": 0.966 } ] } de_TOG-aw1Ao2Y-00437-00132427-00132648 de000/TOG-aw1Ao2Y-00437-00132427-00132648.mp3 2.21 nächsten donnerstag also jetzt in dieser { "txt": "nächsten donnerstag also jetzt in dieser", "words": [ { "word": "nächsten", "start": 0.462, "end": 0.884, "score": 0.947 }, { "word": "donnerstag", "start": 0.924, "end": 1.527, "score": 0.876 }, { "word": "also", "start": 1.567, "end": 1.688, "score": 0.811 }, { "word": "jetzt", "start": 1.748, "end": 1.949, "score": 0.87 }, { "word": "in", "start": 1.969, "end": 2.049, "score": 0.673 }, { "word": "dieser", "start": 2.069, "end": 2.21, "score": 0.494 } ] } de_nzWwhSBOXRY-00253-00083537-00083779 de000/nzWwhSBOXRY-00253-00083537-00083779.mp3 2.42 ein bisschen gewicht aber so ist er ganz { "txt": "ein bisschen gewicht aber so ist er ganz", "words": [ { "word": "ein", "start": 0.04, "end": 0.161, "score": 0.298 }, { "word": "bisschen", "start": 0.161, "end": 0.343, "score": 0.534 }, { "word": "gewicht", "start": 0.363, "end": 0.686, "score": 0.963 }, { "word": "aber", "start": 1.714, "end": 1.916, "score": 0.922 }, { "word": "so", "start": 2.077, "end": 2.138, "score": 0.802 }, { "word": "ist", "start": 2.178, "end": 2.239, "score": 0.3 }, { "word": "er", "start": 2.259, "end": 2.299, "score": 0.371 }, { "word": "ganz", "start": 2.319, "end": 2.42, "score": 0.525 } ] } de_M4c-ze33DH0-00451-00142279-00142779 de000/M4c-ze33DH0-00451-00142279-00142779.mp3 5 endes race am sich das ndr kultur einer { "txt": "endes race am sich das ndr kultur einer", "words": [ { "word": "endes", "start": 0.341, "end": 0.663, "score": 0.864 }, { "word": "race", "start": 0.944, "end": 1.365, "score": 0.308 }, { "word": "am", "start": 1.486, "end": 1.606, "score": 0.518 }, { "word": "sich", "start": 1.767, "end": 2.068, "score": 0.922 }, { "word": "das", "start": 2.45, "end": 2.59, "score": 0.894 }, { "word": "ndr", "start": 2.851, "end": 2.992, "score": 0.413 }, { "word": "kultur", "start": 3.032, "end": 3.414, "score": 0.281 }, { "word": "einer", "start": 3.815, "end": 4.98, "score": 0.331 } ] } de_DwpxnrgvmbY-00054-00028902-00029198 de000/DwpxnrgvmbY-00054-00028902-00029198.mp3 2.95 WAS sich nicht ausschließen muss. { "txt": "WAS sich nicht ausschließen muss.", "words": [ { "word": "WAS", "start": 1.204, "end": 1.365, "score": 0.998 }, { "word": "sich", "start": 1.385, "end": 1.505, "score": 0.998 }, { "word": "nicht", "start": 1.545, "end": 2.087, "score": 0.953 }, { "word": "ausschließen", "start": 2.147, "end": 2.589, "score": 0.992 }, { "word": "muss.", "start": 2.589, "end": 2.749, "score": 0.89 } ] } de_Az0lDvtH1pI-00002-00001112-00001673 de000/Az0lDvtH1pI-00002-00001112-00001673.mp3 5.61 gut das letzte Referat hält wie erwähnt { "txt": "gut das letzte Referat hält wie erwähnt", "words": [ { "word": "gut", "start": 0.701, "end": 1.042, "score": 0.558 }, { "word": "das", "start": 1.723, "end": 1.984, "score": 0.971 }, { "word": "letzte", "start": 2.204, "end": 2.665, "score": 0.757 }, { "word": "Referat", "start": 2.745, "end": 3.366, "score": 0.953 }, { "word": "hält", "start": 3.967, "end": 4.328, "score": 0.568 }, { "word": "wie", "start": 4.448, "end": 4.688, "score": 0.34 }, { "word": "erwähnt", "start": 4.949, "end": 5.51, "score": 0.572 } ] } de_soAd98u94TA-00139-00028438-00028557 de000/soAd98u94TA-00139-00028438-00028557.mp3 1.18 also die bürgerenergiegenossenschaft { "txt": "also die bürgerenergiegenossenschaft", "words": [ { "word": "also", "start": 0.081, "end": 0.183, "score": 0.338 }, { "word": "die", "start": 0.244, "end": 0.346, "score": 0.445 }, { "word": "bürgerenergiegenossenschaft", "start": 0.366, "end": 1.18, "score": 0.121 } ] } de_l2-HdrH4XEc-00103-00066402-00067026 de000/l2-HdrH4XEc-00103-00066402-00067026.mp3 6.24 merkst, was wird gefragt und dann hab ich gemerkt, okay die Frage stelle ich mir normalerweise auch, { "txt": "merkst, was wird gefragt und dann hab ich gemerkt, okay die Frage stelle ich mir normalerweise auch,", "words": [ { "word": "merkst,", "start": 0.08, "end": 0.421, "score": 0.991 }, { "word": "was", "start": 0.482, "end": 0.662, "score": 0.978 }, { "word": "wird", "start": 0.803, "end": 0.983, "score": 0.993 }, { "word": "gefragt", "start": 1.083, "end": 1.705, "score": 0.994 }, { "word": "und", "start": 1.806, "end": 1.926, "score": 0.976 }, { "word": "dann", "start": 1.966, "end": 2.147, "score": 0.956 }, { "word": "hab", "start": 3.15, "end": 3.27, "score": 0.995 }, { "word": "ich", "start": 3.371, "end": 3.491, "score": 0.979 }, { "word": "gemerkt,", "start": 3.551, "end": 3.973, "score": 0.984 }, { "word": "okay", "start": 4.193, "end": 4.454, "score": 0.495 }, { "word": "die", "start": 4.595, "end": 4.695, "score": 0.995 }, { "word": "Frage", "start": 4.715, "end": 4.956, "score": 0.998 }, { "word": "stelle", "start": 4.976, "end": 5.197, "score": 0.796 }, { "word": "ich", "start": 5.197, "end": 5.257, "score": 0.942 }, { "word": "mir", "start": 5.277, "end": 5.357, "score": 0.967 }, { "word": "normalerweise", "start": 5.397, "end": 5.939, "score": 0.999 }, { "word": "auch,", "start": 6.039, "end": 6.22, "score": 0.585 } ] } de_h1ZhNBt6K_A-01011-00238285-00238440 de000/h1ZhNBt6K_A-01011-00238285-00238440.mp3 1.55 initiativen gestartet { "txt": "initiativen gestartet", "words": [ { "word": "initiativen", "start": 0.262, "end": 0.886, "score": 0.988 }, { "word": "gestartet", "start": 0.926, "end": 1.51, "score": 0.882 } ] } de_qO0PuCy0e48-00086-00047282-00047374 de000/qO0PuCy0e48-00086-00047282-00047374.mp3 0.92 körpergewicht { "txt": "körpergewicht", "words": [ { "word": "körpergewicht", "start": 0.082, "end": 0.859, "score": 0.376 } ] } de_Ixw81zpf3GI-00105-00024840-00024964 de000/Ixw81zpf3GI-00105-00024840-00024964.mp3 1.25 bedeutet aber wie gesagt noch lange { "txt": "bedeutet aber wie gesagt noch lange", "words": [ { "word": "bedeutet", "start": 0, "end": 0.282, "score": 0.699 }, { "wor
2026-01-13T09:29:19
https://www.linkedin.com/products/redwood-software-jscape-by-redwood/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_full-click
JSCAPE by Redwood | LinkedIn Skip to main content LinkedIn Redwood Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in JSCAPE by Redwood Managed File Transfer (MFT) Software by Redwood Software See who's skilled in this Add as skill Request demo Report this product About Automate your file transfers while staying secure and compliant with JSCAPE by Redwood, able to support any platform, protocol or environment. Easy to use and set up, IT teams are empowered with a single view into all file transfer operations, with real-time notifications and alerts. With a built-in library of 100+ no-code/low-code workflow automation templates, REST APIs and off-the-shelf integrations, your experts have the tools they need to deliver reliable file transfer operations. Experience the power of scalable file transfer with JSCAPE. Experience the power of secure, seamless and scalable file transfer with JSCAPE. Elevate your enterprise's data management capabilities and unlock new levels of productivity and agility today. Media Products media viewer No more previous content Admin console Network storage Trading partner JSCAPE embedded in RunMyJobs by Redwood No more next content Featured customers of JSCAPE by Redwood SAP Software Development 4,679,808 followers L'Oréal Personal Care Product Manufacturing 6,347,679 followers Starbucks Retail 3,091,714 followers Similar products Axway Managed File Transfer Axway Managed File Transfer Managed File Transfer (MFT) Software Progress MOVEit Progress MOVEit Managed File Transfer (MFT) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software MLADU MLADU Managed File Transfer (MFT) Software dDataBox dDataBox Managed File Transfer (MFT) Software Sign in to see more Show more Show less Redwood Software products ActiveBatch by Redwood ActiveBatch by Redwood Workload Automation Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software Finance Automation by Redwood Finance Automation by Redwood Financial Close Software RunMyJobs by Redwood RunMyJobs by Redwood Workload Automation Software Tidal by Redwood Tidal by Redwood Workload Automation Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/reference/source-replacement.html#source-replacement
Source Replacement - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book Source Replacement This document is about redirecting communication with registries or repositories of git-based dependencies to another data source, such as a server mirroring the original registry or an exact local copy. If you want to patch individual dependencies, see overriding dependencies section of this documentation. If you want to control how Cargo makes network requests, see [http] and [net] configuration. A source is a provider that contains crates that may be included as dependencies for a package. Cargo supports the ability to replace one source with another to express strategies such as: Vendoring — custom sources can be defined which represent crates on the local filesystem. These sources are subsets of the source that they’re replacing and can be checked into packages if necessary. Mirroring — sources can be replaced with an equivalent version which acts as a cache for crates.io itself. Cargo has a core assumption about source replacement that the source code is exactly the same from both sources. Note that this also means that a replacement source is not allowed to have crates which are not present in the original source. As a consequence, source replacement is not appropriate for situations such as patching a dependency or a private registry. Cargo supports patching dependencies through the usage of the [patch] key , and private registry support is described in the Registries chapter . When using source replacement, running commands that need to contact the registry directly 1 requires passing the --registry option. This helps avoid any ambiguity about which registry to contact, and will use the authentication token for the specified registry. Configuration Configuration of replacement sources is done through .cargo/config.toml and the full set of available keys are: # The `source` table is where all keys related to source-replacement # are stored. [source] # Under the `source` table are a number of other tables whose keys are a # name for the relevant source. For example this section defines a new # source, called `my-vendor-source`, which comes from a directory # located at `vendor` relative to the directory containing this `.cargo/config.toml` # file [source.my-vendor-source] directory = "vendor" # The crates.io default source for crates is available under the name # "crates-io", and here we use the `replace-with` key to indicate that it's # replaced with our source above. # # The `replace-with` key can also reference an alternative registry name # defined in the `[registries]` table. [source.crates-io] replace-with = "my-vendor-source" # Each source has its own table where the key is the name of the source [source.the-source-name] # Indicate that `the-source-name` will be replaced with `another-source`, # defined elsewhere replace-with = "another-source" # Several kinds of sources can be specified (described in more detail below): registry = "https://example.com/path/to/index" local-registry = "path/to/registry" directory = "path/to/vendor" # Git sources can optionally specify a branch/tag/rev as well git = "https://example.com/path/to/repo" # branch = "master" # tag = "v1.0.1" # rev = "313f44e8" Registry Sources A “registry source” is one that works like crates.io itself. It’s an index that conforms to the specification at https://doc.rust-lang.org/cargo/reference/registry-index.html with a configuration file indicating where to download crates from. Registry sources can use either git or sparse HTTP protocol : # Git protocol registry = "ssh://git@example.com/path/to/index.git" # Sparse HTTP protocol registry = "sparse+https://example.com/path/to/index" # HTTPS git protocol registry = "https://example.com/path/to/index" Local Registry Sources A “local registry source” is intended to be a subset of another registry source, but available on the local filesystem (aka vendoring). Local registries are downloaded ahead of time, typically sync’d with a Cargo.lock , and are made up of a set of *.crate files and an index like the normal registry is. The primary way to manage and create local registry sources is through the cargo-local-registry subcommand, available on crates.io and can be installed with cargo install cargo-local-registry . Local registries are contained within one directory and contain a number of *.crate files downloaded from crates.io as well as an index directory with the same format as the crates.io-index project (populated with just entries for the crates that are present). Directory Sources A “directory source” is similar to a local registry source where it contains a number of crates available on the local filesystem, suitable for vendoring dependencies. Directory sources are primarily managed by the cargo vendor subcommand. Directory sources are distinct from local registries though in that they contain the unpacked version of *.crate files, making it more suitable in some situations to check everything into source control. A directory source is just a directory containing a number of other directories which contain the source code for crates (the unpacked version of *.crate files). Currently no restriction is placed on the name of each directory. Each crate in a directory source also has an associated metadata file indicating the checksum of each file in the crate to protect against accidental modifications. Git sources Git sources represent repositories used by git-based dependencies . They’re used to specify which git-based dependencies should be replaced with alternative sources. Git sources are not related to the git registries , and can’t be used to replace registry sources. Examples of such commands are in Publishing Commands . ↩
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/guide/creating-a-new-project.html#creating-a-new-package
Creating a New Package - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book Creating a New Package To start a new package with Cargo, use cargo new : $ cargo new hello_world --bin We’re passing --bin because we’re making a binary program: if we were making a library, we’d pass --lib . This also initializes a new git repository by default. If you don’t want it to do that, pass --vcs none . Let’s check out what Cargo has generated for us: $ cd hello_world $ tree . . ├── Cargo.toml └── src └── main.rs 1 directory, 2 files Let’s take a closer look at Cargo.toml : [package] name = "hello_world" version = "0.1.0" edition = "2024" [dependencies] This is called a manifest , and it contains all of the metadata that Cargo needs to compile your package. This file is written in the TOML format (pronounced /tɑməl/). Here’s what’s in src/main.rs : fn main() { println!("Hello, world!"); } Cargo generated a “hello world” program for you, otherwise known as a binary crate . Let’s compile it: $ cargo build Compiling hello_world v0.1.0 (file:///path/to/package/hello_world) And then run it: $ ./target/debug/hello_world Hello, world! You can also use cargo run to compile and then run it, all in one step (You won’t see the Compiling line if you have not made any changes since you last compiled): $ cargo run Compiling hello_world v0.1.0 (file:///path/to/package/hello_world) Running `target/debug/hello_world` Hello, world! You’ll now notice a new file, Cargo.lock . It contains information about your dependencies. Since there are none yet, it’s not very interesting. Once you’re ready for release, you can use cargo build --release to compile your files with optimizations turned on: $ cargo build --release Compiling hello_world v0.1.0 (file:///path/to/package/hello_world) cargo build --release puts the resulting binary in target/release instead of target/debug . Compiling in debug mode is the default for development. Compilation time is shorter since the compiler doesn’t do optimizations, but the code will run slower. Release mode takes longer to compile, but the code will run faster.
2026-01-13T09:29:19
https://www.linkedin.com/products/akamai-technologies-kona-ddos-defender/?trk=products_seo_search
Kona DDoS Defender | LinkedIn Skip to main content LinkedIn Akamai Technologies in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Kona DDoS Defender DDoS Protection Software by Akamai Technologies See who's skilled in this Add as skill Learn more Report this product About Always-on, managed DDoS protection service to stop DDoS for critical websites and web applications. HTTP and HTTPS. Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less Akamai Technologies products Akamai Edge DNS Akamai Edge DNS Managed DNS Services Software Akamai Identity Cloud Akamai Identity Cloud Identity & Access Management (IAM) Software Akamai IoT Edge Connect Akamai IoT Edge Connect Internet of Things (IoT) Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software Aura Managed CDN Aura Managed CDN Content Delivery Network (CDN) Software BOCC BOCC Live Streaming Software DNSi AuthServe DNSi AuthServe Managed DNS Services Software DNSi Big Data Connector DNSi Big Data Connector Managed DNS Services Software Enterprise Threat Protector Enterprise Threat Protector Secure Web Gateways Media Services Live Media Services Live Live Streaming Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/reference/config.html#registrytoken
Configuration - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book Configuration This document explains how Cargo’s configuration system works, as well as available keys or configuration. For configuration of a package through its manifest, see the manifest format . Hierarchical structure Cargo allows local configuration for a particular package as well as global configuration. It looks for configuration files in the current directory and all parent directories. If, for example, Cargo were invoked in /projects/foo/bar/baz , then the following configuration files would be probed for and unified in this order: /projects/foo/bar/baz/.cargo/config.toml /projects/foo/bar/.cargo/config.toml /projects/foo/.cargo/config.toml /projects/.cargo/config.toml /.cargo/config.toml $CARGO_HOME/config.toml which defaults to: Windows: %USERPROFILE%\.cargo\config.toml Unix: $HOME/.cargo/config.toml With this structure, you can specify configuration per-package, and even possibly check it into version control. You can also specify personal defaults with a configuration file in your home directory. If a key is specified in multiple config files, the values will get merged together. Numbers, strings, and booleans will use the value in the deeper config directory taking precedence over ancestor directories, where the home directory is the lowest priority. Arrays will be joined together with higher precedence items being placed later in the merged array. At present, when being invoked from a workspace, Cargo does not read config files from crates within the workspace. i.e. if a workspace has two crates in it, named /projects/foo/bar/baz/mylib and /projects/foo/bar/baz/mybin , and there are Cargo configs at /projects/foo/bar/baz/mylib/.cargo/config.toml and /projects/foo/bar/baz/mybin/.cargo/config.toml , Cargo does not read those configuration files if it is invoked from the workspace root ( /projects/foo/bar/baz/ ). Note: Cargo also reads config files without the .toml extension, such as .cargo/config . Support for the .toml extension was added in version 1.39 and is the preferred form. If both files exist, Cargo will use the file without the extension. Configuration format Configuration files are written in the TOML format (like the manifest), with simple key-value pairs inside of sections (tables). The following is a quick overview of all settings, with detailed descriptions found below. paths = ["/path/to/override"] # path dependency overrides [alias] # command aliases b = "build" c = "check" t = "test" r = "run" rr = "run --release" recursive_example = "rr --example recursions" space_example = ["run", "--release", "--", "\"command list\""] [build] jobs = 1 # number of parallel jobs, defaults to # of CPUs rustc = "rustc" # the rust compiler tool rustc-wrapper = "…" # run this wrapper instead of `rustc` rustc-workspace-wrapper = "…" # run this wrapper instead of `rustc` for workspace members rustdoc = "rustdoc" # the doc generator tool target = "triple" # build for the target triple (ignored by `cargo install`) target-dir = "target" # path of where to place generated artifacts build-dir = "target" # path of where to place intermediate build artifacts rustflags = ["…", "…"] # custom flags to pass to all compiler invocations rustdocflags = ["…", "…"] # custom flags to pass to rustdoc incremental = true # whether or not to enable incremental compilation dep-info-basedir = "…" # path for the base directory for targets in depfiles [credential-alias] # Provides a way to define aliases for credential providers. my-alias = ["/usr/bin/cargo-credential-example", "--argument", "value", "--flag"] [doc] browser = "chromium" # browser to use with `cargo doc --open`, # overrides the `BROWSER` environment variable [env] # Set ENV_VAR_NAME=value for any process run by Cargo ENV_VAR_NAME = "value" # Set even if already present in environment ENV_VAR_NAME_2 = { value = "value", force = true } # `value` is relative to the parent of `.cargo/config.toml`, env var will be the full absolute path ENV_VAR_NAME_3 = { value = "relative/path", relative = true } [future-incompat-report] frequency = 'always' # when to display a notification about a future incompat report [cache] auto-clean-frequency = "1 day" # How often to perform automatic cache cleaning [cargo-new] vcs = "none" # VCS to use ('git', 'hg', 'pijul', 'fossil', 'none') [http] debug = false # HTTP debugging proxy = "host:port" # HTTP proxy in libcurl format ssl-version = "tlsv1.3" # TLS version to use ssl-version.max = "tlsv1.3" # maximum TLS version ssl-version.min = "tlsv1.1" # minimum TLS version timeout = 30 # timeout for each HTTP request, in seconds low-speed-limit = 10 # network timeout threshold (bytes/sec) cainfo = "cert.pem" # path to Certificate Authority (CA) bundle proxy-cainfo = "cert.pem" # path to proxy Certificate Authority (CA) bundle check-revoke = true # check for SSL certificate revocation multiplexing = true # HTTP/2 multiplexing user-agent = "…" # the user-agent header [install] root = "/some/path" # `cargo install` destination directory [net] retry = 3 # network retries git-fetch-with-cli = true # use the `git` executable for git operations offline = true # do not access the network [net.ssh] known-hosts = ["..."] # known SSH host keys [patch.<registry>] # Same keys as for [patch] in Cargo.toml [profile.<name>] # Modify profile settings via config. inherits = "dev" # Inherits settings from [profile.dev]. opt-level = 0 # Optimization level. debug = true # Include debug info. split-debuginfo = '...' # Debug info splitting behavior. strip = "none" # Removes symbols or debuginfo. debug-assertions = true # Enables debug assertions. overflow-checks = true # Enables runtime integer overflow checks. lto = false # Sets link-time optimization. panic = 'unwind' # The panic strategy. incremental = true # Incremental compilation. codegen-units = 16 # Number of code generation units. rpath = false # Sets the rpath linking option. [profile.<name>.build-override] # Overrides build-script settings. # Same keys for a normal profile. [profile.<name>.package.<name>] # Override profile for a package. # Same keys for a normal profile (minus `panic`, `lto`, and `rpath`). [resolver] incompatible-rust-versions = "allow" # Specifies how resolver reacts to these [registries.<name>] # registries other than crates.io index = "…" # URL of the registry index token = "…" # authentication token for the registry credential-provider = "cargo:token" # The credential provider for this registry. [registries.crates-io] protocol = "sparse" # The protocol to use to access crates.io. [registry] default = "…" # name of the default registry token = "…" # authentication token for crates.io credential-provider = "cargo:token" # The credential provider for crates.io. global-credential-providers = ["cargo:token"] # The credential providers to use by default. [source.<name>] # source definition and replacement replace-with = "…" # replace this source with the given named source directory = "…" # path to a directory source registry = "…" # URL to a registry source local-registry = "…" # path to a local registry source git = "…" # URL of a git repository source branch = "…" # branch name for the git repository tag = "…" # tag name for the git repository rev = "…" # revision for the git repository [target.<triple>] linker = "…" # linker to use runner = "…" # wrapper to run executables rustflags = ["…", "…"] # custom flags for `rustc` rustdocflags = ["…", "…"] # custom flags for `rustdoc` [target.<cfg>] linker = "…" # linker to use runner = "…" # wrapper to run executables rustflags = ["…", "…"] # custom flags for `rustc` [target.<triple>.<links>] # `links` build script override rustc-link-lib = ["foo"] rustc-link-search = ["/path/to/foo"] rustc-flags = "-L /some/path" rustc-cfg = ['key="value"'] rustc-env = {key = "value"} rustc-cdylib-link-arg = ["…"] metadata_key1 = "value" metadata_key2 = "value" [term] quiet = false # whether cargo output is quiet verbose = false # whether cargo provides verbose output color = 'auto' # whether cargo colorizes output hyperlinks = true # whether cargo inserts links into output unicode = true # whether cargo can render output using non-ASCII unicode characters progress.when = 'auto' # whether cargo shows progress bar progress.width = 80 # width of progress bar progress.term-integration = true # whether cargo reports progress to terminal emulator Environment variables Cargo can also be configured through environment variables in addition to the TOML configuration files. For each configuration key of the form foo.bar the environment variable CARGO_FOO_BAR can also be used to define the value. Keys are converted to uppercase, dots and dashes are converted to underscores. For example the target.x86_64-unknown-linux-gnu.runner key can also be defined by the CARGO_TARGET_X86_64_UNKNOWN_LINUX_GNU_RUNNER environment variable. Environment variables will take precedence over TOML configuration files. Currently only integer, boolean, string and some array values are supported to be defined by environment variables. Descriptions below indicate which keys support environment variables and otherwise they are not supported due to technical issues . In addition to the system above, Cargo recognizes a few other specific environment variables . Command-line overrides Cargo also accepts arbitrary configuration overrides through the --config command-line option. The argument should be in TOML syntax of KEY=VALUE or provided as a path to an extra configuration file: # With `KEY=VALUE` in TOML syntax cargo --config net.git-fetch-with-cli=true fetch # With a path to a configuration file cargo --config ./path/to/my/extra-config.toml fetch The --config option may be specified multiple times, in which case the values are merged in left-to-right order, using the same merging logic that is used when multiple configuration files apply. Configuration values specified this way take precedence over environment variables, which take precedence over configuration files. When the --config option is provided as an extra configuration file, The configuration file loaded this way follow the same precedence rules as other options specified directly with --config . Some examples of what it looks like using Bourne shell syntax: # Most shells will require escaping. cargo --config http.proxy=\"http://example.com\" … # Spaces may be used. cargo --config "net.git-fetch-with-cli = true" … # TOML array example. Single quotes make it easier to read and write. cargo --config 'build.rustdocflags = ["--html-in-header", "header.html"]' … # Example of a complex TOML key. cargo --config "target.'cfg(all(target_arch = \"arm\", target_os = \"none\"))'.runner = 'my-runner'" … # Example of overriding a profile setting. cargo --config profile.dev.package.image.opt-level=3 … Config-relative paths Paths in config files may be absolute, relative, or a bare name without any path separators. Paths for executables without a path separator will use the PATH environment variable to search for the executable. Paths for non-executables will be relative to where the config value is defined. In particular, rules are: For environment variables, paths are relative to the current working directory. For config values loaded directly from the --config KEY=VALUE option, paths are relative to the current working directory. For config files, paths are relative to the parent directory of the directory where the config files were defined, no matter those files are from either the hierarchical probing or the --config <path> option. Note: To maintain consistency with existing .cargo/config.toml probing behavior, it is by design that a path in a config file passed via --config <path> is also relative to two levels up from the config file itself. To avoid unexpected results, the rule of thumb is putting your extra config files at the same level of discovered .cargo/config.toml in your project. For instance, given a project /my/project , it is recommended to put config files under /my/project/.cargo or a new directory at the same level, such as /my/project/.config . # Relative path examples. [target.x86_64-unknown-linux-gnu] runner = "foo" # Searches `PATH` for `foo`. [source.vendored-sources] # Directory is relative to the parent where `.cargo/config.toml` is located. # For example, `/my/project/.cargo/config.toml` would result in `/my/project/vendor`. directory = "vendor" Executable paths with arguments Some Cargo commands invoke external programs, which can be configured as a path and some number of arguments. The value may be an array of strings like ['/path/to/program', 'somearg'] or a space-separated string like '/path/to/program somearg' . If the path to the executable contains a space, the list form must be used. If Cargo is passing other arguments to the program such as a path to open or run, they will be passed after the last specified argument in the value of an option of this format. If the specified program does not have path separators, Cargo will search PATH for its executable. Credentials Configuration values with sensitive information are stored in the $CARGO_HOME/credentials.toml file. This file is automatically created and updated by cargo login and cargo logout when using the cargo:token credential provider. Tokens are used by some Cargo commands such as cargo publish for authenticating with remote registries. Care should be taken to protect the tokens and to keep them secret. It follows the same format as Cargo config files. [registry] token = "…" # Access token for crates.io [registries.<name>] token = "…" # Access token for the named registry As with most other config values, tokens may be specified with environment variables. The token for crates.io may be specified with the CARGO_REGISTRY_TOKEN environment variable. Tokens for other registries may be specified with environment variables of the form CARGO_REGISTRIES_<name>_TOKEN where <name> is the name of the registry in all capital letters. Note: Cargo also reads and writes credential files without the .toml extension, such as .cargo/credentials . Support for the .toml extension was added in version 1.39. In version 1.68, Cargo writes to the file with the extension by default. However, for backward compatibility reason, when both files exist, Cargo will read and write the file without the extension. Configuration keys This section documents all configuration keys. The description for keys with variable parts are annotated with angled brackets like target.<triple> where the <triple> part can be any target triple like target.x86_64-pc-windows-msvc . paths Type: array of strings (paths) Default: none Environment: not supported An array of paths to local packages which are to be used as overrides for dependencies. For more information see the Overriding Dependencies guide . [alias] Type: string or array of strings Default: see below Environment: CARGO_ALIAS_<name> The [alias] table defines CLI command aliases. For example, running cargo b is an alias for running cargo build . Each key in the table is the subcommand, and the value is the actual command to run. The value may be an array of strings, where the first element is the command and the following are arguments. It may also be a string, which will be split on spaces into subcommand and arguments. The following aliases are built-in to Cargo: [alias] b = "build" c = "check" d = "doc" t = "test" r = "run" rm = "remove" Aliases are not allowed to redefine existing built-in commands. Aliases are recursive: [alias] rr = "run --release" recursive_example = "rr --example recursions" [build] The [build] table controls build-time operations and compiler settings. build.jobs Type: integer or string Default: number of logical CPUs Environment: CARGO_BUILD_JOBS Sets the maximum number of compiler processes to run in parallel. If negative, it sets the maximum number of compiler processes to the number of logical CPUs plus provided value. Should not be 0. If a string default is provided, it sets the value back to defaults. Can be overridden with the --jobs CLI option. build.rustc Type: string (program path) Default: "rustc" Environment: CARGO_BUILD_RUSTC or RUSTC Sets the executable to use for rustc . build.rustc-wrapper Type: string (program path) Default: none Environment: CARGO_BUILD_RUSTC_WRAPPER or RUSTC_WRAPPER Sets a wrapper to execute instead of rustc . The first argument passed to the wrapper is the path to the actual executable to use (i.e., build.rustc , if that is set, or "rustc" otherwise). build.rustc-workspace-wrapper Type: string (program path) Default: none Environment: CARGO_BUILD_RUSTC_WORKSPACE_WRAPPER or RUSTC_WORKSPACE_WRAPPER Sets a wrapper to execute instead of rustc , for workspace members only. When building a single-package project without workspaces, that package is considered to be the workspace. The first argument passed to the wrapper is the path to the actual executable to use (i.e., build.rustc , if that is set, or "rustc" otherwise). It affects the filename hash so that artifacts produced by the wrapper are cached separately. If both rustc-wrapper and rustc-workspace-wrapper are set, then they will be nested: the final invocation is $RUSTC_WRAPPER $RUSTC_WORKSPACE_WRAPPER $RUSTC . build.rustdoc Type: string (program path) Default: "rustdoc" Environment: CARGO_BUILD_RUSTDOC or RUSTDOC Sets the executable to use for rustdoc . build.target Type: string or array of strings Default: host platform Environment: CARGO_BUILD_TARGET The default target platform triples to compile to. Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. Can be overridden with the --target CLI option. [build] target = ["x86_64-unknown-linux-gnu", "i686-unknown-linux-gnu"] build.target-dir Type: string (path) Default: "target" Environment: CARGO_BUILD_TARGET_DIR or CARGO_TARGET_DIR The path to where all compiler output is placed. The default if not specified is a directory named target located at the root of the workspace. Can be overridden with the --target-dir CLI option. For more information see the build cache documentation . build.build-dir Type: string (path) Default: Defaults to the value of build.target-dir Environment: CARGO_BUILD_BUILD_DIR The directory where intermediate build artifacts will be stored. Intermediate artifacts are produced by Rustc/Cargo during the build process. This option supports path templating. Available template variables: {workspace-root} resolves to root of the current workspace. {cargo-cache-home} resolves to CARGO_HOME {workspace-path-hash} resolves to a hash of the manifest path For more information see the build cache documentation . build.rustflags Type: string or array of strings Default: none Environment: CARGO_BUILD_RUSTFLAGS or CARGO_ENCODED_RUSTFLAGS or RUSTFLAGS Extra command-line flags to pass to rustc . The value may be an array of strings or a space-separated string. There are four mutually exclusive sources of extra flags. They are checked in order, with the first one being used: CARGO_ENCODED_RUSTFLAGS environment variable. RUSTFLAGS environment variable. All matching target.<triple>.rustflags and target.<cfg>.rustflags config entries joined together. build.rustflags config value. Additional flags may also be passed with the cargo rustc command. If the --target flag (or build.target ) is used, then the flags will only be passed to the compiler for the target. Things being built for the host, such as build scripts or proc macros, will not receive the args. Without --target , the flags will be passed to all compiler invocations (including build scripts and proc macros) because dependencies are shared. If you have args that you do not want to pass to build scripts or proc macros and are building for the host, pass --target with the host triple . It is not recommended to pass in flags that Cargo itself usually manages. For example, the flags driven by profiles are best handled by setting the appropriate profile setting. Caution : Due to the low-level nature of passing flags directly to the compiler, this may cause a conflict with future versions of Cargo which may issue the same or similar flags on its own which may interfere with the flags you specify. This is an area where Cargo may not always be backwards compatible. build.rustdocflags Type: string or array of strings Default: none Environment: CARGO_BUILD_RUSTDOCFLAGS or CARGO_ENCODED_RUSTDOCFLAGS or RUSTDOCFLAGS Extra command-line flags to pass to rustdoc . The value may be an array of strings or a space-separated string. There are four mutually exclusive sources of extra flags. They are checked in order, with the first one being used: CARGO_ENCODED_RUSTDOCFLAGS environment variable. RUSTDOCFLAGS environment variable. All matching target.<triple>.rustdocflags config entries joined together. build.rustdocflags config value. Additional flags may also be passed with the cargo rustdoc command. Caution : Due to the low-level nature of passing flags directly to the compiler, this may cause a conflict with future versions of Cargo which may issue the same or similar flags on its own which may interfere with the flags you specify. This is an area where Cargo may not always be backwards compatible. build.incremental Type: bool Default: from profile Environment: CARGO_BUILD_INCREMENTAL or CARGO_INCREMENTAL Whether or not to perform incremental compilation . The default if not set is to use the value from the profile . Otherwise this overrides the setting of all profiles. The CARGO_INCREMENTAL environment variable can be set to 1 to force enable incremental compilation for all profiles, or 0 to disable it. This env var overrides the config setting. build.dep-info-basedir Type: string (path) Default: none Environment: CARGO_BUILD_DEP_INFO_BASEDIR Strips the given path prefix from dep info file paths. This config setting is intended to convert absolute paths to relative paths for tools that require relative paths. The setting itself is a config-relative path. So, for example, a value of "." would strip all paths starting with the parent directory of the .cargo directory. build.pipelining This option is deprecated and unused. Cargo always has pipelining enabled. [credential-alias] Type: string or array of strings Default: empty Environment: CARGO_CREDENTIAL_ALIAS_<name> The [credential-alias] table defines credential provider aliases. These aliases can be referenced as an element of the registry.global-credential-providers array, or as a credential provider for a specific registry under registries.<NAME>.credential-provider . If specified as a string, the value will be split on spaces into path and arguments. For example, to define an alias called my-alias : [credential-alias] my-alias = ["/usr/bin/cargo-credential-example", "--argument", "value", "--flag"] See Registry Authentication for more information. [doc] The [doc] table defines options for the cargo doc command. doc.browser Type: string or array of strings ( program path with args ) Default: BROWSER environment variable, or, if that is missing, opening the link in a system specific way This option sets the browser to be used by cargo doc , overriding the BROWSER environment variable when opening documentation with the --open option. [cargo-new] The [cargo-new] table defines defaults for the cargo new command. cargo-new.name This option is deprecated and unused. cargo-new.email This option is deprecated and unused. cargo-new.vcs Type: string Default: "git" or "none" Environment: CARGO_CARGO_NEW_VCS Specifies the source control system to use for initializing a new repository. Valid values are git , hg (for Mercurial), pijul , fossil or none to disable this behavior. Defaults to git , or none if already inside a VCS repository. Can be overridden with the --vcs CLI option. [env] The [env] section allows you to set additional environment variables for build scripts, rustc invocations, cargo run and cargo build . [env] OPENSSL_DIR = "/opt/openssl" By default, the variables specified will not override values that already exist in the environment. This behavior can be changed by setting the force flag. Setting the relative flag evaluates the value as a config-relative path that is relative to the parent directory of the .cargo directory that contains the config.toml file. The value of the environment variable will be the full absolute path. [env] TMPDIR = { value = "/home/tmp", force = true } OPENSSL_DIR = { value = "vendor/openssl", relative = true } [future-incompat-report] The [future-incompat-report] table controls setting for future incompat reporting future-incompat-report.frequency Type: string Default: "always" Environment: CARGO_FUTURE_INCOMPAT_REPORT_FREQUENCY Controls how often we display a notification to the terminal when a future incompat report is available. Possible values: always (default): Always display a notification when a command (e.g. cargo build ) produces a future incompat report never : Never display a notification [cache] The [cache] table defines settings for cargo’s caches. Global caches When running cargo commands, Cargo will automatically track which files you are using within the global cache. Periodically, Cargo will delete files that have not been used for some period of time. It will delete files that have to be downloaded from the network if they have not been used in 3 months. Files that can be generated without network access will be deleted if they have not been used in 1 month. The automatic deletion of files only occurs when running commands that are already doing a significant amount of work, such as all of the build commands ( cargo build , cargo test , cargo check , etc.), and cargo fetch . Automatic deletion is disabled if cargo is offline such as with --offline or --frozen to avoid deleting artifacts that may need to be used if you are offline for a long period of time. Note : This tracking is currently only implemented for the global cache in Cargo’s home directory. This includes registry indexes and source files downloaded from registries and git dependencies. Support for tracking build artifacts is not yet implemented, and tracked in cargo#13136 . Additionally, there is an unstable feature to support manually triggering cache cleaning, and to further customize the configuration options. See the Unstable chapter for more information. cache.auto-clean-frequency Type: string Default: "1 day" Environment: CARGO_CACHE_AUTO_CLEAN_FREQUENCY This option defines how often Cargo will automatically delete unused files in the global cache. This does not define how old the files must be, those thresholds are described above . It supports the following settings: "never" — Never deletes old files. "always" — Checks to delete old files every time Cargo runs. An integer followed by “seconds”, “minutes”, “hours”, “days”, “weeks”, or “months” — Checks to delete old files at most the given time frame. [http] The [http] table defines settings for HTTP behavior. This includes fetching crate dependencies and accessing remote git repositories. http.debug Type: boolean Default: false Environment: CARGO_HTTP_DEBUG If true , enables debugging of HTTP requests. The debug information can be seen by setting the CARGO_LOG=network=debug environment variable (or use network=trace for even more information). Be wary when posting logs from this output in a public location. The output may include headers with authentication tokens which you don’t want to leak! Be sure to review logs before posting them. http.proxy Type: string Default: none Environment: CARGO_HTTP_PROXY or HTTPS_PROXY or https_proxy or http_proxy Sets an HTTP and HTTPS proxy to use. The format is in libcurl format as in [protocol://]host[:port] . If not set, Cargo will also check the http.proxy setting in your global git configuration. If none of those are set, the HTTPS_PROXY or https_proxy environment variables set the proxy for HTTPS requests, and http_proxy sets it for HTTP requests. http.timeout Type: integer Default: 30 Environment: CARGO_HTTP_TIMEOUT or HTTP_TIMEOUT Sets the timeout for each HTTP request, in seconds. http.cainfo Type: string (path) Default: none Environment: CARGO_HTTP_CAINFO Path to a Certificate Authority (CA) bundle file, used to verify TLS certificates. If not specified, Cargo attempts to use the system certificates. http.proxy-cainfo Type: string (path) Default: falls back to http.cainfo if not set Environment: CARGO_HTTP_PROXY_CAINFO Path to a Certificate Authority (CA) bundle file, used to verify proxy TLS certificates. http.check-revoke Type: boolean Default: true (Windows) false (all others) Environment: CARGO_HTTP_CHECK_REVOKE This determines whether or not TLS certificate revocation checks should be performed. This only works on Windows. http.ssl-version Type: string or min/max table Default: none Environment: CARGO_HTTP_SSL_VERSION This sets the minimum TLS version to use. It takes a string, with one of the possible values of "default" , "tlsv1" , "tlsv1.0" , "tlsv1.1" , "tlsv1.2" , or "tlsv1.3" . This may alternatively take a table with two keys, min and max , which each take a string value of the same kind that specifies the minimum and maximum range of TLS versions to use. The default is a minimum version of "tlsv1.0" and a max of the newest version supported on your platform, typically "tlsv1.3" . http.low-speed-limit Type: integer Default: 10 Environment: CARGO_HTTP_LOW_SPEED_LIMIT This setting controls timeout behavior for slow connections. If the average transfer speed in bytes per second is below the given value for http.timeout seconds (default 30 seconds), then the connection is considered too slow and Cargo will abort and retry. http.multiplexing Type: boolean Default: true Environment: CARGO_HTTP_MULTIPLEXING When true , Cargo will attempt to use the HTTP2 protocol with multiplexing. This allows multiple requests to use the same connection, usually improving performance when fetching multiple files. If false , Cargo will use HTTP 1.1 without pipelining. http.user-agent Type: string Default: Cargo’s version Environment: CARGO_HTTP_USER_AGENT Specifies a custom user-agent header to use. The default if not specified is a string that includes Cargo’s version. [install] The [install] table defines defaults for the cargo install command. install.root Type: string (path) Default: Cargo’s home directory Environment: CARGO_INSTALL_ROOT Sets the path to the root directory for installing executables for cargo install . Executables go into a bin directory underneath the root. To track information of installed executables, some extra files, such as .crates.toml and .crates2.json , are also created under this root. The default if not specified is Cargo’s home directory (default .cargo in your home directory). Can be overridden with the --root command-line option. [net] The [net] table controls networking configuration. net.retry Type: integer Default: 3 Environment: CARGO_NET_RETRY Number of times to retry possibly spurious network errors. net.git-fetch-with-cli Type: boolean Default: false Environment: CARGO_NET_GIT_FETCH_WITH_CLI If this is true , then Cargo will use the git executable to fetch registry indexes and git dependencies. If false , then it uses a built-in git library. Setting this to true can be helpful if you have special authentication requirements that Cargo does not support. See Git Authentication for more information about setting up git authentication. net.offline Type: boolean Default: false Environment: CARGO_NET_OFFLINE If this is true , then Cargo will avoid accessing the network, and attempt to proceed with locally cached data. If false , Cargo will access the network as needed, and generate an error if it encounters a network error. Can be overridden with the --offline command-line option. net.ssh The [net.ssh] table contains settings for SSH connections. net.ssh.known-hosts Type: array of strings Default: see description Environment: not supported The known-hosts array contains a list of SSH host keys that should be accepted as valid when connecting to an SSH server (such as for SSH git dependencies). Each entry should be a string in a format similar to OpenSSH known_hosts files. Each string should start with one or more hostnames separated by commas, a space, the key type name, a space, and the base64-encoded key. For example: [net.ssh] known-hosts = [ "example.com ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFO4Q5T0UV0SQevair9PFwoxY9dl4pQl3u5phoqJH3cF" ] Cargo will attempt to load known hosts keys from common locations supported in OpenSSH, and will join those with any listed in a Cargo configuration file. If any matching entry has the correct key, the connection will be allowed. Cargo comes with the host keys for github.com built-in. If those ever change, you can add the new keys to the config or known_hosts file. See Git Authentication for more details. [patch] Just as you can override dependencies using [patch] in Cargo.toml , you can override them in the cargo configuration file to apply those patches to any affected build. The format is identical to the one used in Cargo.toml . Since .cargo/config.toml files are not usually checked into source control, you should prefer patching using Cargo.toml where possible to ensure that other developers can compile your crate in their own environments. Patching through cargo configuration files is generally only appropriate when the patch section is automatically generated by an external build tool. If a given dependency is patched both in a cargo configuration file and a Cargo.toml file, the patch in the configuration file is used. If multiple configuration files patch the same dependency, standard cargo configuration merging is used, which prefers the value defined closest to the current directory, with $HOME/.cargo/config.toml taking the lowest precedence. Relative path dependencies in such a [patch] section are resolved relative to the configuration file they appear in. [profile] The [profile] table can be used to globally change profile settings, and override settings specified in Cargo.toml . It has the same syntax and options as profiles specified in Cargo.toml . See the Profiles chapter for details about the options. [profile.<name>.build-override] Environment: CARGO_PROFILE_<name>_BUILD_OVERRIDE_<key> The build-override table overrides settings for build scripts, proc macros, and their dependencies. It has the same keys as a normal profile. See the overrides section for more details. [profile.<name>.package.<name>] Environment: not supported The package table overrides settings for specific packages. It has the same keys as a normal profile, minus the panic , lto , and rpath settings. See the overrides section for more details. profile.<name>.codegen-units Type: integer Default: See profile docs. Environment: CARGO_PROFILE_<name>_CODEGEN_UNITS See codegen-units . profile.<name>.debug Type: integer or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_DEBUG See debug . profile.<name>.split-debuginfo Type: string Default: See profile docs. Environment: CARGO_PROFILE_<name>_SPLIT_DEBUGINFO See split-debuginfo . profile.<name>.debug-assertions Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_DEBUG_ASSERTIONS See debug-assertions . profile.<name>.incremental Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_INCREMENTAL See incremental . profile.<name>.lto Type: string or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_LTO See lto . profile.<name>.overflow-checks Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_OVERFLOW_CHECKS See overflow-checks . profile.<name>.opt-level Type: integer or string Default: See profile docs. Environment: CARGO_PROFILE_<name>_OPT_LEVEL See opt-level . profile.<name>.panic Type: string Default: See profile docs. Environment: CARGO_PROFILE_<name>_PANIC See panic . profile.<name>.rpath Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_RPATH See rpath . profile.<name>.strip Type: string or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_STRIP See strip . [resolver] The [resolver] table overrides dependency resolution behavior for local development (e.g. excludes cargo install ). resolver.incompatible-rust-versions Type: string Default: See resolver docs Environment: CARGO_RESOLVER_INCOMPATIBLE_RUST_VERSIONS When resolving which version of a dependency to use, select how versions with incompatible package.rust-version s are treated. Values include: allow : treat rust-version -incompatible versions like any other version fallback : only consider rust-version -incompatible versions if no other version matched Can be overridden with --ignore-rust-version CLI option Setting the dependency’s version requirement higher than any version with a compatible rust-version Specifying the version to cargo update with --precise See the resolver chapter for more details. MSRV: allow is supported on any version fallback is respected as of 1.84 [registries] The [registries] table is used for specifying additional registries . It consists of a sub-table for each named registry. registries.<name>.index Type: string (url) Default: none Environment: CARGO_REGISTRIES_<name>_INDEX Specifies the URL of the index for the registry. registries.<name>.token Type: string Default: none Environment: CARGO_REGISTRIES_<name>_TOKEN Specifies the authentication token for the given registry. This value should only appear in the credentials file. This is used for registry commands like cargo publish that require authentication. Can be overridden with the --token command-line option. registries.<name>.credential-provider Type: string or array of path and arguments Default: none Environment: CARGO_REGISTRIES_<name>_CREDENTIAL_PROVIDER Specifies the credential provider for the given registry. If not set, the providers in registry.global-credential-providers will be used. If specified as a string, path and arguments will be split on spaces. For paths or arguments that contain spaces, use an array. If the value exists in the [credential-alias] table, the alias will be used. See Registry Authentication for more information. registries.crates-io.protocol Type: string Default: "sparse" Environment: CARGO_REGISTRIES_CRATES_IO_PROTOCOL Specifies the protocol used to access crates.io. Allowed values are git or sparse . git causes Cargo to clone the entire index of all packages ever published to crates.io from https://github.com/rust-lang/crates.io-index/ . This can have performance implications due to the size of the index. sparse is a newer protocol which uses HTTPS to download only what is necessary from https://index.crates.io/ . This can result in a significant performance improvement for resolving new dependencies in most situations. More information about registry protocols may be found in the Registries chapter . [registry] The [registry] table controls the default registry used when one is not specified. registry.index This value is no longer accepted and should not be used. registry.default Type: string Default: "crates-io" Environment: CARGO_REGISTRY_DEFAULT The name of the registry (from the registries table ) to use by default for registry commands like cargo publish . Can be overridden with the --registry command-line option. registry.credential-provider Type: string or array of path and arguments Default: none Environment: CARGO_REGISTRY_CREDENTIAL_PROVIDER Specifies the credential provider for crates.io . If not set, the providers in registry.global-credential-providers will be used. If specified as a string, path and arguments will be split on spaces. For paths or arguments that contain spaces, use an array. If the value exists in the [credential-alias] table, the alias will be used. See Registry Authentication for more information. registry.token Type: string Default: none Environment: CARGO_REGISTRY_TOKEN Specifies the authentication token for crates.io . This value should only appear in the credentials file. This is used for registry commands like cargo publish that require authentication. Can be overridden with the --token command-line option. registry.global-credential-providers Type: array Default: ["cargo:token"] Environment: CARGO_REGISTRY_GLOBAL_CREDENTIAL_PROVIDERS Specifies the list of global credential providers. If credential provider is not set for a specific registry using registries.<name>.credential-provider , Cargo will use the credential providers in this list. Providers toward the end of the list have precedence. Path and arguments are split on spaces. If the path or arguments contains spaces, the credential provider should be defined in the [credential-alias] table and referenced here by its alias. See Registry Authentication for more information. [source] The [source] table defines the registry sources available. See Source Replacement for more information. It consists of a sub-table for each named source. A source should only define one kind (directory, registry, local-registry, or git). source.<name>.replace-with Type: string Default: none Environment: not supported If set, replace this source with the given named source or named registry. source.<name>.directory Type: string (path) Default: none Environment: not supported Sets the path to a directory to use as a directory source. source.<name>.registry Type: string (url) Default: none Environment: not supported Sets the URL to use for a registry source. source.<name>.local-registry Type: string (path) Default: none Environment: not supported Sets the path to a directory to use as a local registry source. source.<name>.git Type: string (url) Default: none Environment: not supported Sets the URL to use for a git repository source. source.<name>.branch Type: string Default: none Environment: not supported Sets the branch name to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. source.<name>.tag Type: string Default: none Environment: not supported Sets the tag name to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. source.<name>.rev Type: string Default: none Environment: not supported Sets the revision to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. [target] The [target] table is used for specifying settings for specific platform targets. It consists of a sub-table which is either a platform triple or a cfg() expression . The given values will be used if the target platform matches either the <triple> value or the <cfg> expression. [target.thumbv7m-none-eabi] linker = "arm-none-eabi-gcc" runner = "my-emulator" rustflags = ["…", "…"] [target.'cfg(all(target_arch = "arm", target_os = "none"))'] runner = "my-arm-wrapper" rustflags = ["…", "…"] cfg values come from those built-in to the compiler (run rustc --print=cfg to view) and extra --cfg flags passed to rustc (such as those defined in RUSTFLAGS ). Do not try to match on debug_assertions , test , Cargo features like feature="foo" , or values set by build scripts . If using a target spec JSON file, the <triple> value is the filename stem. For example --target foo/bar.json would match [target.bar] . target.<triple>.ar This option is deprecated and unused. target.<triple>.linker Type: string (program path) Default: none Environment: CARGO_TARGET_<triple>_LINKER Specifies the linker which is passed to rustc (via -C linker ) when the <triple> is being compiled for. By default, the linker is not overridden. target.<cfg>.linker This is similar to the target linker , but using a cfg() expression . If both a <triple> and <cfg> runner match, the <triple> will take precedence. It is an error if more than one <cfg> runner matches the current target. target.<triple>.runner Type: string or array of strings ( program path with args ) Default: none Environment: CARGO_TARGET_<triple>_RUNNER If a runner is provided, executables for the target <triple> will be executed by invoking the specified runner with the actual executable passed as an argument. This applies to cargo run , cargo test and cargo bench commands. By default, compiled executables are executed directly. target.<cfg>.runner This is similar to the target runner , but using a cfg() expression . If both a <triple> and <cfg> runner match, the <triple> will take precedence. It is an error if more than one <cfg> runner matches the current target. target.<triple>.rustflags Type: string or array of strings Default: none Environment: CARGO_TARGET_<triple>_RUSTFLAGS Passes a set of custom flags to the compiler for this <triple> . The value may be an array of strings or a space-separated string. See build.rustflags for more details on the different ways to specific extra flags. target.<cfg>.rustflags This is similar to the target rustflags , but using a cfg() expression . If several <cfg> and <triple> entries match the current target, the flags are joined together. target.<triple>.rustdocflags Type: string or array of strings Default: none Environment: CARGO_TARGET_<triple>_RUSTDOCFLAGS Passes a set of custom flags to the compiler for this <triple> . The value may be an array of strings or a space-separated string. See build.rustdocflags for more details on the different ways to specific extra flags. target.<triple>.<links> The links sub-table provides a way to override a build script . When specified, the build script for the given links library will not be run, and the given values will be used instead. [target.x86_64-unknown-linux-gnu.foo] rustc-link-lib = ["foo"] rustc-link-search = ["/path/to/foo"] rustc-flags = "-L /some/path" rustc-cfg = ['key="value"'] rustc-env = {key = "value"} rustc-cdylib-link-arg = ["…"] metadata_key1 = "value" metadata_key2 = "value" [term] The [term] table controls terminal output and interaction. term.quiet Type: boolean Default: false Environment: CARGO_TERM_QUIET Controls whether or not log messages are displayed by Cargo. Specifying the --quiet flag will override and force quiet output. Specifying the --verbose flag will override and disable quiet output. term.verbose Type: boolean Default: false Environment: CARGO_TERM_VERBOSE Controls whether or not extra detailed messages are displayed by Cargo. Specifying the --quiet flag will override and disable verbose output. Specifying the --verbose flag will override and force verbose output. term.color Type: string Default: "auto" Environment: CARGO_TERM_COLOR Controls whether or not colored output is used in the terminal. Possible values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. Can be overridden with the --color command-line option. term.hyperlinks Type: bool Default: auto-detect Environment: CARGO_TERM_HYPERLINKS Controls whether or not hyperlinks are used in the terminal. term.unicode Type: bool Default: auto-detect Environment: CARGO_TERM_UNICODE Control whether output can be rendered using non-ASCII unicode characters. term.progress.when Type: string Default: "auto" Environment: CARGO_TERM_PROGRESS_WHEN Controls whether or not progress bar is shown in the terminal. Possible values: auto (default): Intelligently guess whether to show progress bar. always : Always show progress bar. never : Never show progress bar. term.progress.width Type: integer Default: none Environment: CARGO_TERM_PROGRESS_WIDTH Sets the width for progress bar. term.progress.term-integration Type: bool Default: auto-detect Environment: CARGO_TERM_PROGRESS_TERM_INTEGRATION Report progress to the terminal emulator for display in places like the task bar.
2026-01-13T09:29:19
https://www.linkedin.com/products/categories/workflow-management-software
Best Workflow Management Software | Products | LinkedIn Skip to main content LinkedIn Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Clear text Used by Used by Project Manager (27) Operations Manager (22) Chief Information Officer (15) Software Engineer (15) Information Technology Manager (13) See all products Find top products in Workflow Management Software category Software used to generate and oversee standard sequences of tasks in a business process. - Customize standard workflow templates with low-code or code-free interfaces - Define dependencies and track progress with centralized dashboards - Automate repetitive tasks and receive due date reminders 432 results n8n Workflow Management Software by n8n n8n is a flexible workflow automation tool. With a fair-code distribution model, n8n will always have visible source code, be available to self-host, and allow you to add your own custom functions, logic and integrations. n8n's modular, node-based approach makes it highly versatile, enabling you to connect systems and build powerful AI-infused automations. View product Automation Engine Workflow Management Software by Esko Automation Engine is designed to maximize automation within your prepress department and beyond – supplying right first time proofs to your customers and high-quality print-ready files for your digital, flexo, offset, and gravure presses. Automation Engine enables you to standardize your processes, remove time wasted activities from your prepress, and minimize pre-production errors. View product Pipefy Workflow Management Software by Pipefy AI Agents work for you. Pipefy is on a mission to redefine workflow automation. With Pipefy, you can design and execute seamless processes across departments like Finance, HR, IT, and Customer Support. Integrate tools, automate repetitive tasks, and rely on AI-driven insights to optimize every step—so your team can focus on what truly matters: driving results and creating value. View product Jotform Workflows Workflow Management Software by Jotform Jotform Workflows streamlines processes within your company or organization. Build automated flows for HR requests, IT tickets, job applications, and other processes. Design workflows that include multiple forms, integrations, conditional logic, payment processing, and more. Create, customize, and assign workflow tasks to teammates, and white-label workflows to match company guidelines. Track every step in your business processes from any device — so no task gets left behind. View product TeamViewer Frontline Workflow Management Software by TeamViewer Empower your Frontline Workforce with Augmented Reality! Digitalize and streamline processes for frontline employees in desk-free workspaces with AR-guided workflows – seamlessly integrated with wearables and mobile devices. With TeamViewer Frontline you can increase your company’s productivity, efficiency, and quality along the entire value chain. View product Find products trusted by professionals in your network See which products are used by connections in your network and those that share similar job titles Sign in to view full insights Value Stream Management Workflow Management Software by Planview Planview’s value stream management solution enables CIOs and Transformation Leaders to align technology and the business with the right metrics and automated workflows to enable more effective enterprise digital transformation. View product ApprovalMax Workflow Management Software by ApprovalMax ApprovalMax is an award-winning B2B software platform used by 15,000+ businesses around the world to build robust financial controls. It is a powerful all-in-one solution that integrates with Xero, QuickBooks Online, and Oracle NetSuite to unlock powerful efficiencies for accounts payable and accounts receivable functions in any finance team. Key features and capabilities: – Multi-stage and multi-level approval automation for accounts payable and accounts receivable – Unlimited users, fully digital approval process – Automatic audit trails for all approvals – Mobile app approvals for decision makers and instant notifications – Bill to purchase order matching – Budget checking capability – Auto-approvals – Direct integration with Dext Prepare, Airwallex, Cin7 Core, and Slack – Compatibility with all data capture tools New add ons as at 12 June 2024: - ApprovalMax Capture - ApprovalMax Pay View product Process Street Workflow Management Software by Process Street Process Street is the Compliance Operations Platform that turns critical procedures into automated, AI-enforced workflows. Built for teams in high-stakes environments, it ensures every policy is followed, every task is auditable, and every risk is caught before it becomes a problem. Start with employee onboarding or client implementation. Then expand into audit prep, policy management, vendor approvals, or any workflow where compliance matters. Thousands of teams use Process Street to streamline operations, enforce standards, and scale with confidence. Customers include Salesforce, Colliers, Cargill, and Hartford Healthcare. View product Formstack Workflow Management Software by Intellistack Empower your organization to digitize and automate workflows with powerful workplace productivity tools. Explore our workflow automation solutions today. View product awork Workflow Management Software by awork awork is the #1 project platform for agencies. ✨ Generic project tools fall short. They don’t fit the fast-paced agency workflow and leave valuable potential untapped. That’s exactly where awork comes in. awork is the only project platform built specifically for agency workflows: Managing complex client projects across internal teams and external collaborators – planning, scheduling, collaboration, and time tracking all in one place. ✔️ Overview: Visibility across teams and deliverables. ✔️ Reliable planning: Accurate plans, better utilisation. ✔️ Collaboration: Add clients & freelancers for free. ✔️ Time tracking: Increased billability, less chasing. ✔️ AI assistance: Smart project copilot. ✔️ Onboarding: Within 6 weeks, success guaranteed. ✔️ GDPR-compliant and ISO 27001-certified Join over 10,000 teams working better and happier with awork. 💜 View product See more How it works Explore Discover the best product for your need from a growing catalog of 25,000 products and categories trusted by LinkedIn professionals Learn Evaluate new tools, explore trending products in your industry and see who in your network is skilled in the product Grow Join communities of product users to learn best practices, celebrate your progress and accelerate your career LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English Language
2026-01-13T09:29:19
https://git-scm.com/book/gr/v2/%ce%a4%ce%b1-%ce%b8%ce%b5%ce%bc%ce%b5%ce%bb%ce%b9%cf%8e%ce%b4%ce%b7-%cf%83%cf%84%ce%bf%ce%b9%cf%87%ce%b5%ce%af%ce%b1-%cf%84%ce%bf%cf%85-Git-%ce%95%cf%84%ce%b9%ce%ba%ce%ad%cf%84%ce%b5%cf%82
Git - Ετικέτες About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Ξεκινώντας με το Git 1.1 Σχετικά με τον έλεγχο εκδόσεων 1.2 Σύντομο ιστορικό του Git 1.3 Τι είναι το Git; 1.4 Η γραμμή εντολών 1.5 Εγκατάσταση του Git 1.6 Ρύθμιση του Git για πρώτη φορά 1.7 Χρησιμοποιώντας τη βοήθεια 1.8 Ανακεφαλαίωση 2. Τα θεμελιώδη στοιχεία του Git 2.1 Απόκτηση αποθετηρίου Git 2.2 Καταγραφή αλλαγών στο αποθετήριο 2.3 Χρησιμοποιώντας το ιστορικό υποβολών 2.4 Αναιρέσεις (undoing) 2.5 Δουλεύοντας με απομακρυσμένα αποθετήρια 2.6 Ετικέτες 2.7 Συντομεύεσεις στο Git 2.8 Ανακεφαλαίωση 3. Διακλαδώσεις στο Git 3.1 Οι κλάδοι με λίγα λόγια 3.2 Βασικές έννοιες διακλαδώσεων και συγχωνεύσεων 3.3 Διαχείριση κλάδων 3.4 Ροές εργασίας με διακλαδώσεις 3.5 Απομακρυσμένοι κλάδοι 3.6 Αλλαγή βάσης (rebasing) 3.7 Ανακεφαλαίωση 4. Το Git στον διακομιστή 4.1 Τα πρωτόκολλα 4.2 Εγκατάσταση του Git σε διακομιστή 4.3 Δημιουργία δημόσιου κλειδιού SSH 4.4 Στήσιμο του διακομιστή 4.5 Δαίμονες του Git 4.6 Έξυπνο HTTP 4.7 GitWeb 4.8 GitLab 4.9 Επιλογές φιλοξενίας από τρίτους 4.10 Ανακεφαλαίωση 5. Κατανεμημένο Git 5.1 Κατανεμημένες ροές εργασίας 5.2 Συνεισφέροντας σε ένα έργο 5.3 Συντήρηση ενός έργου 5.4 Ανακεφαλαίωση 6. GitHub 6.1 Δημιουργία λογαριασμού και ρύθμισή του 6.2 Συνεισφορά σε έργο 6.3 Συντήρηση ενός έργου 6.4 Διαχείριση οργανισμώνν 6.5 Συγγραφή script στο GitHub 6.6 Ανακεφαλαίωση 7. Εργαλεία του Git 7.1 Επιλογή αναθεώρησης 7.2 Διαδραστική εργασία με το στάδιο καταχώρισης 7.3 Αποθέματα και Καθαρισμός 7.4 Υπογραφή της δουλειάς μας 7.5 Αναζήτηση 7.6 Η ιστορία ξαναγράφεται 7.7 Απομυθοποίηση της reset 7.8 Προχωρημένη Συγχώνευση 7.9 Rerere 7.10 Αποσφαλμάτωση με το Git 7.11 Υπομονάδες 7.12 Δεμάτιασμα δεδομένων 7.13 Replace 7.14 Αποθήκευση διαπιστευτηρίων 7.15 Ανακεφαλαίωση 8. Εξατομίκευση του Git 8.1 Διαμόρφωση Git 8.2 Γνωρίσματα του Git 8.3 Τα άγκιστρα του Git 8.4 Ένα παράδειγμα επιβολής πολιτικής από το Git 8.5 Ανακεφαλαίωση 9. Το Git και άλλα συστήματα 9.1 Το Git ως πελάτης 9.2 Μετανάστευση στο Git 9.3 Ανακεφαλαίωση 10. Εσωτερική λειτουργία του Git 10.1 Διοχετεύσεις και πορσελάνες 10.2 Αντικείμενα του Git 10.3 Αναφορές του Git 10.4 Πακετάρισμα αρχείων 10.5 Τα refspec 10.6 Πρωτόκολλα μεταφοράς 10.7 Διατήρηση και ανάκτηση δεδομένων 10.8 Μεταβλητές περιβάλλοντος 10.9 Ανακεφαλαίωση A1. Appendix A: Το Git σε άλλα περιβάλλοντα A1.1 Γραφικές διεπαφές A1.2 Το Git στο Visual Studio A1.3 Git στο Visual Studio Code A1.4 Git στο IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git στο Sublime Text A1.6 Το Git στο Bash A1.7 Το Git στο Zsh A1.8 Το Git στο Powershell A1.9 Ανακεφαλαίωση A2. Appendix B: Ενσωμάτωση του Git στις εφαρμογές μας A2.1 Γραμμή εντολών Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Εντολές Git A3.1 Ρύθμιση και διαμόρφωση A3.2 Λήψη και δημιουργία έργων A3.3 Βασική λήψη στιγμιοτύπων A3.4 Διακλάδωση και συγχώνευση A3.5 Κοινή χρήση και ενημέρωση έργων A3.6 Επιθεώρηση και σύγκριση A3.7 Αποσφαλμάτωση A3.8 Επιθέματα A3.9 Ηλεκτρονικό ταχυδρομείο A3.10 Εξωτερικά Συστήματα A3.11 Διοίκηση A3.12 Εντολές διοχέτευσης 2nd Edition 2.6 Τα θεμελιώδη στοιχεία του Git - Ετικέτες Ετικέτες Το Git, όπως και τα περισσότερα VCS, δίνει τη δυνατότητα να βάζουμε ετικέτες (tags) σε συγκεκριμένα σημεία του ιστορικού ενός έργου. Η λειτουργικότητα αυτή χρησιμοποείται συνήθως για σημειωθούν συγκεκριμένες εκδόσεις (π.χ. έκδοση 1.0 , 2.0 κ.ο.κ.). Σε αυτή την ενότητα θα μάθουμε πώς να βλέπουμε ποιες ετικέτες έχει ένα έργο, πώς να δημιουργούμε ετικέτες και ποια είδη ετικετών υπάρχουν. Παράθεση ετικετών Για να δούμε τις ετικέτες ενός έργου, η εντολή στο Git είναι αρκετά απλή. Απλά πληκτρολογούμε git tag (και προαιρετικά -l ή --list ): $ git tag v1.0 v2.0 Η εντολή παραθέτει τις ετικέτες σε αλφαβητική σειρά, αν και η σειρά αυτή δεν έχει κάποια ιδιαίτερη σημασία. Μπορούμε επίσης να αναζητήσουμε ετικέτες που ακολουθούν κάποιο μοτίβο. Για παράδειγμα, το αποθετήριο με τον πηγαίο κώδικα του Git περιέχει περισσότερες από 500 ετικέτες. Αν μας ενδιφέρει να δούμε μόνο αυτές που έχουν σχέση με την έκδοση 1.8.5, τότε μπορούμε να εκτελέσουμε: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Note Παράθεση ετικετών με χαρακτήρες υποκατάστασης απαιτεί την επιλογή -l ή --list Αν θέλετε μόνο την πλήρη λίστα των ετικετών, τότε αν εκτελέσετε την εντολή git tag υποθέτει ότι θέλετε μια λίστα και την εμφανίζει· η χρήση του -l ή του --list σε αυτή την περίπτωση είναι προαιρετική. Αν όμως δώσετε και ένα μοτίβο με χαρακτήρες υποκατάσταση (wildcards), τότε η χρήση ενός από τα -l ή --list είναι υποχρεωτική. Δημιουργία ετικετών Το Git χρησιμοποιεί δύο κατηγορίες ετικετών, τις απλές ( lightweight ) και τις ετικέτες με επισημείωση ( annotated ). Μια απλή ετικέτα μοιάζει πολύ με έναν κλάδο που δεν αλλάζει· είναι απλά ένας δείκτης σε μια συγκεκριμένη υποβολή. Οι ετικέτες με επισημειώσεις από την άλλη, αποθηκεύονται στη βάση δεδομένων του Git ως πλήρη αντικείμενα. Για καθεμία επισημειωμένη ετικέτα: υπολογίζεται το άθροισμα ελέγχου (checksum) της· περιέχει το όνομα, και τη διεύθυνση e-mail αυτού που βάζει την ετικέτα και την ημερομηνία· έχει ένα μήνυμα· μπορεί να υπογραφεί και να επαληθευθεί με τον GNU Privacy Guard (GPG). Γενικά συνιστάται να δημιουργούμε ετικέτες με επισημειώσεις (annotations) έτσι ώστε να έχουμε όλες αυτές τις πληροφορίες, αλλά αν για κάποιο λόγο θέλουμε μία προσωρινή ετικέτα ή μια ετικέτα χωρίς περαιτέρω πληροφορίες, μπορούμε να χρησιμοποιήσουμε τις απλές ετικέτες. Επισημασμένες ετικέτες Η δημιουργία επισημασμένων ετικετών είναι απλή. Ο ευκολότερος τρόπος είναι να χρησιμοποιήσουμε την επιλογή -a όταν εκτελούμε την εντολή tag : $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 Η επιλογή -m σημαίνει ότι ακολουθεί το μήνυμα της ετικέτας, το οποίο και αποθηκεύεται μαζί με αυτή. Αν δεν προσδιορίσουμε μήνυμα σε μια καινούρια επισημειωμένη ετικέτα, τότε το Git θα εκκινήσει τον επεξεργαστή κειμένου μας, ώστε να γράψουμε το μήνυμα εκεί. Χρησιμοποιώντας την εντολή git show μπορούμε να δούμε τις πληροφορίες που περιέχει μια ετικέτα: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Η εντολή αυτή μας δείχνει πληροφορίες για τον χρήστη που έφτιαξε την ετικέτα, την ημερομηνία που τοποθετήθηκε η ετικέτα στην υποβολή (commit) και το μήνυμά της. Απλές ετικέτες Ένας άλλος τρόπος για να βάζουμε ετικέτες στις υποβολές είναι οι απλές (lightweight) ετικέτες. Μία τέτοια ετικέτα δεν είναι τίποτα άλλο από το άθροισμα ελέγχου της υποβολής μας, που αποθηκεύεται σε ένα αρχείο· δεν διατηρείται καμία άλλη πληροφορία. Για να δημιουργήσουμε μια απλή ετικέτα, δεν θα πρέπει να χρησιμοποιήσουμε τις επιλογές -a , -s ή -m : $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Αν τώρα εκτελέσουμε git show για τη συγκεκριμένη ετικέτα, δεν θα δούμε τις επιπλέον πληροφορίες που έχουν οι επισημασμένες ετικέτες. Θα δούμε μόνο την υποβολή: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Προσάρτηση ετικέτας εκ των υστέρων Μπορούμε επίσης να προσαρτήσουμε ετικέτες σε παλαιότερες υποβολές. Ας υποθέσουμε ότι το ιστορικό υποβολών μας είναι κάπως έτσι: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme Ας υποθέσουμε τώρα ότι ξεχάσαμε να βάλουμε ετικέτα στο έργο μας στην έκδοση v1.2 που ήταν η υποβολή με το μήνυμα “updated rakefile”. Μπορούμε να προσαρτήσουμε την ετικέτα αργότερα. Για να το κάνουμε αυτό, θα πρέπει να προσδιορίσουμε το άθροισμα ελέγχου της υποβολής μας (ή ένα μέρος του) (commit checksum) στο τέλος της εντολής: $ git tag -a v1.2 9fceb02 Μπορούμε να δούμε ότι έχουμε προσθέσει την ετικέτα στην υποβολή: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... Κοινοποίηση ετικετών Εξ ορισμού, η εντολή git push δεν μεταφέρει ετικέτες στους απομακρυσμένους διακομιστές. Θα πρέπει να ορίσουμε ρητά ότι θέλουμε να ωθήσουμε τις ετικέτες στον διακομιστή, αφού προηγουμένως τις έχουμε δημιουργήσει. Η διαδικασία είναι παρόμοια με την κοινοποίηση απομακρυσμένων κλάδων — εκτελούμε την εντολή git push origin <tagname> . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Αν έχουμε πολλές ετικέτες που θέλουμε να ωθήσουμε με τη μία, μπορούμε επίσης να χρησιμοποιήσουμε την επιλογή --tags στην εντολή git push . Με τον τρόπο αυτό θα μεταφέρουμε στον διακομιστή όλες τις ετικέτες που δεν είναι ήδη εκεί. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Πλέον, όταν κάποιος κλωνοποιήσει ή ελκύσει (pulls) δεδομένα από το αποθετήριό μας, θα λάβει μαζί και όλες τις ετικέτες μας. Note Η εντολή git push ωθεί και τα δύο είδη ετικετών Η εντολή git push <remote> --tags θα ωθήσει τόσο τις απλές όσο και τις επισημασμένες ετικέτες. Προς το παρόν δεν υπάρχει κάποιος τρόπος να ωθήσετε μόνο τις απλές ετικέτες, αλλά αν εκτελέσετε git push <remote> --follow-tags μόνο οι επισημασμένες ετικέτες θα ωθηθούν στον απομακρυσμένο διακομιστή. Διαγραφή ετικετών Για να διαγράψουμε μια ετικέτα στο τοπικό μας αποθετήριο, μπορούμε να εκτελέσουμε git tag -d <tagname> . Για παράδειγμα, μπορούμε να διαγράψουμε την απλή ετικέτα που δημιουργήσαμε παραπάνω ως εξής: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Σημειώνουμε ότι η εντολή αυτή δεν διαγράφει την ετικέτα από τους απομακρυσμένους διακομιστές. Υπάρχουν δύο παραλλαγές για τη διαγραφή ετικετών από έναν απομακρυσμένο διακομιστή. Η πρώτη είναι η εντολή git push <remote> :refs/tags/<tagname> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw Αυτό ερμηνεύεται ως εξής: η ετικέτα της οποίας το ονομα είναι ο χαρακτήρας null πριν από το : ωθείται στο απομακρυσμένο όνομα ετικέτας και ουσιαστικά το διαγράφει. Η δεύτερη (και πιο διαισθητική) είναι να διαγράψουμε την ετικέτα ως εξής: $ git push origin --delete <tagname> Ενημέρωση (check out) ετικετών Για να δούμε τις εκδόσεις των αρχείων μας στα οποία δείχνει μία ετικέτα, μπορούμε να εκτελέσουμε git checkout για αυτή την ετικέτα, αλλά αυτό θέτει το αποθετήριό μας σε κατάσταση “detached HEAD”, κάτι που έχει κάποιες παρενέργειες: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image Στην κατάσταση “detached HEAD”, αν κάνουμε αλλαγές και δημιουργήσουμε μια υποβολή (commit), η ετικέτα θα παραμείνει η ίδια, αλλά η νέα υποβολή δεν θα ανήκει σε κανέναν κλάδο και δεν θα είναι προσβάσιμη με κανέναν τρόπο εκτός κι αν χρησιμοποιήσουμε το ακριβές hash της υποβολής. Επιπλέον, αν χρειάζεται να κάνουμε αλλαγές — ας πούμε ότι διορθώνουμε κάποιο σφάλμα κάποιας παλιότερης έκδοσης, για παράδειγμα — καλό θα είναι να δημιουργήσουμε έναν κλάδο: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Αν βέβαια εκτελέσουμε την παραπάνω εντολή και πραγματοποιήσουμε μια υποβολή, ο κλάδος version2 θα είναι λίγο διαφορετικός από την ετικέτα v2.0.0 διότι θα έχει προχωρήσει με τις νέες αλλαγές· συνεπώς προσοχή. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-De-protocollen
Git - De protocollen About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.1 Git op de server - De protocollen Je zou nu de alledaagse taken waarvoor je Git zult gebruiken moeten kunnen uitvoeren. Echter, om enige vorm van samenwerking te hebben in Git is een remote Git repository nodig. Technisch gezien kan je wijzigingen pushen naar en pullen van de persoonlijk repositories van anderen, maar dat wordt afgeraden omdat je vrij gemakkelijk het werk waar anderen mee bezig zijn in de war kunt schoppen als je niet oppast. Daarnaast wil je dat je medewerkers de repository kunnen bereiken, zelfs als jouw computer van het netwerk is; het hebben van een meer betrouwbare gezamenlijke repository is vaak nuttig. De voorkeursmethode om met iemand samen te werken is daarom een tussenliggende repository in te richten waar alle partijen toegang tot hebben en om daar naartoe te pushen en vandaan te pullen. Een Git server draaien is relatief eenvoudig. Als eerste kies je welke protocollen je de server wilt laten ondersteunen. In het eerste gedeelte van dit hoofdstuk zullen we de beschikbare protocollen bespreken met de voor- en nadelen van elk. De daarop volgende paragrafen zullen we een aantal veel voorkomende opstellingen bespreken die van die protocollen gebruik maken en hoe je je server ermee kunt opzetten. Als laatste laten we een paar servers van derden zien, als je het niet erg vindt om je code op de server van een ander te zetten en niet het gedoe wilt hebben van het opzetten en onderhouden van je eigen server. Als je niet van plan bent om je eigen server te draaien, dan kun je de direct naar de laatste paragraaf van dit hoofdstuk gaan om wat mogelijkheden van online accounts te zien en dan door gaan naar het volgende hoofdstuk, waar we diverse zaken bespreken die komen kijken bij het werken met een gedistribueerde versiebeheer omgeving. Een remote repository is over het algemeen een bare repository (kale repository): een Git repository dat geen werkdirectory heeft. Omdat de repository alleen gebruikt wordt als een samenwerkingspunt, is er geen reden om een snapshot op de schijf te hebben; het is alleen de Git data. Een kale repository is eenvoudigweg de inhoud van de .git directory in je project, en niets meer. De protocollen Git kan vier verschillende netwerk protocollen gebruiken om data uit te wisselen: Lokaal, HTTP, Beveiligde Shell (Secure Shell, SSH) en Git. Hier bespreken we wat deze zijn en in welke omstandigheden je ze zou willen gebruiken (of juist niet). Lokaal protocol Het simpelste is het Lokale protocol , waarbij de remote repository in een andere directory op de schijf van dezelfde host staat. Deze opzet wordt vaak gebruikt als iedereen in het team toegang heeft op een gedeeld bestandssyteem zoals een NFS mount, of in het minder waarschijnlijke geval dat iedereen op dezelfde computer werkt. Het laatste zou niet ideaal zijn, want dan zouden alle repositories op dezelfde computer staan, wat de kans op een catastrofaal verlies van gegevens veel groter maakt. Als je een gedeeld bestandssyteem hebt, dan kun je clonen, pushen en pullen van een op een lokale bestands-gebaseerde repository. Om een dergelijk repository te clonen, of om er een als een remote aan een bestaand project toe te voegen, moet je het pad naar het repository als URL gebruiken. Bijvoorbeeld, om een lokaal repository te clonen, kun je zoiets als het volgende uitvoeren: $ git clone /srv/git/project.git Of je kunt dit doen: $ git clone file:///srv/git/project.git Git werkt iets anders als je expliciet file:// aan het begin van de URL zet. Als je alleen het pad specificeert, probeert Git hardlinks te gebruiken naar de bestanden die het nodig heeft of direct de bestanden te kopieren. Als je file:// specificeert, dan start Git de processen die het normaal gesproken gebruikt om data uit te wisselen over een netwerk, wat over het algemeen een veel minder efficiënte methode is om gegevens over te dragen. De belangrijkste reden om file:// wel te specificeren is als je een schone kopie van de repository wilt met de systeemvreemde referenties of objecten eruit gelaten — over het algemeen na een import uit een ander versiebeheer systeem of iets dergelijks (zie Git Binnenwerk voor onderhoudstaken). We zullen het normale pad hier gebruiken, omdat het bijna altijd sneller is om het zo te doen. Om een lokale repository aan een bestaand Git project toe te voegen, kun je iets als het volgende uitvoeren: $ git remote add local_proj /srv/git/project.git Daarna kan je op gelijke wijze pushen naar, en pullen van die remote via je nieuwe reomte genaamd local_projc zoals je over een netwerk zou doen. De voordelen De voordelen van bestands-gebaseerde repositories zijn dat ze eenvoudig zijn en ze maken gebruik van bestaande bestandspermissies en netwerk toegang. Als je al een gedeeld bestandssysteem hebt waar het hele team al toegang toe heeft, dan is een repository opzetten heel gemakkelijk. Je stopt de kale repository ergens waar iedereen gedeelde toegang tot heeft, en stelt de lees- en schrijfrechten in zoals je dat bij iedere andere gedeelde directory zou doen. We zullen de manier om een kopie van een kale repositiory te exporteren voor dit doel bespreken in Git op een server krijgen . Dit is ook een prettige optie om snel werk uit een repository van iemand anders te pakken. Als jij en een collega aan hetzelfde project werken, en hij wil dat je iets bekijkt, dan is het uitvoeren van een commando zoals git pull /home/john/project vaak makkelijker dan wanneer hij naar een remote server moet pushen, en jij er van moet fetchen. De nadelen Eén van de nadelen van deze methode is dat gedeelde toegang over het algemeen moeilijker op te zetten en te bereiken is vanaf meerdere lokaties dan simpele netwerk toegang. Als je wilt pushen van je laptop als je thuis bent, dan moet je de remote schijf aankoppelen, wat moeilijk en langzaam kan zijn in vergelijking met met netwerk gebaseerde toegang. Het is ook belangrijk om te vermelden dat het niet altijd de snelste optie is als je een gedeeld koppelpunt (mount) of iets dergelijks gebruikt. Een lokale repository is alleen snel als je snelle toegang tot de data hebt. Een repository op NFS is vaak langzamer dan een repository via SSH op dezelfde server omdat dit laatste Git in staat stelt op lokale schijven te werken op elk van de betrokken systemen. De HTTP protocollen Git kan op twee verschillende manieren via HTTP communiceren. Voor Git versie 1.6.6 was er maar een manier waarop dit kon en dat was erg basaal en over het algemeen kon je alleen lezen. In versie 1.6.6 en later, is een slimmere protocol geïntroduceerd waardoor Git in staat is gesteld om de gegevens-uitwisseling iets slimmer aan te pakken, op een manier die lijkt op de SSH aanpak. In de laatste paar jaar is dit nieuwe HTTP protocol erg populair geworden omdat het eenvoudiger is voor de gebruiker en slimmer in de manier waarop het communiceert. Aan deze nieuwere versie wordt vaak gerefereerd als het slimme HTTP protocol en het oude als het domme HTTP. We zullen eerst het nieuwere slimme HTTP protocol behandelen. Slimme HTTP Het slimme HTTP protocol werkt ongeveer gelijk aan het SSH of Git protocol, maar verloopt via standaard HTTPS poorten en kan verscheidene HTTP authenticatie mechanismen gebruiken, wat betekent dat het vaak eenvoudiger is voor de gebruiker dan iets als SSH, omdat je zaken als basale gebruikersnaam/wachtwoord authenticatie kunt gebruiken in plaats van SSH sleutels in te richten. Het is waarschijnlijk de meest populaire manier om Git te gebruiken geworden, omdat het zowel ingericht kan worden om anoniem gebruik te faciliteren zoals bij het git:// protocol, maar ook met authenticatie en encryptie zoals bij het SSH protocol. In plaats van verschillende URLs op te zetten voor deze dingen, kan je nu een enkele URL voor beide gebruiken. Als je probeert te pushen en de repository heeft authenticatie nodig (wat normaalgesproken wel zou moeten), kan de server om gebruikernaam en wachtwoord vragen. Hetzelfde geldt voor lees-toegang. Sterker nog, voor services als GitHub is de URL die je gebruikt om de repository online te bekijken (bijvoorbeeld, https://github.com/schacon/simplegit ) is dezelfde URL die je kunt gebruiken om te clonen en, als je toegang hebt, om te pushen. Domme HTTP Als de server niet werkt met een Git HTTP slimme service, zal de Git client terug proberen te vallen op het simpelere domme HTTP protocol. Het domme protocol verwacht dat de kale Git repository moet worden verspreid als gewone bestanden van de web server. Het mooie van het domme HTTP protocol is de eenvoud van het inrichten. Alles wat je hoeft te doen is een kale Git repository onder je HTTP document-root neer te zetten en een speciale post-update hook in te richten, en je bent klaar (Zie Git Hooks ). Vanaf dat moment, kan iedereen die de webserver waarop je de repository hebt gezet kan bereiken ook je repository klonen. Om lees-toegang tot je repository toe te staan over HTTP, kan je iets als dit doen: $ cd /var/www/htdocs/ $ git clone --bare /path/to/git_project gitproject.git $ cd gitproject.git $ mv hooks/post-update.sample hooks/post-update $ chmod a+x hooks/post-update Dat is alles. De post-update hook dat standaard bij Git geleverd wordt activeert het juiste commando ( git update-server-info ) om HTTP fetching en cloning correct te laten werken. Dit commando wordt gedraaid als je naar deze repository pusht (misschien met SSH); en andere mensen kunnen clonen met zoiets als $ git clone https://example.com/gitproject.git In dit specifieke geval gebruiken we het /var/www/htdocs pad dat gebruikelijk is voor Apache setups, maar je kunt elke statische web server gebruiken — zet gewoon de kale repository in het pad. De Git data wordt verspreid als gewone statische bestanden (zie Git Binnenwerk voor details hoe precies het wordt verspreid). Over het algemeen zou je kunnen kiezen om een lees/schrijf slimme HTTP server te draaien of om de bestanden beschikbaar te stellen als alleen lezen op de domme manier. Het is zeldzaam om een combinatie van beide services te draaien. De voordelen Ze zullen ons voornamelijk richten op de voordelen van de slimme versie van het HTTP protocol. De eenvoud van het hebben van een enkele URL voor alle typen van toegang en de server alleen te laten vragen om authenticatie wanneer het noodzakelijk is maakt het heel erg eenvoudig voor de eindgebruiker. Het kunnen authenticeren met gebruikersnaam en wachtwoord is ook een groot voordeel ten opzichte van SSH, omdat de gebruikers geen SSH sleutels lokaal hoeven te genereren en hun publieke sleutels niet naar de server hoeven te uploaden voordat ze in staat zijn met deze te communiceren. Voor minder kundige gebruikers, of gebruikers op systemen waar SSH minder gebruikelijk is, is dit een groot voordeel in bruikbaarheid. Het is ook een erg snel en efficiënt protocol, vergelijkbaar met SSH. Je kunt ook je repositories met alleen leesrechten verspreiden middels HTTPS, wat inhoudt dat je de inhoud versleuteld verstuurt; of je kunt zelfs zover gaan dat je de clients speciaal getekende SSL certificaten laat gebruiken. Een ander aardigheid is dat het HTTPS protocol zo gebruikelijk is dat de firewalls van bedrijven vaak zo zijn opgezet dat ze verkeer via deze poorten toestaan. De nadelen Git via HTTPS kan op sommige servers iets moeilijker op te zetten zijn in vergelijking tot SSH. Anders dan dat, is er weinig dat andere protocollen in het voordeel laat zijn als we het hebben over het slimme HTTP protocol om Git inhoud te bedienen. Als je HTTP gebruikt voor geauthenticeerde pushen, is het invullen van credentials (gebruikersnaam en wachtwoord) soms ingewikkelder dan het gebruik sleutels via SSH. Er zijn echter een aantal credential caching tools die je kunt gebruiken, waaronder Keychain toegang op macOS en Credential Manager op Windows om dit redelijk gladjes te laten verlopen. Lees Het opslaan van inloggegevens om te zien hoe veilige HTTP wachtwoord caching op jouw systeem op te zetten. Het SSH Protocol Een gebruikelijk transport protocol voor Git in het geval van zelf-hosting is SSH. Dit is omdat SSH toegang tot servers in de meeste gevallen al ingericht is — en als dat niet het geval is, is het eenvoudig te doen. SSH is ook een authenticerend netwerk protocol, en omdat het alom aanwezig is, is het over het algemeen eenvoudig om in te richten en te gebruiken. Om een Git repository via SSH te clonen, kan je een ssh:// URL specificeren op deze manier: $ git clone ssh://[user@]server/project.git Of je kunt het kortere scp-achtige syntax voor het SSH protocol gebruiken: $ git clone [user@]server:project.git Als je in beide bovenstaande gevallen de optionele gebruikersnaam niet opgeeft, neemt Git de gebruikernaam over waarmee je op dat moment ingelogd bent. De voordelen Er zijn vele voordelen om SSH te gebruiken. Ten eerste is SSH relatief eenvoudig op te zetten — SSH daemons komen veel voor, veel systeembeheerders hebben er ervaring mee, en veel operating systemen zijn er mee uitgerust of hebben applicaties om ze te beheren. Daarnaast is toegang via SSH veilig — alle data transporten zijn versleuteld en geauthenticeerd. En als laatste is SSH efficiënt, net zoals bij het HTTPS, Git en lokale protocol worden de gegevens zo compact mogelijk gemaakt voordat het getransporteerd wordt. De nadelen Het negatieve aspect van SSH is dat je er geen anonieme toegang naar je repository over kunt geven. Mensen moeten via SSH toegang hebben om er gebruik van te kunnen maken zelfs als het alleen lezen is, dit maakt dat SSH toegang niet echt bevordelijk is voor open source projecten. Als je het alleen binnen je bedrijfsnetwerk gebruikt, is SSH wellicht het enige protocol waar je mee in aanraking komt. Als je anonieme alleen-lezen toegang wilt toestaan tot je projecten, dan moet je SSH voor jezelf instellen om over te pushen, maar iets anders voor anderen om te fetchen. Het Git Protocol Tot slot is er het Git protocol. Dit is een speciale daemon dat met Git mee wordt geleverd, het luistert naar een toegewezen poort (9418) en biedt een service die vergelijkbaar is met het SSH protocol, maar zonder enige vorm van authenticatie. Om een repository via het Git protocol te laten verspreiden, moet je eerst een git-daemon-export-ok bestand maken — de daemon zal een repository zonder dat bestand niet verspreiden, maar anders dan dat is er geen beveiliging. De Git repository is beschikbaar voor iedereen om te klonen of het is het voor niemand. Dit betekent dat er over het algemeen niet wordt gepusht met dit protocol. Je kunt push-toegang beschikbaar maken, maar gegeven het ontbreken van authenticatie kan iedereen op het internet die de vingers kan leggen op het URL van je project naar jouw project kan pushen. Laten we zeggen dat dit zelden voorkomt. De voordelen Het Git protocol is vaak het snelste netwerk overdrachtsprotocol beschikbaar. Als je veel verkeer voor een publiek project moet bedienen of een erg groot project dat geen authenticatie behoeft voor lees-toegang, is het waarschijnlijk dat je een Git daemon wilt opzetten om je project te bedienen. Het gebruikt dezelfde data-transfer mechanisme als het SSH protocol, maar zonder de encryptie en authenticatie-overhead. De nadelen Het nadeel van het Git protocol is het ontbreken van authenticatie. Het is over het algemeen niet wenselijk om alleen middels het Git protocol toegang te geven tot je project. Over het algemeen zal je dit koppelen met SSH of HTTPS toegang voor de enkele ontwikkelaars die push (schrijf)rechten hebben en voor alle anderen het git:// protocol voor leestoegang. Daarbij is het waarschijnlijk het moeilijkste protocol om in te richten. Het moet zijn eigen daemon draaien, wat betekent dat de xinetd of systemd configuratie (of iets verglijkbaars) wat niet het eenvoudigste is om te doen. Ook moet de firewall toegang tot poort 9418 worden opengezet, wat geen standaard poort is die in de firewalls van bedrijven wordt toegestaan. In de firewall van grote bedrijven is deze obscure poort doorgaans geblokkeerd. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/commands/cargo-publish.html#option-cargo-publish---allow-dirty
cargo publish - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book cargo-publish(1) NAME cargo-publish — Upload a package to the registry SYNOPSIS cargo publish [ options ] DESCRIPTION This command will create a distributable, compressed .crate file with the source code of the package in the current directory and upload it to a registry. The default registry is https://crates.io . This performs the following steps: Performs a few checks, including: Checks the package.publish key in the manifest for restrictions on which registries you are allowed to publish to. Create a .crate file by following the steps in cargo-package(1) . Upload the crate to the registry. The server will perform additional checks on the crate. The client will poll waiting for the package to appear in the index, and may timeout. In that case, you will need to check for completion manually. This timeout does not affect the upload. This command requires you to be authenticated using cargo-login(1) or environment variables of the registry.token and registries.<name>.token config fields. See the reference for more details about packaging and publishing. OPTIONS Publish Options --dry-run Perform all checks without uploading. --no-verify Don’t verify the contents by building them. --allow-dirty Allow working directories with uncommitted VCS changes to be packaged. --index index The URL of the registry index to use. --registry registry Name of the registry to publish to. Registry names are defined in Cargo config files . If not specified, and there is a package.publish field in Cargo.toml with a single registry, then it will publish to that registry. Otherwise it will use the default registry, which is defined by the registry.default config key which defaults to crates-io . Package Selection By default, when no package selection options are given, the packages selected depend on the selected manifest file (based on the current working directory if --manifest-path is not given). If the manifest is the root of a workspace then the workspaces default members are selected, otherwise only the package defined by the manifest will be selected. The default members of a workspace can be set explicitly with the workspace.default-members key in the root manifest. If this is not set, a virtual workspace will include all workspace members (equivalent to passing --workspace ), and a non-virtual workspace will include only the root crate itself. -p spec … --package spec … Publish only the specified packages. See cargo-pkgid(1) for the SPEC format. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. --workspace Publish all members in the workspace. --all Deprecated alias for --workspace . --exclude SPEC … Exclude the specified packages. Must be used in conjunction with the --workspace flag. This flag may be specified multiple times and supports common Unix glob patterns like * , ? and [] . However, to avoid your shell accidentally expanding glob patterns before Cargo handles them, you must use single quotes or double quotes around each pattern. Compilation Options --target triple Publish for the specified target architecture. Flag may be specified multiple times. The default is the host architecture. The general format of the triple is <arch><sub>-<vendor>-<sys>-<abi> . Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. This may also be specified with the build.target config value . Note that specifying this flag makes Cargo run in a different mode where the target artifacts are placed in a separate directory. See the build cache documentation for more details. --target-dir directory Directory for all generated artifacts and intermediate files. May also be specified with the CARGO_TARGET_DIR environment variable, or the build.target-dir config value . Defaults to target in the root of the workspace. Feature Selection The feature flags allow you to control which features are enabled. When no feature options are given, the default feature is activated for every selected package. See the features documentation for more details. -F features --features features Space or comma separated list of features to activate. Features of workspace members may be enabled with package-name/feature-name syntax. This flag may be specified multiple times, which enables all specified features. --all-features Activate all available features of all selected packages. --no-default-features Do not activate the default feature of the selected packages. Manifest Options --manifest-path path Path to the Cargo.toml file. By default, Cargo searches for the Cargo.toml file in the current directory or any parent directory. --locked Asserts that the exact same dependencies and versions are used as when the existing Cargo.lock file was originally generated. Cargo will exit with an error when either of the following scenarios arises: The lock file is missing. Cargo attempted to change the lock file due to a different dependency resolution. It may be used in environments where deterministic builds are desired, such as in CI pipelines. --offline Prevents Cargo from accessing the network for any reason. Without this flag, Cargo will stop with an error if it needs to access the network and the network is not available. With this flag, Cargo will attempt to proceed without the network if possible. Beware that this may result in different dependency resolution than online mode. Cargo will restrict itself to crates that are downloaded locally, even if there might be a newer version as indicated in the local copy of the index. See the cargo-fetch(1) command to download dependencies before going offline. May also be specified with the net.offline config value . --frozen Equivalent to specifying both --locked and --offline . --lockfile-path PATH Changes the path of the lockfile from the default ( <workspace_root>/Cargo.lock ) to PATH . PATH must end with Cargo.lock (e.g. --lockfile-path /tmp/temporary-lockfile/Cargo.lock ). Note that providing --lockfile-path will ignore existing lockfile at the default path, and instead will either use the lockfile from PATH , or write a new lockfile into the provided PATH if it doesn’t exist. This flag can be used to run most commands in read-only directories, writing lockfile into the provided PATH . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #14421 ). Miscellaneous Options -j N --jobs N Number of parallel jobs to run. May also be specified with the build.jobs config value . Defaults to the number of logical CPUs. If negative, it sets the maximum number of parallel jobs to the number of logical CPUs plus provided value. If a string default is provided, it sets the value back to defaults. Should not be 0. --keep-going Build as many crates in the dependency graph as possible, rather than aborting the build on the first one that fails to build. For example if the current package depends on dependencies fails and works , one of which fails to build, cargo publish -j1 may or may not build the one that succeeds (depending on which one of the two builds Cargo picked to run first), whereas cargo publish -j1 --keep-going would definitely run both builds, even if the one run first fails. Display Options -v --verbose Use verbose output. May be specified twice for “very verbose” output which includes extra output such as dependency warnings and build script output. May also be specified with the term.verbose config value . -q --quiet Do not print cargo log messages. May also be specified with the term.quiet config value . --color when Control when colored output is used. Valid values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. May also be specified with the term.color config value . Common Options + toolchain If Cargo has been installed with rustup, and the first argument to cargo begins with + , it will be interpreted as a rustup toolchain name (such as +stable or +nightly ). See the rustup documentation for more information about how toolchain overrides work. --config KEY=VALUE or PATH Overrides a Cargo configuration value. The argument should be in TOML syntax of KEY=VALUE , or provided as a path to an extra configuration file. This flag may be specified multiple times. See the command-line overrides section for more information. -C PATH Changes the current working directory before executing any specified operations. This affects things like where cargo looks by default for the project manifest ( Cargo.toml ), as well as the directories searched for discovering .cargo/config.toml , for example. This option must appear before the command name, for example cargo -C path/to/my-project build . This option is only available on the nightly channel and requires the -Z unstable-options flag to enable (see #10098 ). -h --help Prints help information. -Z flag Unstable (nightly-only) flags to Cargo. Run cargo -Z help for details. ENVIRONMENT See the reference for details on environment variables that Cargo reads. EXIT STATUS 0 : Cargo succeeded. 101 : Cargo failed to complete. EXAMPLES Publish the current package: cargo publish SEE ALSO cargo(1) , cargo-package(1) , cargo-login(1)
2026-01-13T09:29:19
https://www.linkedin.com/products/flip-app-flip/
Flip | LinkedIn Skip to main content LinkedIn Flip in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Flip Employee Communications Software by Flip See who's skilled in this Add as skill Learn more Report this product About The AI employee experience platform where work meets wow. Streamline tasks, spark conversations, and inspire your teams to reach new heights – from head office to the floor, and door-to-door. This product is intended for Human Resources Specialist Human Resources Manager Human Resources Director Internal Communications Manager Internal Communication Specialist Media Products media viewer No more previous content Meet Flip: The Employee Super-App Built for Frontline Teams Empower every employee with Flip, the all-in-one employee app built for frontline teams. Streamline tasks, spark conversations, and inspire teams – from head office to the floor, and door-to-door. This is Flip. Flip Intelligence | AI-Powered Automation for Frontline Teams Flip Intelligence combines your company’s brainpower — communications, internal knowledge, and workflows — into one AI-powered system that gets things done. While GenAI equips every employee with the right information at the right time, Agentic AI proactively takes care of the next steps. Knowledge Base: Put Internal Knowledge at Their Fingertips Flip Intelligence combines three powerful features that centralise key information, automate everyday tasks, and connect your people to what matters. With Knowledge Base, you can put internal knowledge right in your team’s pocket. Centralise everything from SOPs to HR policies, and make it searchable, secure, and instantly accessible — all from the app your employees use every day. How Ben & Jerry’s Connects Their Team to Everything They Need with One App Discover how Ben & Jerry’s ANZ gave every employee a simpler way to stay informed, engaged, and connected – replacing scattered tools with one branded hub and achieving 70% adoption within just three days. No more next content Featured customers of Flip EUROPART Group Wholesale 9,249 followers DB Schenker Transportation, Logistics, Supply Chain and Storage 1,063,239 followers Magna International Motor Vehicle Manufacturing 879,907 followers REWE Retail 22,641 followers GTÜ Gesellschaft für Technische Überwachung mbH Public Safety 3,185 followers Dirk Rossmann GmbH Retail 54,880 followers Schwarzwald-Baar Klinikum Hospitals and Health Care 1,646 followers Paul Bauder GmbH & Co. KG Construction 7,542 followers MAHLE Motor Vehicle Manufacturing 646,289 followers Klinikum Würzburg Mitte gGmbH Hospitals 760 followers Bosch Software Development 1,600,877 followers elasto GmbH & Co. KG Wholesale 2,425 followers TEDi Retail 25,127 followers Baxter International Inc. Medical Equipment Manufacturing 1,222,575 followers GLS Transportation, Logistics, Supply Chain and Storage 70,323 followers Städtisches Klinikum Karlsruhe gGmbH Hospitals and Health Care 2,713 followers BABOR BEAUTY GROUP Personal Care Product Manufacturing 23,064 followers Kodi Handels GmbH Retail 2,347 followers toom Baumarkt Retail 13,315 followers Show more Show less Similar products Zoho Cliq Zoho Cliq Employee Communications Software HCL Sametime HCL Sametime Employee Communications Software Workvivo by Zoom Workvivo by Zoom Employee Communications Software Poppulo Poppulo Employee Communications Software Staffbase Staffbase Employee Communications Software Firstup Firstup Employee Communications Software Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/ms/v2/Git-on-the-Server-GitLab
Git - GitLab About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Getting Started 1.1 About Version Control 1.2 A Short History of Git 1.3 What is Git? 1.4 The Command Line 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Getting Help 1.8 Summary 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.8 Git on the Server - GitLab GitLab GitWeb is pretty simplistic though. If you’re looking for a modern, fully featured Git server, there are several open source solutions out there that you can install instead. As GitLab is one of the popular ones, we’ll cover installing and using it as an example. This is harder than the GitWeb option and will require more maintenance, but it is a fully featured option. Installation GitLab is a database-backed web application, so its installation is more involved than some other Git servers. Fortunately, this process is well-documented and supported. GitLab strongly recommends installing GitLab on your server via the official Omnibus GitLab package. The other installation options are: GitLab Helm chart, for use with Kubernetes. Dockerized GitLab packages for use with Docker. From the source files. Cloud provider such as AWS, Google Cloud Platform, Azure, OpenShift and Digital Ocean. For more information read the GitLab Community Edition (CE) readme . Administration GitLab’s administration interface is accessed over the web. Simply point your browser to the hostname or IP address where GitLab is installed, and log in as the admin user. The default username is admin@local.host , and the default password is 5iveL!fe (which you must change right away). After you’ve logged in, click the “Admin area” icon in the menu at the top right. Figure 50. The “Admin area” item in the GitLab menu Users Everybody using your GitLab server must have a user account. User accounts are quite simple, they mainly contain personal information attached to login data. Each user account has a namespace , which is a logical grouping of projects that belong to that user. If the user jane had a project named project, that project’s url would be http://server/jane/project . Figure 51. The GitLab user administration screen You can remove a user account in two ways: “Blocking” a user prevents them from logging into the GitLab instance, but all of the data under that user’s namespace will be preserved, and commits signed with that user’s email address will still link back to their profile. “Destroying” a user, on the other hand, completely removes them from the database and filesystem. All projects and data in their namespace is removed, and any groups they own will also be removed. This is obviously a much more permanent and destructive action, and you will rarely need it. Groups A GitLab group is a collection of projects, along with data about how users can access those projects. Each group has a project namespace (the same way that users do), so if the group training has a project materials, its url would be http://server/training/materials . Figure 52. The GitLab group administration screen Each group is associated with a number of users, each of which has a level of permissions for the group’s projects and the group itself. These range from “Guest” (issues and chat only) to “Owner” (full control of the group, its members, and its projects). The types of permissions are too numerous to list here, but GitLab has a helpful link on the administration screen. Projects A GitLab project roughly corresponds to a single Git repository. Every project belongs to a single namespace, either a user or a group. If the project belongs to a user, the owner of the project has direct control over who has access to the project; if the project belongs to a group, the group’s user-level permissions will take effect. Every project has a visibility level, which controls who has read access to that project’s pages and repository. If a project is Private , the project’s owner must explicitly grant access to specific users. An Internal project is visible to any logged-in user, and a Public project is visible to anyone. Note that this controls both git fetch access as well as access to the web UI for that project. Hooks GitLab includes support for hooks, both at a project or system level. For either of these, the GitLab server will perform an HTTP POST with some descriptive JSON whenever relevant events occur. This is a great way to connect your Git repositories and GitLab instance to the rest of your development automation, such as CI servers, chat rooms, or deployment tools. Basic Usage The first thing you’ll want to do with GitLab is create a new project. You can do this by clicking on the “+” icon on the toolbar. You’ll be asked for the project’s name, which namespace it should belong to, and what its visibility level should be. Most of what you specify here isn’t permanent, and can be changed later through the settings interface. Click “Create Project”, and you’re done. Once the project exists, you’ll probably want to connect it with a local Git repository. Each project is accessible over HTTPS or SSH, either of which can be used to configure a Git remote. The URLs are visible at the top of the project’s home page. For an existing local repository, this command will create a remote named gitlab to the hosted location: $ git remote add gitlab https://server/namespace/project.git If you don’t have a local copy of the repository, you can simply do this: $ git clone https://server/namespace/project.git The web UI provides access to several useful views of the repository itself. Each project’s home page shows recent activity, and links along the top will lead you to views of the project’s files and commit log. Working Together The simplest way of working together on a GitLab project is by giving each user direct push access to the Git repository. You can add a user to a project by going to the “Members” section of that project’s settings, and associating the new user with an access level (the different access levels are discussed a bit in Groups ). By giving a user an access level of “Developer” or above, that user can push commits and branches directly to the repository. Another, more decoupled way of collaboration is by using merge requests. This feature enables any user that can see a project to contribute to it in a controlled way. Users with direct access can simply create a branch, push commits to it, and open a merge request from their branch back into master or any other branch. Users who don’t have push permissions for a repository can “fork” it to create their own copy, push commits to their copy, and open a merge request from their fork back to the main project. This model allows the owner to be in full control of what goes into the repository and when, while allowing contributions from untrusted users. Merge requests and issues are the main units of long-lived discussion in GitLab. Each merge request allows a line-by-line discussion of the proposed change (which supports a lightweight kind of code review), as well as a general overall discussion thread. Both can be assigned to users, or organized into milestones. This section is focused mainly on the Git-related features of GitLab, but as a mature project, it provides many other features to help your team work together, such as project wikis and system maintenance tools. One benefit to GitLab is that, once the server is set up and running, you’ll rarely need to tweak a configuration file or access the server via SSH; most administration and general usage can be done through the in-browser interface. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/sl/v2/Osnove-Git-Delo-z-daljavami
Git - Delo z daljavami About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Začetek 1.1 O nadzoru različic 1.2 Kratka zgodovina Gita 1.3 Kaj je Git? 1.4 Ukazna vrstica 1.5 Namestitev Gita 1.6 Prva nastavitev Gita 1.7 Pridobivanje pomoči 1.8 Povzetek 2. Osnove Git 2.1 Pridobivanje repozitorija Git 2.2 Snemanje sprememb v repozitorij 2.3 Pregled zgodovine potrditev 2.4 Razveljavljanje stvari 2.5 Delo z daljavami 2.6 Označevanje 2.7 Aliasi Git 2.8 Povzetek 3. Veje Git 3.1 Veje na kratko 3.2 Osnove vej in združevanja 3.3 Upravljanje vej 3.4 Poteki dela z vejami 3.5 Oddaljene veje 3.6 Ponovno baziranje 3.7 Povzetek 4. Git na strežniku 4.1 Protokoli 4.2 Pridobitev Gita na strežniku 4.3 Generiranje vaših javnih ključev SSH 4.4 Nastavitev strežnika 4.5 Prikriti proces Git 4.6 Pametni HTTP 4.7 GitWeb 4.8 GitLab 4.9 Možnosti gostovanja pri tretjih ponudnikih 4.10 Povzetek 5. Porazdeljeni Git 5.1 Porazdeljeni poteki dela 5.2 Prispevek k projektu 5.3 Vzdrževanje projekta 5.4 Povzetek 6. GitHub 6.1 Namestitev in konfiguracija računa 6.2 Prispevek k projektu 6.3 Vzdrževanje projekta 6.4 Upravljanje organizacije 6.5 Skriptni GitHub 6.6 Povzetek 7. Orodja Git 7.1 Izbira revizije 7.2 Interaktivno pripravljanje 7.3 Shranjevanje na varno (angl. stashing) in čiščenje 7.4 Podpisovanje vašega dela 7.5 Iskanje 7.6 Prepisovanje zgodovine 7.7 Demistifikacija ponastavitve 7.8 Napredno združevanje 7.9 Rerere 7.10 Razhroščevanje z Gitom 7.11 Podmoduli 7.12 Povezovanje v pakete 7.13 Zamenjava 7.14 Shramba poverilnic 7.15 Povzetek 8. Prilagoditev Gita 8.1 Konfiguracija Git 8.2 Atributi Git 8.3 Kljuke Git 8.4 Primer pravilnika, ki ga uveljavlja Git 8.5 Povzetek 9. Git in ostali sistemi 9.1 Git kot odjemalec 9.2 Migracija na Git 9.3 Povzetek 10. Notranjost Gita 10.1 Napeljava in keramika 10.2 Objekti Git 10.3 Reference Git 10.4 Packfiles (datoteke zmanjšanih podatkov) 10.5 Refspec 10.6 Protokoli prenosa 10.7 Vzdrževanje in obnovitev podatkov 10.8 Spremenljivke okolja 10.9 Povzetek A1. Dodatek A: Git v drugih okoljih A1.1 Grafični vmesniki A1.2 Git v Visual Studio A1.3 Git v Visual Studio Code A1.4 Git v IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git v Sublime Text A1.6 Git v Bashu A1.7 Git v Zsh A1.8 Git v Powershellu A1.9 Povzetek A2. Dodatek B: Vdelava Gita v vašo aplikacijo A2.1 Git v ukazni vrstici A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Dodatek C: Ukazi Git A3.1 Nastavitev in konfiguracija A3.2 Pridobivanje in ustvarjanje projektov A3.3 Osnove posnetkov A3.4 Veje in združevanje A3.5 Deljenje in posodabljanje projektov A3.6 Pregled in primerjava A3.7 Razhroščevanje A3.8 Popravljanje A3.9 E-pošta A3.10 Zunanji sistemi A3.11 Administracija A3.12 Orodja za sisteme napeljave 2nd Edition 2.5 Osnove Git - Delo z daljavami Delo z daljavami Da lahko sodelujete na kateremkoli projektu Git, morate vedeti, kako upravljati vaše oddaljene repozitorije. Oddaljeni repozitoriji so različice vašega projekta, ki gostujejo nekje na internetu ali omrežju. Imate jih lahko več, v splošnem je za vas vsak od njih samo za branje ali pa za branje/pisanje. Sodelovanje z drugimi vključuje upravljanje teh oddaljenih repozitorijev, potiskanje in vlečenje podatkov vanje ali iz njih, ko morate deliti delo. Upravljanje oddaljenih repozitorijev vključuje vedeti, kako dodati oddaljene repozitorije, odstraniti daljave, ki niso več veljavne, upravljati različne oddaljene veje in kako jih definirati kot sledene ali ne in še več. V tem razdelku bomo pokrili nekaj od teh veščin upravljanja daljav. Opomba Oddaljeni repozitoriji so lahko na vaši lokalni napravi. Popolnoma mogoče je, da delate z oddaljenim (angl. remote ) repozitorijem, ki je v resnici na istem gostitelju kot vi. Beseda »oddaljen« ne pomeni nujno, da je repozitorij drugje na omrežju ali na internetu, temveč samo, da je drugje. Delo s takšnim oddaljenim repozitorijem še vedno vključuje vse standardne operacije potiskanja, vlečenja in pridobivanja kot pri katerem koli drugem oddaljenem repozitoriju. Prikaz vaših daljav Da pogledate, katere oddaljene strežnike ste nastavili, lahko poženete ukaz git remote . Izpiše kratka imena vsakega oprimka daljave, ki ste ga določili. Če ste klonirali svoj repozitorij, bi morali videti vsaj t. i. origin (izvor) — to je privzeto ime, ki ga Git da strežniku, iz katerega ste klonirali: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin Določite lahko tudi -v , ki vam pokaže URL-je, ki jih je Git shranil za kratko ime, ki bo uporabljeno, ko se bo bralo in pisalo na to daljavo: $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) Če imate več kot eno daljavo, ukaz izpiše vse. Na primer, repozitorij z več daljavami za delo z več sodelavci je lahko videti nekako takole. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) To pomeni, da lahko precej enostavno povlečemo prispevke od kateregakoli od teh uporabnikov. Morda moramo dodatno imeti pravice za potiskanje v enega ali več od le-teh, vendar tega tu ne moremo vedeti. Bodite pozorni, saj te daljave uporabljajo številne protokole; več o tem bomo pokrili v Pridobitev Gita na strežniku . Dodajanje oddaljenih repozitorijev Omenili in podali smo nekaj demonstracij, kako vam ukaz git clone posredno doda daljavo origin . Tako se doda nova daljava izrecno. Da dodate nov oddaljeni repozitorij Git kot kratko ime, na katerega se lahko enostavno sklicujete, poženite git remote add <shortname> <url> : $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) Sedaj lahko v ukazni vrstici uporabite niz pb namesto celotnega URL-ja. Na primer, če želite prenesti vse informacije, ki jih ima Paul, vendar jih vi še nimate v svojem repozitoriju, lahko poženete git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit Paulova veja master je sedaj dostopna lokalno kot pb/master  — lahko jo združite v eno izmed svojih vej, ali pa izvlečete lokalno vejo na tisti točki, ki jo želite preiskati. Kaj so veje in kako jih uporabljati, bomo šli bolj podrobno skozi v Veje Git . Pridobivanje in vlečenje iz vaših daljav Kot ste ravnokar videli, da dobite podatke iz svojih oddaljenih projektov, lahko poženete: $ git fetch <remote> Ukaz gre v oddaljeni projekt in iz tega oddaljenega projekta povleče vse podatke, ki jih še nimate. Ko to naredite, bi morali imeti reference na vse veje iz te daljave, ki jih lahko kadarkoli združite ali raziščete. Če klonirate repozitorij, ukaz avtomatsko doda ta oddaljeni repozitorij pod ime »origin«. Torej, git fetch origin prinese katerokoli delo, ki je bilo potisnjeno na ta strežnik, odkar ste ga klonirali (ali zadnje prenešeno iz njega). Pomembno je opaziti, da ukaz git fetch samo prenese podatke v vaš lokalni repozitorij — avtomatsko jih ne združi s katerimkoli delom ali spremeni česa, na čemer trenutno delate. Združiti jih morate ročno v svoje delo, ko ste pripravljeni. Če je vaša trenutna veja nastavljena, da sledi oddaljeni veji (za več informacij glejte naslednji razdelek in poglavje Veje Git ), lahko uporabite ukaz git pull , da avtomatsko prinese in nato združi oddaljeno vejo v vašo trenutno vejo. To je lahko za vas enostavnejši ali bolj priročen potek dela; in privzeto ukaz git clone avtomatsko nastavi vašo lokalno vejo master , da sledi oddaljeni veji master (ali kakorkoli se privzeta veja imenuje) na strežniku, iz katerega ste klonirali. Pogon git pull v splošnem prinese podatke iz strežnika, iz katerega ste prvotno klonirali, in jih skuša avtomatsko združiti v kodo na kateri trenutno delate. Opomba Od verzije Git 2.27 naprej bo git pull izpisal opozorilo, če spremenljivka pull.rebase ni nastavljena. Git bo obdržal opozorilo, dokler ne nastavite te spremenljivke. Če želite privzeto obnašanje Gita (fast-forward, če je možen, drugače ustvarite potrditev združevanja): git config --global pull.rebase "false" Če želite ponovno bazirati, ko povlečete: git config --global pull.rebase "true" Potiskanje na vaše daljave Ko imate svoj projekt na točki, ki jo želite deliti, ga morate potisniti v povratni tok. Ukaz za to je enostaven: git push <remote> <branch> . Če želite potisniti vašo vejo master na vaš strežnik origin (ponovno, kloniranje vam v splošnem nastavi obe od teh imen avtomatsko), potem lahko to poženete, da potisnete nazaj na strežnik katerekoli potrditve, ki ste jih naredili: $ git push origin master Ta ukaz deluje samo, če ste klonirali iz strežnika, za katerega imate pravice pisanja in če nihče vmes ni ničesar potisnil. Če vi in še nekdo klonirata istočasno in drugi potisne proti toku ter nato potisnete proti toku še vi, bo vaše potiskanje pravilno zavrnjeno. Najprej boste morali prinesti delo drugega in ga vdelati v vaše, preden lahko potiskate. Za več podrobnih informacij si oglejte poglavje Veje Git , kako lahko potiskate na oddaljene strežnike. Preverjanje daljave Če želite videti več informacij o določeni daljavi, lahko uporabite ukaz git remote show <remote> . Če poženete ta ukaz z določenim kratkim imenom, kot je na primer origin , dobite nekaj takega: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) Izpiše URL za oddaljeni repozitorij kot tudi informacije sledene veje. Ukaz vam koristno pove, da če ste na veji master in poženete git pull , bo avtomatsko združil oddaljeno vejo master v lokalno, ko se jo prinese. Izpiše tudi vse oddaljene reference, ki jih je povlekel. To je enostaven primer, na katerega boste verjetno naleteli. Ko uporabljate Git bolj intenzivno, boste lahko videli veliko več informacij preko git remote show : $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) Ta ukaz pokaže, na katero vejo se avtomatsko potiska, ko poženete git push , medtem ko ste na določenih vejah. Pokaže vam tudi, katerih oddaljenih vej na strežniku še nimate pri sebi, katere oddaljene veje imate, ki so bile odstranjene iz strežnika, in več lokalnih vej, ki se lahko avtomatsko združijo s svojimi sledenimi oddaljenimi vejami, ko poženete git pull . Preimenovanje in odstranjevanje daljav Da spremenite kratko ime daljave, lahko poženete git remote rename . Na primer, če želite preimenovati pb v paul , lahko to naredite z git remote rename : $ git remote rename pb paul $ git remote origin paul Vredno je omeniti, da to spremeni tudi vsa imena vaših sledenih oddaljenih vej. Kar se je včasih sklicevalo na pb/master , je sedaj na paul/master . Če želite zaradi nekega razloga odstraniti daljavo — prenesli ste strežnik, ali pa določene zrcaljene lokacije ne uporabljate več, ali pa mogoče nekdo, ki je prispeval, sedaj ne dela več — lahko uporabite bodisi git remote remove ali pa git remote rm : $ git remote remove paul $ git remote origin Ko enkrat na ta način odstranite referenco na daljavo, bodo izbrisane tudi vse sledene oddaljene veje in konfiguracijske nastavitve povezane s to daljavo. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/uz/v2/GitHub-Maintaining-a-Project
Git - Maintaining a Project About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Иш бошланиши 1.1 Талқинларни бошқариш ҳақида 1.2 Git нинг қисқача тарихи 1.3 Git асоси 1.4 Командалар сатри 1.5 Git ни ўрнатиш 1.6 Git да биринчи созлашлар 1.7 Қандай ёрдам олиш мумкин? 1.8 Хулосалар 2. Git асослари 2.1 Git омборини яратиш 2.2 Ўзгаришларни омборга ёзиш 2.3 Фиксирлашлар тарихини кўриш 2.4 Ўзгаришларни бекор қилиш 2.5 Узоқ масофадаги омборлар билан ишлаш 2.6 Тамғалаш 2.7 Git да таҳаллуслар 2.8 Хулоса 3. Git да тармоқланиш 3.1 Тармоқланиш ҳақида икки оғиз сўз 3.2 Тармоқланиш ва бирлашиш асослари 3.3 Тармоқларни бошқариш 3.4 Иш жараёнларини тармоқлаш 3.5 Узоқ масофадаги тармоқлар 3.6 Қайта асосланиш 3.7 Хулосалар 4. Git серверда 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Sizning SSH ochiq (public) kalitingizni generatsiyalash 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Хулосалар 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Qism modullar (Submodule) 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 6.3 GitHub - Maintaining a Project Maintaining a Project Now that we’re comfortable contributing to a project, let’s look at the other side: creating, maintaining and administering your own project. Creating a New Repository Let’s create a new repository to share our project code with. Start by clicking the “New repository” button on the right-hand side of the dashboard, or from the {plus} button in the top toolbar next to your username as seen in The “New repository” dropdown. . Figure 110. The “Your repositories” area. Figure 111. The “New repository” dropdown. This takes you to the “new repository” form: Figure 112. The “new repository” form. All you really have to do here is provide a project name; the rest of the fields are completely optional. For now, just click the “Create Repository” button, and boom – you have a new repository on GitHub, named <user>/<project_name> . Since you have no code there yet, GitHub will show you instructions for how create a brand-new Git repository, or connect an existing Git project. We won’t belabor this here; if you need a refresher, check out Git асослари . Now that your project is hosted on GitHub, you can give the URL to anyone you want to share your project with. Every project on GitHub is accessible over HTTP as https://github.com/<user>/<project_name> , and over SSH as git@github.com:<user>/<project_name> . Git can fetch from and push to both of these URLs, but they are access-controlled based on the credentials of the user connecting to them. Note It is often preferable to share the HTTP based URL for a public project, since the user does not have to have a GitHub account to access it for cloning. Users will have to have an account and an uploaded SSH key to access your project if you give them the SSH URL. The HTTP one is also exactly the same URL they would paste into a browser to view the project there. Adding Collaborators If you’re working with other people who you want to give commit access to, you need to add them as “collaborators”. If Ben, Jeff, and Louise all sign up for accounts on GitHub, and you want to give them push access to your repository, you can add them to your project. Doing so will give them “push” access, which means they have both read and write access to the project and Git repository. Click the “Settings” link at the bottom of the right-hand sidebar. Figure 113. The repository settings link. Then select “Collaborators” from the menu on the left-hand side. Then, just type a username into the box, and click “Add collaborator.” You can repeat this as many times as you like to grant access to everyone you like. If you need to revoke access, just click the “X” on the right-hand side of their row. Figure 114. Repository collaborators. Managing Pull Requests Now that you have a project with some code in it and maybe even a few collaborators who also have push access, let’s go over what to do when you get a Pull Request yourself. Pull Requests can either come from a branch in a fork of your repository or they can come from another branch in the same repository. The only difference is that the ones in a fork are often from people where you can’t push to their branch and they can’t push to yours, whereas with internal Pull Requests generally both parties can access the branch. For these examples, let’s assume you are “tonychacon” and you’ve created a new Arduino code project named “fade”. Email Notifications Someone comes along and makes a change to your code and sends you a Pull Request. You should get an email notifying you about the new Pull Request and it should look something like Email notification of a new Pull Request. . Figure 115. Email notification of a new Pull Request. There are a few things to notice about this email. It will give you a small diffstat — a list of files that have changed in the Pull Request and by how much. It gives you a link to the Pull Request on GitHub. It also gives you a few URLs that you can use from the command line. If you notice the line that says git pull <url> patch-1 , this is a simple way to merge in a remote branch without having to add a remote. We went over this quickly in Checking Out Remote Branches . If you wish, you can create and switch to a topic branch and then run this command to merge in the Pull Request changes. The other interesting URLs are the .diff and .patch URLs, which as you may guess, provide unified diff and patch versions of the Pull Request. You could technically merge in the Pull Request work with something like this: $ curl http://github.com/tonychacon/fade/pull/1.patch | git am Collaborating on the Pull Request As we covered in The GitHub Flow , you can now have a conversation with the person who opened the Pull Request. You can comment on specific lines of code, comment on whole commits or comment on the entire Pull Request itself, using GitHub Flavored Markdown everywhere. Every time someone else comments on the Pull Request you will continue to get email notifications so you know there is activity happening. They will each have a link to the Pull Request where the activity is happening and you can also directly respond to the email to comment on the Pull Request thread. Figure 116. Responses to emails are included in the thread. Once the code is in a place you like and want to merge it in, you can either pull the code down and merge it locally, either with the git pull <url> <branch> syntax we saw earlier, or by adding the fork as a remote and fetching and merging. If the merge is trivial, you can also just hit the “Merge” button on the GitHub site. This will do a “non-fast-forward” merge, creating a merge commit even if a fast-forward merge was possible. This means that no matter what, every time you hit the merge button, a merge commit is created. As you can see in Merge button and instructions for merging a Pull Request manually. , GitHub gives you all of this information if you click the hint link. Figure 117. Merge button and instructions for merging a Pull Request manually. If you decide you don’t want to merge it, you can also just close the Pull Request and the person who opened it will be notified. Pull Request Refs If you’re dealing with a lot of Pull Requests and don’t want to add a bunch of remotes or do one time pulls every time, there is a neat trick that GitHub allows you to do. This is a bit of an advanced trick and we’ll go over the details of this a bit more in The Refspec , but it can be pretty useful. GitHub actually advertises the Pull Request branches for a repository as sort of pseudo-branches on the server. By default you don’t get them when you clone, but they are there in an obscured way and you can access them pretty easily. To demonstrate this, we’re going to use a low-level command (often referred to as a “plumbing” command, which we’ll read about more in Plumbing and Porcelain ) called ls-remote . This command is generally not used in day-to-day Git operations but it’s useful to show us what references are present on the server. If we run this command against the “blink” repository we were using earlier, we will get a list of all the branches and tags and other references in the repository. $ git ls-remote https://github.com/schacon/blink 10d539600d86723087810ec636870a504f4fee4d HEAD 10d539600d86723087810ec636870a504f4fee4d refs/heads/master 6a83107c62950be9453aac297bb0193fd743cd6e refs/pull/1/head afe83c2d1a70674c9505cc1d8b7d380d5e076ed3 refs/pull/1/merge 3c8d735ee16296c242be7a9742ebfbc2665adec1 refs/pull/2/head 15c9f4f80973a2758462ab2066b6ad9fe8dcf03d refs/pull/2/merge a5a7751a33b7e86c5e9bb07b26001bb17d775d1a refs/pull/4/head 31a45fc257e8433c8d8804e3e848cf61c9d3166c refs/pull/4/merge Of course, if you’re in your repository and you run git ls-remote origin or whatever remote you want to check, it will show you something similar to this. If the repository is on GitHub and you have any Pull Requests that have been opened, you’ll get these references that are prefixed with refs/pull/ . These are basically branches, but since they’re not under refs/heads/ you don’t get them normally when you clone or fetch from the server — the process of fetching ignores them normally. There are two references per Pull Request - the one that ends in /head points to exactly the same commit as the last commit in the Pull Request branch. So if someone opens a Pull Request in our repository and their branch is named bug-fix and it points to commit a5a775 , then in our repository we will not have a bug-fix branch (since that’s in their fork), but we will have pull/<pr#>/head that points to a5a775 . This means that we can pretty easily pull down every Pull Request branch in one go without having to add a bunch of remotes. Now, you could do something like fetching the reference directly. $ git fetch origin refs/pull/958/head From https://github.com/libgit2/libgit2 * branch refs/pull/958/head -> FETCH_HEAD This tells Git, “Connect to the origin remote, and download the ref named refs/pull/958/head .” Git happily obeys, and downloads everything you need to construct that ref, and puts a pointer to the commit you want under .git/FETCH_HEAD . You can follow that up with git merge FETCH_HEAD into a branch you want to test it in, but that merge commit message looks a bit weird. Also, if you’re reviewing a lot of pull requests, this gets tedious. There’s also a way to fetch all of the pull requests, and keep them up to date whenever you connect to the remote. Open up .git/config in your favorite editor, and look for the origin remote. It should look a bit like this: [remote "origin"] url = https://github.com/libgit2/libgit2 fetch = +refs/heads/*:refs/remotes/origin/* That line that begins with fetch = is a “refspec.” It’s a way of mapping names on the remote with names in your local .git directory. This particular one tells Git, "the things on the remote that are under refs/heads should go in my local repository under refs/remotes/origin ." You can modify this section to add another refspec: [remote "origin"] url = https://github.com/libgit2/libgit2.git fetch = +refs/heads/*:refs/remotes/origin/* fetch = +refs/pull/*/head:refs/remotes/origin/pr/* That last line tells Git, “All the refs that look like refs/pull/123/head should be stored locally like refs/remotes/origin/pr/123 .” Now, if you save that file, and do a git fetch : $ git fetch # … * [new ref] refs/pull/1/head -> origin/pr/1 * [new ref] refs/pull/2/head -> origin/pr/2 * [new ref] refs/pull/4/head -> origin/pr/4 # … Now all of the remote pull requests are represented locally with refs that act much like tracking branches; they’re read-only, and they update when you do a fetch. This makes it super easy to try the code from a pull request locally: $ git checkout pr/2 Checking out files: 100% (3769/3769), done. Branch pr/2 set up to track remote branch pr/2 from origin. Switched to a new branch 'pr/2' The eagle-eyed among you would note the head on the end of the remote portion of the refspec. There’s also a refs/pull/#/merge ref on the GitHub side, which represents the commit that would result if you push the “merge” button on the site. This can allow you to test the merge before even hitting the button. Pull Requests on Pull Requests Not only can you open Pull Requests that target the main or master branch, you can actually open a Pull Request targeting any branch in the network. In fact, you can even target another Pull Request. If you see a Pull Request that is moving in the right direction and you have an idea for a change that depends on it or you’re not sure is a good idea, or you just don’t have push access to the target branch, you can open a Pull Request directly to it. When you go to open a Pull Request, there is a box at the top of the page that specifies which branch you’re requesting to pull to and which you’re requesting to pull from. If you hit the “Edit” button at the right of that box you can change not only the branches but also which fork. Figure 118. Manually change the Pull Request target fork and branch. Here you can fairly easily specify to merge your new branch into another Pull Request or another fork of the project. Mentions and Notifications GitHub also has a pretty nice notifications system built in that can come in handy when you have questions or need feedback from specific individuals or teams. In any comment you can start typing a @ character and it will begin to autocomplete with the names and usernames of people who are collaborators or contributors in the project. Figure 119. Start typing @ to mention someone. You can also mention a user who is not in that dropdown, but often the autocompleter can make it faster. Once you post a comment with a user mention, that user will be notified. This means that this can be a really effective way of pulling people into conversations rather than making them poll. Very often in Pull Requests on GitHub people will pull in other people on their teams or in their company to review an Issue or Pull Request. If someone gets mentioned on a Pull Request or Issue, they will be “subscribed” to it and will continue getting notifications any time some activity occurs on it. You will also be subscribed to something if you opened it, if you’re watching the repository or if you comment on something. If you no longer wish to receive notifications, there is an “Unsubscribe” button on the page you can click to stop receiving updates on it. Figure 120. Unsubscribe from an Issue or Pull Request. The Notifications Page When we mention “notifications” here with respect to GitHub, we mean a specific way that GitHub tries to get in touch with you when events happen and there are a few different ways you can configure them. If you go to the “Notification center” tab from the settings page, you can see some of the options you have. Figure 121. Notification center options. The two choices are to get notifications over “Email” and over “Web” and you can choose either, neither or both for when you actively participate in things and for activity on repositories you are watching. Web Notifications Web notifications only exist on GitHub and you can only check them on GitHub. If you have this option selected in your preferences and a notification is triggered for you, you will see a small blue dot over your notifications icon at the top of your screen as seen in Notification center. . Figure 122. Notification center. If you click on that, you will see a list of all the items you have been notified about, grouped by project. You can filter to the notifications of a specific project by clicking on it’s name in the left hand sidebar. You can also acknowledge the notification by clicking the checkmark icon next to any notification, or acknowledge all of the notifications in a project by clicking the checkmark at the top of the group. There is also a mute button next to each checkmark that you can click to not receive any further notifications on that item. All of these tools are very useful for handling large numbers of notifications. Many GitHub power users will simply turn off email notifications entirely and manage all of their notifications through this screen. Email Notifications Email notifications are the other way you can handle notifications through GitHub. If you have this turned on you will get emails for each notification. We saw examples of this in Comments sent as email notifications and Email notification of a new Pull Request. . The emails will also be threaded properly, which is nice if you’re using a threading email client. There is also a fair amount of metadata embedded in the headers of the emails that GitHub sends you, which can be really helpful for setting up custom filters and rules. For instance, if we look at the actual email headers sent to Tony in the email shown in Email notification of a new Pull Request. , we will see the following among the information sent: To: tonychacon/fade <fade@noreply.github.com> Message-ID: <tonychacon/fade/pull/1@github.com> Subject: [fade] Wait longer to see the dimming effect better (#1) X-GitHub-Recipient: tonychacon List-ID: tonychacon/fade <fade.tonychacon.github.com> List-Archive: https://github.com/tonychacon/fade List-Post: <mailto:reply+i-4XXX@reply.github.com> List-Unsubscribe: <mailto:unsub+i-XXX@reply.github.com>,... X-GitHub-Recipient-Address: tchacon@example.com There are a couple of interesting things here. If you want to highlight or re-route emails to this particular project or even Pull Request, the information in Message-ID gives you all the data in <user>/<project>/<type>/<id> format. If this were an issue, for example, the <type> field would have been “issues” rather than “pull”. The List-Post and List-Unsubscribe fields mean that if you have a mail client that understands those, you can easily post to the list or “Unsubscribe” from the thread. That would be essentially the same as clicking the “mute” button on the web version of the notification or “Unsubscribe” on the Issue or Pull Request page itself. It’s also worth noting that if you have both email and web notifications enabled and you read the email version of the notification, the web version will be marked as read as well if you have images allowed in your mail client. Special Files There are a couple of special files that GitHub will notice if they are present in your repository. README The first is the README file, which can be of nearly any format that GitHub recognizes as prose. For example, it could be README , README.md , README.asciidoc , etc. If GitHub sees a README file in your source, it will render it on the landing page of the project. Many teams use this file to hold all the relevant project information for someone who might be new to the repository or project. This generally includes things like: What the project is for How to configure and install it An example of how to use it or get it running The license that the project is offered under How to contribute to it Since GitHub will render this file, you can embed images or links in it for added ease of understanding. CONTRIBUTING The other special file that GitHub recognizes is the CONTRIBUTING file. If you have a file named CONTRIBUTING with any file extension, GitHub will show Opening a Pull Request when a CONTRIBUTING file exists. when anyone starts opening a Pull Request. Figure 123. Opening a Pull Request when a CONTRIBUTING file exists. The idea here is that you can specify specific things you want or don’t want in a Pull Request sent to your project. This way people may actually read the guidelines before opening the Pull Request. Project Administration Generally there are not a lot of administrative things you can do with a single project, but there are a couple of items that might be of interest. Changing the Default Branch If you are using a branch other than “master” as your default branch that you want people to open Pull Requests on or see by default, you can change that in your repository’s settings page under the “Options” tab. Figure 124. Change the default branch for a project. Simply change the default branch in the dropdown and that will be the default for all major operations from then on, including which branch is checked out by default when someone clones the repository. Transferring a Project If you would like to transfer a project to another user or an organization in GitHub, there is a “Transfer ownership” option at the bottom of the same “Options” tab of your repository settings page that allows you to do this. Figure 125. Transfer a project to another GitHub user or Organization. This is helpful if you are abandoning a project and someone wants to take it over, or if your project is getting bigger and want to move it into an organization. Not only does this move the repository along with all it’s watchers and stars to another place, it also sets up a redirect from your URL to the new place. It will also redirect clones and fetches from Git, not just web requests. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/ko/v2/Git%ec%9d%98-%ea%b8%b0%ec%b4%88-%eb%a6%ac%eb%aa%a8%ed%8a%b8-%ec%a0%80%ec%9e%a5%ec%86%8c
Git - 리모트 저장소 About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. 시작하기 1.1 버전 관리란? 1.2 짧게 보는 Git의 역사 1.3 Git 기초 1.4 CLI 1.5 Git 설치 1.6 Git 최초 설정 1.7 도움말 보기 1.8 요약 2. Git의 기초 2.1 Git 저장소 만들기 2.2 수정하고 저장소에 저장하기 2.3 커밋 히스토리 조회하기 2.4 되돌리기 2.5 리모트 저장소 2.6 태그 2.7 Git Alias 2.8 요약 3. Git 브랜치 3.1 브랜치란 무엇인가 3.2 브랜치와 Merge 의 기초 3.3 브랜치 관리 3.4 브랜치 워크플로 3.5 리모트 브랜치 3.6 Rebase 하기 3.7 요약 4. Git 서버 4.1 프로토콜 4.2 서버에 Git 설치하기 4.3 SSH 공개키 만들기 4.4 서버 설정하기 4.5 Git 데몬 4.6 스마트 HTTP 4.7 GitWeb 4.8 GitLab 4.9 또 다른 선택지, 호스팅 4.10 요약 5. 분산 환경에서의 Git 5.1 분산 환경에서의 워크플로 5.2 프로젝트에 기여하기 5.3 프로젝트 관리하기 5.4 요약 6. GitHub 6.1 계정 만들고 설정하기 6.2 GitHub 프로젝트에 기여하기 6.3 GitHub 프로젝트 관리하기 6.4 Organization 관리하기 6.5 GitHub 스크립팅 6.6 요약 7. Git 도구 7.1 리비전 조회하기 7.2 대화형 명령 7.3 Stashing과 Cleaning 7.4 내 작업에 서명하기 7.5 검색 7.6 히스토리 단장하기 7.7 Reset 명확히 알고 가기 7.8 고급 Merge 7.9 Rerere 7.10 Git으로 버그 찾기 7.11 서브모듈 7.12 Bundle 7.13 Replace 7.14 Credential 저장소 7.15 요약 8. Git맞춤 8.1 Git 설정하기 8.2 Git Attributes 8.3 Git Hooks 8.4 정책 구현하기 8.5 요약 9. Git과 여타 버전 관리 시스템 9.1 Git: 범용 Client 9.2 Git으로 옮기기 9.3 요약 10. Git의 내부 10.1 Plumbing 명령과 Porcelain 명령 10.2 Git 개체 10.3 Git Refs 10.4 Packfile 10.5 Refspec 10.6 데이터 전송 프로토콜 10.7 운영 및 데이터 복구 10.8 환경변수 10.9 요약 A1. 부록 A: 다양한 환경에서 Git 사용하기 A1.1 GUI A1.2 Visual Studio A1.3 Eclipse A1.4 Bash A1.5 Zsh A1.6 Git in Powershell A1.7 요약 A2. 부록 B: 애플리케이션에 Git 넣기 A2.1 Git 명령어 A2.2 Libgit2 A2.3 JGit A2.4 go-git A3. 부록 C: Git 명령어 A3.1 설치와 설정 A3.2 프로젝트 가져오기와 생성하기 A3.3 스냅샷 다루기 A3.4 Branch와 Merge A3.5 공유하고 업데이트하기 A3.6 보기와 비교 A3.7 Debugging A3.8 Patch 하기 A3.9 Email A3.10 다른 버전 관리 시스템 A3.11 관리 A3.12 Plumbing 명령어 2nd Edition 2.5 Git의 기초 - 리모트 저장소 리모트 저장소 리모트 저장소를 관리할 줄 알아야 다른 사람과 함께 일할 수 있다. 리모트 저장소는 인터넷이나 네트워크 어딘가에 있는 저장소를 말한다. 저장소는 여러 개가 있을 수 있는데 어떤 저장소는 읽고 쓰기 모두 할 수 있고 어떤 저장소는 읽기만 가능할 수 있다. 간단히 말해서 다른 사람들과 함께 일한다는 것은 리모트 저장소를 관리하면서 데이터를 거기에 Push 하고 Pull 하는 것이다. 리모트 저장소를 관리한다는 것은 저장소를 추가, 삭제하는 것뿐만 아니라 브랜치를 관리하고 추적할지 말지 등을 관리하는 것을 말한다. 이번에는 리모트 저장소를 관리하는 방법에 대해 설명한다. 노트 원격 저장소라 하더라도 로컬 시스템에 위치할 수도 있다. “remote” 저장소라고 이름이 붙어있어도 이 원격 저장소가 사실 같은 로컬 시스템에 존재할 수도 있다. 여기서 “remote” 라는 이름은 반드시 저장소가 네트워크나 인터넷을 통해 어딘가 멀리 떨어져 있어야만 한다는 것을 의미하지 않는다. 물론 일반적인 원격 저장소와 마찬가지로 Push, Pull 등의 기능은 동일하게 사용한다. 리모트 저장소 확인하기 git remote 명령으로 현재 프로젝트에 등록된 리모트 저장소를 확인할 수 있다. 이 명령은 리모트 저장소의 단축 이름을 보여준다. 저장소를 Clone 하면 origin 이라는 리모트 저장소가 자동으로 등록되기 때문에 origin 이라는 이름을 볼 수 있다. $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin -v 옵션을 주어 단축이름과 URL을 함께 볼 수 있다. $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) 리모트 저장소가 여러 개 있다면 이 명령은 등록된 전부를 보여준다. 여러 사람과 함께 작업하는 리모트 저장소가 여러개라면 아래와 같은 결과를 얻을 수도 있다. $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) 이렇게 리모트 저장소가 여러 개 등록되어 있으면 다른 사람이 기여한 내용(Contributions)을 쉽게 가져올 수 있다. 어떤 저장소에는 Push 권한까지 제공하기도 하지만 일단 이 화면에서 Push 가능 권한까지는 확인할 수 없다. 리모트 저장소와 데이터를 주고받는데 사용하는 다양한 프로토콜에 대해서는 서버에 Git 설치하기 에서 자세히 살펴보기로 한다. 리모트 저장소 추가하기 이전 절에서도 git clone 명령이 묵시적으로 origin 리모트 저장소를 어떻게 추가되는지 설명했었지만 수박 겉핥기식으로 살펴봤을 뿐이었다. 여기에서는 리모트 저장소를 추가하는 방법을 자세하게 설명한다. 기존 워킹 디렉토리에 새 리모트 저장소를 쉽게 추가할 수 있는데 git remote add <단축이름> <url> 명령을 사용한다. $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) 이제 URL 대신에 pb 라는 이름을 사용할 수 있다. 예를 들어 로컬 저장소에는 없지만 Paul의 저장소에 있는 것을 가져오려면 아래과 같이 실행한다. $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit 로컬에서 pb/master 가 Paul의 master 브랜치이다. 이 브랜치를 로컬 브랜치중 하나에 Merge 하거나 Checkout 해서 브랜치 내용을 자세히 확인할 수 있다. (브랜치를 어떻게 사용하는지는 Git 브랜치 에서 자세히 살펴본다) 리모트 저장소를 Pull 하거나 Fetch 하기 앞서 설명했듯이 리모트 저장소에서 데이터를 가져오려면 간단히 아래와 같이 실행한다. $ git fetch <remote> 이 명령은 로컬에는 없지만, 리모트 저장소에는 있는 데이터를 모두 가져온다. 그러면 리모트 저장소의 모든 브랜치를 로컬에서 접근할 수 있어서 언제든지 Merge를 하거나 내용을 살펴볼 수 있다. 저장소를 Clone 하면 명령은 자동으로 리모트 저장소를 “origin” 이라는 이름으로 추가한다. 그래서 나중에 git fetch origin 명령을 실행하면 Clone 한 이후에(혹은 마지막으로 가져온 이후에) 수정된 것을 모두 가져온다. git fetch 명령은 리모트 저장소의 데이터를 모두 로컬로 가져오지만, 자동으로 Merge 하지 않는다. 그래서 당신이 로컬에서 하던 작업을 정리하고 나서 수동으로 Merge 해야 한다. 그냥 쉽게 git pull 명령으로 리모트 저장소 브랜치에서 데이터를 가져올 뿐만 아니라 자동으로 로컬 브랜치와 Merge 시킬 수 있다(다음 섹션과 Git 브랜치 에서 좀더 자세히 살펴본다). 먼저 git clone 명령은 자동으로 로컬의 master 브랜치가 리모트 저장소의 master 브랜치를 추적하도록 한다(물론 리모트 저장소에 master 브랜치가 있다는 가정에서). 그리고 git pull 명령은 Clone 한 서버에서 데이터를 가져오고 그 데이터를 자동으로 현재 작업하는 코드와 Merge 시킨다. 리모트 저장소에 Push 하기 프로젝트를 공유하고 싶을 때 Upstream 저장소에 Push 할 수 있다. 이 명령은 git push <리모트 저장소 이름> <브랜치 이름> 으로 단순하다. master 브랜치를 origin 서버에 Push 하려면(다시 말하지만 Clone 하면 보통 자동으로 origin 이름이 생성된다) 아래와 같이 서버에 Push 한다. $ git push origin master 이 명령은 Clone 한 리모트 저장소에 쓰기 권한이 있고, Clone 하고 난 이후 아무도 Upstream 저장소에 Push 하지 않았을 때만 사용할 수 있다. 다시 말해서 Clone 한 사람이 여러 명 있을 때, 다른 사람이 Push 한 후에 Push 하려고 하면 Push 할 수 없다. 먼저 다른 사람이 작업한 것을 가져와서 Merge 한 후에 Push 할 수 있다. Git 브랜치 에서 서버에 Push 하는 방법에 대해 자세히 설명할 것이다. 리모트 저장소 살펴보기 git remote show <리모트 저장소 이름> 명령으로 리모트 저장소의 구체적인 정보를 확인할 수 있다. origin 같은 단축이름으로 이 명령을 실행하면 아래와 같은 정보를 볼 수 있다. $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) 리모트 저장소의 URL과 추적하는 브랜치를 출력한다. 이 명령은 git pull 명령을 실행할 때 master 브랜치와 Merge 할 브랜치가 무엇인지 보여 준다. git pull 명령은 리모트 저장소 브랜치의 데이터를 모두 가져오고 나서 자동으로 Merge 할 것이다. 그리고 가져온 모든 리모트 저장소 정보도 출력한다. 좀 더 Git을 열심히 사용하다 보면 git remote show 명령으로 더 많은 정보를 보는 날이 온다. 여러분도 언젠가는 아래와 같은 메시지(역주 - 다수의 브랜치를 사용하는 메시지)를 볼 날이 올 것이다. $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) 브랜치명을 생략하고 git push 명령을 실행할 때 어떤 브랜치가 어떤 브랜치로 Push 되는지 보여준다. 또 아직 로컬로 가져오지 않은 리모트 저장소의 브랜치는 어떤 것들이 있는지, 서버에서는 삭제됐지만 아직 가지고 있는 브랜치는 어떤 것인지, git pull 명령을 실행했을 때 자동으로 Merge 할 브랜치는 어떤 것이 있는지 보여준다. 리모트 저장소 이름을 바꾸거나 리모트 저장소를 삭제하기 git remote rename 명령으로 리모트 저장소의 이름을 변경할 수 있다. 예를 들어 pb 를 paul 로 변경하려면 git remote rename 명령을 사용한다. $ git remote rename pb paul $ git remote origin paul 로컬에서 관리하던 리모트 저장소의 브랜치 이름도 바뀐다는 점을 생각해두자. 여태까지 pb/master 로 리모트 저장소 브랜치를 사용했으면 이제는 paul/master 라고 사용해야 한다. 리모트 저장소를 삭제해야 한다면 git remote remove 나 git remote rm 명령을 사용한다. 서버 정보가 바뀌었을 때, 더는 별도의 미러가 필요하지 않을 때, 더는 기여자가 활동하지 않을 때 필요하다. $ git remote remove paul $ git remote origin 위와 같은 방법으로 리모트 저장소를 삭제하면 해당 리모트 저장소에 관련된 추적 브랜치 정보나 모든 설정 내용도 함께 삭제된다. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/text-livechat/?trk=products_seo_search
LiveChat | LinkedIn Skip to main content LinkedIn Text in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in LiveChat Live Chat Software by Text See who's skilled in this Add as skill Get started Report this product About LiveChat - premium live chat software for business. Over 37 000 companies from 150 countries use LiveChat. Try now, chat for free! Media Products media viewer No more previous content Setting up LiveChat In this tutorial, we will walk you through the process of setting up LiveChat, exploring the agent app, and installing LiveChat on your website. By the end of this lesson, you'll be ready to engage with your customers in real-time. Engaging Customers with Chat Check out how to effectively use chat campaigns, pre-chat surveys, customer data, and CRM integrations to enhance your customer interactions in LiveChat. Managing Customers and Agents in LiveChat Discover LiveChat's powerful traffic and team sections, covering real-time visitor tracking, personalized messaging, and effective agent management. Learn to set up filters, create custom segments, and manage agent roles and groups to streamline support operations. Discover 200+ LiveChat integrations Explore the powerful integrations in LiveChat's App Marketplace and the convenience of its iOS and Android mobile apps for on-the-go support. Integrate communication channels like Instagram, Facebook Messenger, WhatsApp, and e-commerce plugins like Shopify for streamlined customer engagement and productivity! Archives and Reports at LiveChat Discover LiveChat's Archives and Reports sections to analyze, organize, and optimize your customer support. Learn to navigate past conversations, use tags, and download chat transcripts in the Archives section, and gain insights into performance metrics like chat volume and customer satisfaction in the Reports section. Export reports as CSV files for further analysis and create custom reports to enhance your support strategies and boost business success. No more next content Featured customers of LiveChat McDonald's Restaurants 2,378,938 followers Valley Driving School Truck Transportation 383 followers Atlassian Software Development 2,340,591 followers Henkel Manufacturing 1,697,595 followers KPMG Accounting 5,197,788 followers Virgin Venture Capital and Private Equity Principals 320,544 followers Adobe Software Development 5,252,483 followers Better Business Bureau Technology, Information and Media 27,501 followers Unilever Manufacturing 20,625,355 followers Lexus Motor Vehicle Manufacturing 137,581 followers Show more Show less Similar products Freshchat Freshchat Live Chat Software ChatR ChatR Live Chat Software ICE Chat ICE Chat Live Chat Software Webchat Webchat Live Chat Software Microsoft Teams Microsoft Teams Live Chat Software Chat Chat Live Chat Software Sign in to see more Show more Show less Text products ChatBot ChatBot Chatbot Platforms Software HelpDesk HelpDesk Help Desk Software OpenWidget OpenWidget Text Text Live Chat Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.netscout.com/product/arbor-cloud
Cloud DDoS Protection Service - On Demand DDoS Protection Skip to main content myNETSCOUT Login Solutions ← Back Enterprise Communication Service Provider By Industry Technology Partners Enterprise Observability nGenius Solutions for Observability Network Performance Management Application Performance Management Remote Site Observability Cloud Performance Monitoring VPN/VDI SaaS AIOps AIOps for Enterprises Collaboration UC&C and UCaaS Contact Centers Cybersecurity Network Detection and Response NetOps and SecOps Collaboration DDoS Protection Adaptive DDoS Protection Cybersecurity Stack Optimization DNS Infrastructure Protection 5G Transformation 5G for Enterprise Private 4G/5G Communication Service Provider AIOps AIOps for Communication Service Providers GeoMarketing Mobile Networks 5G 5G Fixed Wireless Access LTE/VoLTE/VoWi-Fi RAN Planning, Optimization, and Automation Cybersecurity & DDoS Protection Arbor DDoS Protection Carpet Bombing DDoS Protection DNS Infrastructure Protection Mobile Network Security Network Detection and Response Cloud and Digital Transformation Cloud Visibility Internet of Things (IoT) Network Function Virtualization (NFV) OTT Visibility Cable/MSO and Fixed Networks Cable/MSO and Fixed Networks Fiber (FTTx) and DAA Monitoring Analytics Automated Analytics Business Analytics By Industry Banking Capital Markets Education Government Healthcare Insurance Manufacturing Pharmaceutical Retail Telecommunications Transportation Utilities Featured Industry: Healthcare Optimize productivity, enhance the patient care experience, and improve bottom-line with powerful healthcare IT solutions. Learn more Technology Partners AWS Cisco Systems Citrix Dell Technologies F5 Google Cloud Microsoft Oracle Palo Alto Networks Red Hat ServiceNow Splunk VMware Products ← Back Enterprise Network Security Arbor DDoS Protection Communication Service Provider AIOps Enterprise nGeniusONE for Enterprise Enable real-time visibility into any enterprise network infrastructure environment to improve performance and user experience. InfiniStreamNG Capture, store, and analyze packet-level data in real-time to help troubleshoot issues and optimize performance across IT infrastructures. nGenius Edge Sensor Help improve user experience at remote sites and other network edges to achieve holistically strong network and application performance. vSTREAM Improve visibility into virtualized and cloud environments with deep packet inspection to further performance management and security. nGeniusPULSE Continuously test and track the performance of applications, network services, and user experience across all IT infrastructures. nGenius Packet Switches Packet flow switch architectures can combine with other NETSCOUT solutions to enable pervasive visibility across network environments. Network Security Omnis Cyber Intelligence NDR Leverage packet-level insights to power network detection and response (NDR) solutions and cyber threat hunting. Omnis CyberStream Detect, investigate, and respond to cyberthreats in real-time by leveraging packet-level data to identify suspicious activity across IT environments. nGenius Decryption Appliance Decrypt SSL and TLS-encrypted traffic to enable visibility into hidden threats within network traffic. Arbor DDoS Protection Arbor Sightline Leverage network intelligence to detect, analyze, and respond to DDoS threats across the most complex service provider and enterprise networks. Arbor Sightline Mobile Safeguard mobile services and ensure network performance and security with a DDoS protection solution specifically designed for mobile networks. Arbor Threat Mitigation System Detect, mitigate, and prevent DDoS attacks with real-time traffic analysis and automated attack mitigation. Arbor Edge Defense (AED) Powerful AI & ML powered DDoS protection backed by unmatched actionable threat intelligence maintains availability to critical services and applications. Arbor Global DDoS Threat Intelligence NETSCOUT global DDoS threat intelligence provides insights into the most current cyberthreats and their tactics, enhancing products and solutions. Arbor Cloud Cloud DDoS protection that defends against large-scale, volumetric attacks. Arbor Managed Services Round-the-clock support and assistance from Arbor product experts is available with managed services. Communication Service Provider nGenius Business Analytics Drive better business decisions at scale with efficient data collection, enrichment, and third-party application feeds. nGenius Session Analyzer Isolate and resolve congestion and interference problems contributing to 70% of all subscriber issues. nGeniusONE for Carrier Assure a premium user experience with end-to-end visibility across network domains, service enablers, and devices. InfiniStreamNG Access real-time views of RAN control plane for end-to-end visibility across physical and virtual network environments. ISNG-RAN Prioritize the mobile subscriber experience with scalable, cost-effective and cloud-ready RAN network monitoring. Omnis RAN Stay ahead of capacity and mobility targets with distinct automation modules to address growing RAN complexities. TrueCall Improve your understanding of network capacity and complexity with true user performance insights. AIOps Omnis AI Sensor Unparalleled telemetry insights at enterprise speed and scale across your entire IT ecosystem. Omnis AI Streamer Build data pipelines that sift through the noise to improve the efficiency, reliability, and scalability of IT operations and AI/ML workflows. Support & Resources ← Back Blogs, Resources, & Webinars Intelligence & Reports Support & Training Blogs, Resources, & Webinars Blog Stay updated with the latest insights, tips, and news on our blog. Resource Library Explore the latest case studies, whitepapers, eBooks, and more to discover real-world applications of our products and solutions. Learning Center Quickly learn the definitions of key phrases in networking, security, and DDoS. Webinars Gain insights from product and industry experts through our live & on-demand webinars. NETSCOUT Blog Whether you're a seasoned network performance professional or a novice in the industry, you're guaranteed to learn something new from the NETSCOUT blog. Read the blog Intelligence & Reports ASERT Threat Intelligence Team Meet the ASERT team, the leading experts on DDoS and cyber threat intelligence. DDoS Threat Intelligence Report Learn about the latest DDoS trends and statistics powered by NETSCOUT’s industry-leading global internet traffic visibility. Cyber Threat Horizon Look at live DDoS attack maps to discover what attacks are occurring in real-time. DDoS Threat Intelligence Report NETSCOUT’s latest DDoS Threat Intelligence dives into the most recent trends in DDoS adversarial tactics. Get the report Support & Training Overview Get support for your products and solutions to maximize the value of your investment. My.NETSCOUT Existing customers, log in to MyNETSCOUT to access additional assets, resources, and support channels. NETSCOUT University Our mission is to help you and your team to successfully implement NETSCOUT’s solutions so that your network functions seamlessly and effectively. NO LATENCY Explore tech tips, use cases, and more in our NO LATENCY newsletter. MasterCare Support NETSCOUT's award-winning global support service. Learn more Company ← Back About Us News & Events Partners About Us About Us Board of Directors Careers Customer Reviews Environmental, Social, and Governance Executive Team Industry Analyst Perspectives Investor Relations The Future of Intelligence The quality of your AI outcomes hinges on the data available. With NETSCOUT's Smart Data, ensure you have the network visibility you need to overcome the challenges of tomorrow. NETSCOUT provides the Data that Drives You. Watch the video News & Events Events Newsroom Press Releases Read the latest NETSCOUT news to stay informed on the latest updates, press, and more. Read the latest news Partners Partner Portal Become a Partner Learn about our NETSCOUT Connect 360 partner program. Search Under Attack? Contact Us Under Attack? Home Products Arbor Cloud DDoS Protection Services Arbor Cloud DDoS Protection Services Fully Managed, Intelligently Automated On-Demand DDoS Protection Overview Overview Benefits Features Resources Related Pages Proven Cloud DDoS Mitigation Provider NETSCOUT’s Arbor hybrid DDoS protection solution using Arbor Cloud employs an adaptive DDoS attack mitigation mechanism that operates both from data centers and the cloud. This approach is designed to ensure comprehensive protection against various DDoS threats. These services specifically counter high-volume attacks targeting bandwidth, "low and slow" attacks aimed at applications and infrastructure, and simultaneous multi-vector attacks. With Arbor Cloud, inbound traffic is routed via BGP or DNS through 16 DDoS scrubbing centers with over 15Tbps capacity. Even the largest attacks are well mitigated, and only clean traffic is returned to the internet access links and servers. Benefits Purpose Built Global DDoS Protection Arbor Cloud provides over 15 Tbps of DDoS attack mitigation capacity via 16 worldwide scrubbing centers located in Asia, Europe and The Americas. Seamless Hybrid Protection to Industry Best Practices On-premise protection guards against state-exhausting and application attacks aimed at your infrastructure in tandem with Arbor Cloud. It protects against attacks that target Firewalls, Intrusion Prevention Systems (IPS) and business-critical applications. Expertise The Arbor Cloud DDoS protection service is supported by NETSCOUT’s 24x7 ASERT Team. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Sub Minute Mitigation SLA With the intelligent automation capabilities of Arbor Cloud, we can provide mitigation initiation within 60-seconds of detecting an attack via cloud signaling from your AED, via flow-detection or as part of your always-on DDoS solution. Click to enlarge image Arbor Cloud Managed DDoS Protection Service for Enterprises Arbor Cloud combines on-premise DDoS defense with cloud-based traffic scrubbing services that are tightly integrated via an automated cloud signal. This multi-layered, adaptive DDoS protection approach is a proven industry best practice and is the only way to mitigate today's full spectrum of DDoS threats for both Service Providers and Enterprises, all from a single cloud DDoS protection provider. DDoS Scrubbing Capabilities NETSCOUT's Arbor Cloud is fueled by 16 DDoS scrubbing centers worldwide, boasting a capacity of more than 15 Tbps. These scrubbing centers are located across the globe to best protect against attacks originating from any corner of the world. This industry-leading capability helps monitor and maintain availability in the face of large-scale volumetric DDoS attacks . Solution Brief Why NETSCOUT's Arbor DDoS Attack Protection Solution is Better NETSCOUT Arbor has the industry’s broadest portfolio of DDoS attack protection products and services that enable organizations of any size to customize a solution to meet their technical and financial requirements of today – and the future. Learn more Features Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Using stateless packet-processing technology and/or cloud-based IP flow analysis, DDoS attacks can be automatically detected and routed to Arbor Cloud global DDoS scrubbing centers for mitigation. Automated DDoS Attack Detection and Mitigation A single solution offering carrier-agnostic, global on-demand DDoS protection, backed by world-class security intelligence and industry-leading on-premises and global cloud DDoS protection services. Click to enlarge image Arbor Cloud brings the availability of your network, services, and applications back under your control.  Multi-Tbps of In-Cloud Protection 16 worldwide DDoS scrubbing centers with over 15 Tbps of network mitigation capacity offers comprehensive global protection from the largest DDoS attacks. All supported by NETSCOUT’s 24x7 Security Operations Center staffed by our cloud DDoS protection experts. Click to enlarge image It can also be employed in an intelligent combination of in-cloud and on-premise hybrid DDoS protection, providing you the flexibility to design a comprehensive DDoS mitigation solution that fits your environment. Cloud Only and/or Hybrid Protection Arbor Cloud DDoS protection services can be deployed as a stand-alone cloud-only solution that can be invoked on-demand upon an attack or combined with on-premises Arbor Edge Defense . Click to enlarge image ATLAS Security Engineering & Response Team (ASERT) delivers world-class network security research and analysis for the benefit of today’s enterprise and network operators. ASERT engineers and researchers are part of an elite group of institutions that are referred to as ‘super remediators’ and represent the best in information security. Powered by Global Threat Intelligence Arbor Cloud's on-demand DDoS protection is continuously armed with the latest global threat intelligence from ATLAS and ATLAS Security Engineering & Response Team (ASERT). Using automated malware analysis pipelines, sinkholes, scanners, honeypots, open-source intelligence data sets and ASERT analysis, we can provide a unique view in the threat landscape.  Incident Management Arbor Cloud DDoS experts provide real-time feedback to you using a comprehensive ticketing system. Data Sheet NETSCOUT Arbor Cloud for Service Providers Augment your local mitigation capacity and staff with Arbor Cloud for Service Providers. Learn more Data Sheet NETSCOUT Arbor Cloud DDoS Protection for Enterprises NETSCOUT Arbor Cloud combines on-premise AED DDoS defense with global cloud-based traffic scrubbing services for best-of-breed, hybrid, DDoS attack protection. Read More Featured Resources Case Study Globe Telecom Protects Critical Infrastructure and Customer Experience with NETSCOUT Arbor DDoS Attack Protection Globe Telecom struggled with DDoS attacks due to a lack of real-time systems for detection and prevention. This caused multiple issues for the company. Report SPARK Matrix for Distributed Denial of Service (DDoS) Mitigation, Q3 2025 NETSCOUT Named both a Technology Leader and Ace Performer in DDoS Mitigation for 2025 White Paper IDC - Security Requirements Driving Comprehensive DDoS Protection Security buyers must maintain a vigilant approach toward assessing the state of their defenses—an incomplete strategy can be as costly of a mistake as a reactive strategy. White Paper EMA PRISM Report – DDoS Mitigation Solutions NETSCOUT Solution Profile Quick Look Healthcare Organizations are Increasingly Targets of DDoS Attacks Healthcare Organization Uses NETSCOUT Solution To Solve Cybersecurity Challenges by Stopping DDoS Attacks Impacting Availability of Services Solution Brief On-Premise Protection is the Best First Step Against DDoS and Cyberattacks for Academic Institutions Due to the complexity of DDoS attacks, Educational Institutions need to understand the risk reduction that on-premise edge protection can provide to their on line learning... Use Case Using Arbor Cloud to Protect Multi-Cloud and Hybrid Environments Against DDoS Attacks Arbor Cloud provides a superior means for protecting critical assets hosted in a single or multi-cloud environment, offering several advantages over less capable protection... Use Case Defending Government Agencies Against High-Volume Multi-Vector DDoS Attacks Attacks are often waged against national security infrastructure. NETSCOUT has observed a massive increase from the previous half year in attacks against the US national security... View more resources Related Pages Product ATLAS Intelligence Feed (AIF) ATLAS Intelligence Feed (AIF) provides up-to-date threat intelligence on the latest DDoS threats to our DDoS protection products. Automate defenses against trending attacks with AIF. Learn more Product Arbor Edge Defense (AED) Arbor Edge Defense (AED) is an inline security appliance deployed at the network perimeter, between the internet router and firewall. This AI and ML-powered solution provides stateless, on-premise DDoS protection. Learn more Product Arbor Sightline DDoS Attack Detection Solution Whether you are a service provider or have a complex enterprise network, Arbor Sightline helps monitor and identify networking and security issues at any network scale with AI and ML-powered insights. Learn more Product Arbor Threat Mitigation System for DDoS Attacks Filter out malicious traffic while allowing legitimate traffic to pass through, at any scale, with Arbor Threat Mitigation System (TMS). Learn more Contact NETSCOUT Today To learn more about how NETSCOUT can help your company, speak to one of our highly experienced subject matter experts. Contact Us Guardians of the Connected World ® Discover Our Data Observability Solutions Service Provider Service Assurance AIOps Omnis Network Security Arbor DDoS Protection All Solutions All Products Learn Blog Resource Library Learning Center MyNETSCOUT Events Webinars About About Us Careers Environmental, Social, and Governance For Investors Newsroom Partners Connect 1-888-357-7667 Contact Us Under DDoS Attack? © 2026 NETSCOUT Terms & Conditions Website Terms of Use Data Privacy and Trust Center Transparency in Supply Chains Privacy Policy Vendor Portal Do Not Sell or Share My Personal Information California Privacy Notice UK Disclosure
2026-01-13T09:29:19
https://git-scm.com/book/ms/v2/Git-Basics-Getting-a-Git-Repository
Git - Getting a Git Repository About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Getting Started 1.1 About Version Control 1.2 A Short History of Git 1.3 What is Git? 1.4 The Command Line 1.5 Installing Git 1.6 First-Time Git Setup 1.7 Getting Help 1.8 Summary 2. Git Basics 2.1 Getting a Git Repository 2.2 Recording Changes to the Repository 2.3 Viewing the Commit History 2.4 Undoing Things 2.5 Working with Remotes 2.6 Tagging 2.7 Git Aliases 2.8 Summary 3. Git Branching 3.1 Branches in a Nutshell 3.2 Basic Branching and Merging 3.3 Branch Management 3.4 Branching Workflows 3.5 Remote Branches 3.6 Rebasing 3.7 Summary 4. Git on the Server 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Generating Your SSH Public Key 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Summary 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.1 Git Basics - Getting a Git Repository If you can read only one chapter to get going with Git, this is it. This chapter covers every basic command you need to do the vast majority of the things you’ll eventually spend your time doing with Git. By the end of the chapter, you should be able to configure and initialize a repository, begin and stop tracking files, and stage and commit changes. We’ll also show you how to set up Git to ignore certain files and file patterns, how to undo mistakes quickly and easily, how to browse the history of your project and view changes between commits, and how to push and pull from remote repositories. Getting a Git Repository You typically obtain a Git repository in one of two ways: You can take a local directory that is currently not under version control, and turn it into a Git repository, or You can clone an existing Git repository from elsewhere. In either case, you end up with a Git repository on your local machine, ready for work. Initializing a Repository in an Existing Directory If you have a project directory that is currently not under version control and you want to start controlling it with Git, you first need to go to that project’s directory. If you’ve never done this, it looks a little different depending on which system you’re running: for Linux: $ cd /home/user/my_project for macOS: $ cd /Users/user/my_project for Windows: $ cd C:/Users/user/my_project and type: $ git init This creates a new subdirectory named .git that contains all of your necessary repository files — a Git repository skeleton. At this point, nothing in your project is tracked yet. See Git Internals for more information about exactly what files are contained in the .git directory you just created. If you want to start version-controlling existing files (as opposed to an empty directory), you should probably begin tracking those files and do an initial commit. You can accomplish that with a few git add commands that specify the files you want to track, followed by a git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' We’ll go over what these commands do in just a minute. At this point, you have a Git repository with tracked files and an initial commit. Cloning an Existing Repository If you want to get a copy of an existing Git repository — for example, a project you’d like to contribute to — the command you need is git clone . If you’re familiar with other VCSs such as Subversion, you’ll notice that the command is "clone" and not "checkout". This is an important distinction — instead of getting just a working copy, Git receives a full copy of nearly all data that the server has. Every version of every file for the history of the project is pulled down by default when you run git clone . In fact, if your server disk gets corrupted, you can often use nearly any of the clones on any client to set the server back to the state it was in when it was cloned (you may lose some server-side hooks and such, but all the versioned data would be there — see Getting Git on a Server for more details). You clone a repository with git clone <url> . For example, if you want to clone the Git linkable library called libgit2 , you can do so like this: $ git clone https://github.com/libgit2/libgit2 That creates a directory named libgit2 , initializes a .git directory inside it, pulls down all the data for that repository, and checks out a working copy of the latest version. If you go into the new libgit2 directory that was just created, you’ll see the project files in there, ready to be worked on or used. If you want to clone the repository into a directory named something other than libgit2 , you can specify the new directory name as an additional argument: $ git clone https://github.com/libgit2/libgit2 mylibgit That command does the same thing as the previous one, but the target directory is called mylibgit . Git has a number of different transfer protocols you can use. The previous example uses the https:// protocol, but you may also see git:// or user@server:path/to/repo.git , which uses the SSH transfer protocol. Getting Git on a Server will introduce all of the available options the server can set up to access your Git repository and the pros and cons of each. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/bmc-software-controlm/
Control-M | LinkedIn Skip to main content LinkedIn BMC Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Control-M Workload Automation Software by BMC Software See who's skilled in this Add as skill Learn more Report this product About Transform your business with award-winning application and data workflow orchestration. Control-M simplifies application and data workflow orchestration on premises or as a service. It makes it easy to build, define, schedule, manage, and monitor production workflows, ensuring visibility and reliability and improving service level agreements (SLAs). Take advantage of: - Control-M SaaS or - Control-M for Mainframe This product is intended for Developer Director of Information Technology Development Manager Application Development Manager Information Technology Operations Manager Data Architect Information Technology Infrastructure Manager Cloud Architect DevOps Engineer Big Data Analyst Media Products media viewer No more previous content See How To Optimize and Automate Workflows Lack of integration between business applications can result in confusion and increased costs. Control-M provides a single point of control improving integration between applications and optimizing workloads within applications, enabling them to perform more efficiently. Learn more about application integrations and Control-M at bmc.com/control-m-integrations Control-M: Harness the Power of the Cloud The integration, automation, and orchestration platform, Control-M, enables businesses to harness the power of the cloud to automate the management and delivery of critical business services. Watch to learn how Control-M is the magic solution you need to thrive in a multi-cloud world. Control-M: The Secret Sauce to Big Data Success While there are many factors that contribute to the success of a big project, the secret sauce is automation in driving quick time to market and building a reliable and scalable solution driven by big data. BMC’s Control-M is an enterprise grade automation solution that can automate every aspect of the big data journey. Control-M's native support for Hadoop, Spark and NoSQL allows for seamless processing of the ingested data and managing all processes from any application from a single pane of DevOps: Don’t Get Sidetracked by What’s in Your Blind Spot (Jobs as Code) Still challenged with a DevOps blind spot in your side view mirror? Learn how a Jobs-as-Code with Control-M can modernize your traditional job and workload schedule automation by reducing costs and increasing productivity while accelerating your app dev delivery. No more next content Featured customers of Control-M Raymond James Financial Services 217,486 followers SAP Software Development 4,679,809 followers Itaú Chile Banking 453,254 followers REWE digital Software Development 30,583 followers Carrefour Retail 1,370,253 followers Amadeus IT Services and IT Consulting 516,845 followers Navistar Inc Motor Vehicle Manufacturing 142,066 followers Management Science Associates, Inc. Market Research 5,250 followers AG Insurance Insurance 51,304 followers Unum Insurance 96,511 followers CARFAX Software Development 24,673 followers Coop Retail 173,677 followers Colruyt France Retail Groceries 8,303 followers Railinc Corp. IT Services and IT Consulting 7,097 followers ING Banking 617,583 followers Sky Italia Broadcast Media Production and Distribution 372,262 followers W&W-Gruppe Financial Services 9,698 followers Domino's Restaurants 507,168 followers CMC Markets Financial Services 63,547 followers Sí Vale Financial Services 30,506 followers The Hershey Company Manufacturing 595,354 followers Show more Show less Similar products RunMyJobs by Redwood RunMyJobs by Redwood Workload Automation Software ActiveBatch by Redwood ActiveBatch by Redwood Workload Automation Software Altair PBS Professional Altair PBS Professional Workload Automation Software Stonebranch Stonebranch Workload Automation Software Tidal by Redwood Tidal by Redwood Workload Automation Software Odin Automator Odin Automator Workload Automation Software Sign in to see more Show more Show less BMC Software products BMC AMI (Mainframe Transformation) BMC AMI (Mainframe Transformation) Mainframe Management Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT2jynv_wZlhhh0Kc6i7-3C09sWVWSFDLQsCv-oL323p5C3FDdb4_-W1Vj3y34PJEC89ywL62_PxmBRaMAr9S-DSfPKk_u5S67oaU3b74ETfMROUGkfvPlPCP9NB_XMRkVyiX1SVZSj_jsq4
Facebook Facebook 이메일 또는 휴대폰 비밀번호 계정을 잊으셨나요? 새 계정 만들기 일시적으로 차단됨 일시적으로 차단됨 회원님의 이 기능 사용 속도가 너무 빠른 것 같습니다. 이 기능 사용에서 일시적으로 차단되었습니다. Back 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch 가입하기 로그인 Messenger Facebook Lite 동영상 Meta Pay Meta 스토어 Meta Quest Ray-Ban Meta Meta AI Meta AI 콘텐츠 더 보기 Instagram Threads 투표 정보 센터 개인정보처리방침 개인정보 보호 센터 정보 광고 만들기 페이지 만들기 개발자 채용 정보 쿠키 AdChoices 이용 약관 고객 센터 연락처 업로드 및 비사용자 설정 활동 로그 Meta © 2026
2026-01-13T09:29:19
https://www.linkedin.com/products/bmc-software-mainframe/
BMC AMI (Mainframe Transformation) | LinkedIn Skip to main content LinkedIn BMC Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in BMC AMI (Mainframe Transformation) Mainframe Management Software by BMC Software See who's skilled in this Add as skill Learn more Report this product About BMC Automated Mainframe Intelligence (AMI) solutions leverage AI, machine learning, and predictive analytics to achieve a self-managing mainframe. Take advantage of: - AMI Cloud - AMI Data for DB2 - AMI Data for IMS - AMI DevX - AMI Ops - AMI Security Featured customers of BMC AMI (Mainframe Transformation) CUA Financial Services 4,780 followers Sichuan Rural Credit Union & Cooperative Bank Accounting 127 followers Iberdrola Electric Power Generation 582,243 followers PT Bank Central Asia Tbk (BCA) Banking 709,265 followers ATPCO Airlines and Aviation 27,831 followers Dillard's Retail 77,946 followers HUK-COBURG Insurance 22,076 followers CPT Global Information Technology & Services 9,833 followers Show more Show less Similar products HCL Mainframe Solutions HCL Mainframe Solutions Mainframe Management Software Vertali ZTRUST for SMP/e Vertali ZTRUST for SMP/e Mainframe Management Software Vertali ZTRUST Password Synchronisation Vertali ZTRUST Password Synchronisation Mainframe Management Software Tectia® SSH Server for IBM z/OS Tectia® SSH Server for IBM z/OS Mainframe Management Software Lozen™ Connector for MuleSoft Lozen™ Connector for MuleSoft Mainframe Management Software iCAN iCAN Mainframe Management Software Sign in to see more Show more Show less BMC Software products Control-M Control-M Workload Automation Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/experian-datashare/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
Experian DataShare | LinkedIn Skip to main content LinkedIn Experian in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Experian DataShare Customer Relationship Management (CRM) Software by Experian See who's skilled in this Add as skill Learn more Report this product About Experian DataShare will accelerate new client onboarding and help increase retention through a simplified user experience. The platform integrates with over 100 versions of cloud and on-premise packages to deliver unparalleled accounting connectivity, flexible platform deployment, and instant audit or credit intelligence. It supports 100% of desktop and online accounting packages... including document uploading. Clients get faster decisions and have technical support if they experience any issues during their onboarding journey. Experian DataShare creates a digital pathway that removes the friction associated with the document gathering process. Start your digital journey to collect, format and interpret your client financials today! Media Products media viewer No more previous content Experian DataShare Experian DataShare facilitates access to permissioned data, by creating a digital client experience that delivers faster time to cash through automation, eliminating the need to reach out for updates. This solution enables you to migrate 100% of online and offline accounting packages, including document upload, for automated data retrieval. So you can generate standardized reports, all accessible via our web portal or API. No more next content Similar products Sales Cloud Sales Cloud Customer Relationship Management (CRM) Software Zoho CRM Zoho CRM Customer Relationship Management (CRM) Software Bigin by Zoho CRM Bigin by Zoho CRM Customer Relationship Management (CRM) Software Odoo CRM Odoo CRM Customer Relationship Management (CRM) Software Freshsales Freshsales Customer Relationship Management (CRM) Software Clickup Clickup Customer Relationship Management (CRM) Software Sign in to see more Show more Show less Experian products Aperture Data Studio Aperture Data Studio Data Quality Software Ascend Commercial Suite Ascend Commercial Suite Business Intelligence (BI) Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/ovhgroup-ovh-antiddos-protection/?trk=products_seo_search
OVHcloud Anti-DDoS Protection | LinkedIn Skip to main content LinkedIn OVHcloud in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in OVHcloud Anti-DDoS Protection DDoS Protection Software by OVHcloud See who's skilled in this Add as skill Learn more Report this product About Mitigate your risk of DDoS attacks and keep your services available at all times. Free with every OVH service. Learn more. Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less OVHcloud products OVHcloud Block Storage OVHcloud Block Storage Block Storage Software OVHcloud Marketplace OVHcloud Marketplace Marketplace Platforms OVHcloud Object Storage OVHcloud Object Storage Object Storage Software OVHcloud Zerto DRP OVHcloud Zerto DRP Disaster Recovery as a Service (DRaaS) Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/uz/v2/Git-%d0%b0%d1%81%d0%be%d1%81%d0%bb%d0%b0%d1%80%d0%b8-%d0%a2%d0%b0%d0%bc%d2%93%d0%b0%d0%bb%d0%b0%d1%88
Git - Тамғалаш About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Иш бошланиши 1.1 Талқинларни бошқариш ҳақида 1.2 Git нинг қисқача тарихи 1.3 Git асоси 1.4 Командалар сатри 1.5 Git ни ўрнатиш 1.6 Git да биринчи созлашлар 1.7 Қандай ёрдам олиш мумкин? 1.8 Хулосалар 2. Git асослари 2.1 Git омборини яратиш 2.2 Ўзгаришларни омборга ёзиш 2.3 Фиксирлашлар тарихини кўриш 2.4 Ўзгаришларни бекор қилиш 2.5 Узоқ масофадаги омборлар билан ишлаш 2.6 Тамғалаш 2.7 Git да таҳаллуслар 2.8 Хулоса 3. Git да тармоқланиш 3.1 Тармоқланиш ҳақида икки оғиз сўз 3.2 Тармоқланиш ва бирлашиш асослари 3.3 Тармоқларни бошқариш 3.4 Иш жараёнларини тармоқлаш 3.5 Узоқ масофадаги тармоқлар 3.6 Қайта асосланиш 3.7 Хулосалар 4. Git серверда 4.1 The Protocols 4.2 Getting Git on a Server 4.3 Sizning SSH ochiq (public) kalitingizni generatsiyalash 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Хулосалар 5. Distributed Git 5.1 Distributed Workflows 5.2 Contributing to a Project 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Qism modullar (Submodule) 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.6 Git асослари - Тамғалаш Тамғалаш Кўпгина ТБТ лар каби Git аниқ бир вазиятларни тарихдаги муҳимлар сифатида белгилаб кетишни билади. Қоидага асосан ушбу функционалдан талқинларни йўлга қўйишда фойдаланишади. Ушбу бўлимда сиз мавжуд тамғаларни (tag) қандай кўриш ва янгиларини қандай яратишни ўрганасиз. Сиз шунингдек, тамғалар тури ўзида нималарни мужассамлаштиришини билиб оласиз. Тамғаларни кўриш Git да мавжуд тамғаларни кўриш онсонлик билан бажарилади. git tag териш етарли: $ git tag v0.1 v1.3 Уушбу команда тамғаларни алфавит тарзда тартибланган ҳолда намойиш этади; уларни қай бири биринчи яратилганлигини аҳамияти йўқ. Тамғаларни қидириш учун шаблонларни қўллашингиз мумкин. Масалан Git омбори 500 дан зиёд тамғаларга эга. Агар сизни фақат 1.8.5 тамғалар кетма кетлиги қизиқтирса сиз қуйидагини бажаришингиз мумкин: $ git tag -l 'v1.8.5*' v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Тамғаларни яратиш Git икки хил турдаги тамғаларни ишлатади: енгил вазнли ва аннотацияланган. Енгил вазнли тамға – бу тармоққа ўхшаб кетади ва ўзгартирилмайдиган аниқ бир фиксирлашга кўрсаткич. Аннотацияланган тамғалар Git нинг маълумотлар омборида тўлиқ бир объектлар тарзида сақланади. Улар текширув йиғиндисига эга бўлишади, тамғани қўйганни номини, e-mail ва санани сақлайди, шарҳларга эга бўлиши, имзоланган ва GNU Privacy Guard (GPG) ёрдамида текширилиши мумкин. Одатда юқорида санаб ўтилган маълумотларга эга бўлиш учун аннотацияланган тамғалар яратиш тавсия этилади. Лекин сиз вақтинчаликка тамға яратмоқчи бўлсангиз ёки қайсидир сабабга кўра бошқа маълумотларни сақлашни хоҳламасангиз у ҳолда енгил вазнли тамғаларни яратиш ҳам етарли. Аннотацияланган тамғалар Git да аннотацияланган тамғалар енгилгина яратилади. Энг содда услуб tag командаси бажарилаётган вақтда -a параметрини кўрсатиш: $ git tag -a v1.4 -m 'my version 1.4' $ git tag v0.1 v1.3 v1.4 -m опцияси тамға билан сақланадиган тамғага тегишли ҳабарни киритиш имконини беради. Агар аннотацияланган тамғага ҳабар киритилмаса Git уни киритишингиз учун таҳрирловчини ишга туширади. Сиз git show командаси ёрдамида тамғаларни тамғаланган фиксирлашлар билан кўришингиз мумкин: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number У ўрнатилган тамға ҳақида, тамғаланган фиксирлашлар санаси тўғрисидаги маълумотни беради ва фиксирлаш ҳақидаги маълумотдан олдин келувчи аннотацияловчи ҳабарни ҳам беради. Енгил вазнли тамғалар Фиксирлашларни тамғалашни яна бир усули бу - енгил вазнли тамғалаш. Мазмунан бу фиксирлашнинг текширув йиғиндисини файлга сақлангани - бошқа ҳеч нарса эмас. Енгил вазнли тамғани яратиш учун -a , -s ёки -m опцияларини берманг: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Бу сафар сиз агар git show командасини ушбу тамғада чақирсангиз қўшимча маълумотларни кўрмайсиз. Команда фақат тамғаланган фиксирлашни кўрсатади: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number Тамғаларни кечроқ ўрнатиш Сиз шунингдек ўтиб кетган фиксирлашларга тамғалар яратишингиз мумкин бўлади. Фараз қиламиз сизнинг фиксирлашларингиз тарихи қуйидаги кўринишда: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 beginning write support 0d52aaab4479697da7686c15f77a3d64d9165190 one more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc added a commit function 4682c3261057305bdd616e23b64b0857d832627b added a todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a started write support 9fceb02d0ae598e95dc970b74767f19372d61af8 updated rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a updated readme Энди айтайлик сиз “updated rakefile” фиксирлаши бўлган жойда v1.2 ном билан тамғалашни унутгансиз. Сиз кечроқ ҳам тамғалашни киритишингиз мумкин. Фиксирлашни тамғалаш учун уни текширув йиғиндисини (ёки уни қисмини) команда охирида келтиринг: $ git tag -a v1.2 9fceb02 Сиз фиксирлаш тамғаланганини кўришингиз мумкин: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... Тамғаларни улашиш Одатий ҳолатда git push командаси тамғаларни узоқ масофадаги серверга юбормайди. Уларни яратганингиздан сўнг тамғаларни жўнатиш ҳақида кўрсатиш керак. Ушбу жараён тармоқларни юбориш каби амалга оширилади – git push origin [тамға номи] . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Агар сизда бир қанча тамғалар бўлса ва сиз уларни бир команда билан ҳаммасини юборишни хоҳласангиз git push командасида --tags опциясини ишлатишингиз мумкин. Бундай вазиятда сизни барча тамғаларингиз юборилади агарда албатта улар серверда бор бўлмаса. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Энди кимдир сизни омборингиздан клонласа ёки pull командасини бажарса у қўшимча тарзда сиз яратган тамғаларни ҳам олади. Тамғаларни текшириш Сиз аслида тамғаларни текшира олмайсиз Git да, токи улар ҳамма жойда бўлмаса. Агар сиз сизнинг ишчи каталогингиздаги омборингизга махсус тамға сифатида талқинни қўймоқчи бўлсангиз сиз махсус тамғали тармоқ яратишингиз мумкин: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Албатта агар сиз буни қилсангиз ва фиксирлашни бажарсангиз, сизнинг version2 талқинингизда янги ўзгаришлар бўлиши мумкинлиги туфайли у сизнинг v2.0.0 тамғангиздан бирозгина фарқ қилиши мумкин, шунинг учун уни бажаришда эътиборли бўлиш керак бўлади. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/zh-tw/v2/Git-%e5%9f%ba%e7%a4%8e-%e6%a8%99%e7%b1%a4
Git - 標籤 About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. 開始 1.1 關於版本控制 1.2 Git 的簡史 1.3 Git 基礎要點 1.4 命令列 1.5 Git 安裝教學 1.6 初次設定 Git 1.7 取得說明文件 1.8 摘要 2. Git 基礎 2.1 取得一個 Git 倉儲 2.2 紀錄變更到版本庫中 2.3 檢視提交的歷史記錄 2.4 復原 2.5 與遠端協同工作 2.6 標籤 2.7 Git Aliases 2.8 總結 3. 使用 Git 分支 3.1 簡述分支 3.2 分支和合併的基本用法 3.3 分支管理 3.4 分支工作流程 3.5 遠端分支 3.6 衍合 3.7 總結 4. 伺服器上的 Git 4.1 通訊協定 4.2 在伺服器上佈署 Git 4.3 產生你的 SSH 公鑰 4.4 設定伺服器 4.5 Git 常駐程式 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 第3方 Git 託管方案 4.10 總結 5. 分散式的 Git 5.1 分散式工作流程 5.2 對專案進行貢獻 5.3 維護一個專案 5.4 Summary 6. GitHub 6.1 建立帳戶及設定 6.2 參與一個專案 6.3 維護專案 6.4 Managing an organization 6.5 Scripting GitHub 6.6 總結 7. Git 工具 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 總結 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Git Internals 10.1 Plumbing and Porcelain 10.2 Git Objects 10.3 Git References 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Environment Variables 10.9 Summary A1. 附錄 A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. 附錄 B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. 附錄 C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.6 Git 基礎 - 標籤 標籤 跟大多數的版本管理系統一樣,Git 有能力對專案歷史中比較特別的時間點貼標籤,來表示其重要性。 通常大家都會用這個功能來標出發行版本,如 v1.0 …等等。 在這個章節中,你將會學到如何列出所有的標籤,如何建立新的標籤和各種不同的標籤類型。 列出你的標籤 想要列出 Git 中所有標籤的方法非常直覺。 只要輸入 git tag 如下: $ git tag v0.1 v1.3 這個指令將依字母序列出所有標籤;雖然說標籤用什麼方式列出不是很重要。 你也可以使用特定的 pattern 來搜尋標籤。 舉例來說,在 Git 原始碼的版本庫中,已經包含了超過 500 個標籤。 如果你只想看到 1.8.5 系列的標籤,你可以執行以下指令: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 建立新的標籤 Git 主要使用兩種類型的標籤:輕量級標籤和有註解的標籤。 一個輕量級的標籤就像是一個不會移動的分支——這個標籤只會指向一個特定的提交。 然而,有註解的標籤,會在 Git 的資料庫中儲存成完整的物件。 它們將被計算校驗碼;包含貼標籤那個人的名字、電子郵件和日期;能夠紀錄一個標籤訊息;並且可以簽署及透過 GNU Privacy Guard (GPG) 驗證。 通常建議你可以建立一個有註解的標籤,以便你可以保留跟這個標籤有關的所有資訊;但是你如果只想要一個暫時的標籤,或是因為某些原因不想保留額外的資訊,你也可以只用輕量級標籤。 有註解的標籤 建立一個有註解的標籤很簡單。 最簡單的方法是在你建立標籤時,同時指定 -a 的選項如下: $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 指令中的 -m 選項後面同時指定了一個標籤訊息,這個訊息會和這個標籤一起保存。 如果你沒有為標籤指定一個訊息,那麼 Git 會開啟你的編輯器以便你輸入。 當你使用 git show 指令時,你可以查看標籤的資訊,還有這個標籤所標記的提交資訊如下: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number 這樣就可以在提交資訊前顯示出標籤的資訊、標籤被建立的時間以及標籤的訊息。 輕量級標籤 另外一種能標記提交的標籤是輕量級標籤。 這基本上是把該提交的校驗碼存在一個檔案中,不包含其他資訊。 如果想要建立一個輕量級的標籤,不要指定 -a 、 -s 或 -m 的選項如下: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 此時如果對該標籤使用 git show ,你將不會看到這個標籤的額外資訊。 這個指令就只會顯示標籤所在的提交資訊: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 changed the version number 對以前的提交貼標籤 你也可以對過去的提交貼標籤。 假設你的提交歷史看起來如下: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 beginning write support 0d52aaab4479697da7686c15f77a3d64d9165190 one more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc added a commit function 4682c3261057305bdd616e23b64b0857d832627b added a todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a started write support 9fceb02d0ae598e95dc970b74767f19372d61af8 updated rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a updated readme 現在,假設你忘記在專案的「updated rakefile」提交貼 v1.2 的標籤。 你可以在後來再補貼標籤。 要在那個提交上面貼標籤,你需要在指令後面指定那個提交的校驗碼(可以省略後半段): $ git tag -a v1.2 9fceb02 你可以看到你已經在那個提交上面貼標籤了: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 updated rakefile ... 分享標籤 git push 指令預設不會傳送標籤到遠端伺服器。 在你建立標籤後,你必須明確的要求 Git 將標籤推送到共用的伺服器上面。 這個動作就像是在分享遠端分支一樣——你可以執行 git push origin [tagname] 。 $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 如果想要一次推送很多標籤,也可以在使用 git push 指令的時候加上 --tags 選項。 這將會把你所有不在伺服器上面的標籤傳送給遠端伺服器。 $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw 現在,當其他人從版本庫克隆或拉取時,他們就能同時拿到你所貼的標籤, 檢出標籤 在 Git 中你不能真的檢出一個標籤,因為它們並不能像分支一樣四處移動。 如果你希望工作目錄和版本庫中特定的標籤版本完全一樣,你可以使用 git checkout -b [branchname] [tagname] 在該標籤上建立一個新分支: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' 當然,如果在建立新分支以後又進行了一次提交, version2 分支將會和 v2.0.0 標籤有所差異,因為這個分支已經因為你的提交而改變了,請特別小心。 prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-Git-Daemon
Git - Git Daemon About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.5 Git op de server - Git Daemon Git Daemon Nu gaan we een daemon opzetten die repositories verspreid via het “Git” protocol. Dit is een gebruikelijke keuze voor snelle, ongeauthenticerde toegang tot je Git data. Onthoud dat omdat dit geen geauthenticeerde service is, alles wat je verspreid publiek is in het netwerk. Als je deze server buiten je firewall draait, zou het alleen gebruikt moeten worden voor projecten die publiekelijk zichtbaar zijn naar de wereld. Als de server die je draait binnen je firewall staat, zou je het kunnen gebruiken voor projecten met een groot aantal mensen of computers (continue integratie of build servers) die alleen-lees-toegang hebben, waarvoor je niet een SSH sleutel wilt toevoegen voor elk van deze. In elk geval, het Git protocol is relatief simpel om op te zetten. Het enige wat je hoeft te doen is dit commando in de achtergrond (daemonize) draaien: $ git daemon --reuseaddr --base-path=/srv/git/ /srv/git/ De --reuseaddr optie staat de server toe om te herstarten zonder te wachten tot oude connecties een time-out krijgen, de --base-path optie staat mensen toe om projecten te clonen zonder het volledige pad te specificeren, en het pad aan het einde vertelt de Git daemon waar hij moet kijken voor de te exporteren repositories. Als je een firewall draait, zul je er poort 9418 open moeten zetten op de machine waar je dit op gaat doen. Je kunt dit proces op een aantal manieren daemoniseren, afhankelijk van het besturingssystem waarop je draait. Omdat systemd het meest gebruikte init systeem is onder de moderne Linux distributies, kan je dat gebruiken voor dat doel. Simpelweg een bestand in /etc/systemd/system/git-daemon.service zetten met deze inhoud: [Unit] Description=Start Git Daemon [Service] ExecStart=/usr/bin/git daemon --reuseaddr --base-path=/srv/git/ /srv/git/ Restart=always RestartSec=500ms StandardOutput=syslog StandardError=syslog SyslogIdentifier=git-daemon User=git Group=git [Install] WantedBy=multi-user.target Het zal je op zijn gevallen dat de Git daemon hier wordt gestart met git als zowel de group als user. Pas het aan aan jouw standaard en org ervoor dat de gebruiker bestaat op het systeem. Controleer ook dat de Git binary inderdaad op /usr/bin/git staat en wijzig het pad indien noodzakelijk. Als laatste roep je systemctl enable git-daemon aan om de service automatisch te starten op boot-tijd, en je kunt de service starten en stoppen met respectievelijk systemctl start git-daemon en systemctl stop git-daemon . Op andere systemen zul je misschien xinetd willen gebruiken, een script in je sysvinit systeem, of iets anders — zolang je dat commando maar ge-daemoniseerd krijgt en deze op een of andere manier in de gaten gehouden wordt. Vervolgens zul je Git moeten vertellen op welke repositories je onauthenticeerde Gitserver gebaseerde toegang toestaat. Je kunt dit doen voor elke repository door een bestand genaamd git-deamon-export-ok te maken. $ cd /path/to/project.git $ touch git-daemon-export-ok Het feit dat dit bestand er is geeft aan dat Git dat project zonder authenticatie kan verspreiden. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://huggingface.co/datasets/TeichAI/claude-4.5-opus-high-reasoning-250x
TeichAI/claude-4.5-opus-high-reasoning-250x · Datasets at Hugging Face Hugging Face Models Datasets Spaces Community Docs Enterprise Pricing Log In Sign Up Datasets: TeichAI / claude-4.5-opus-high-reasoning-250x like 161 Follow TeichAI 238 Modalities: Text Formats: json Size: < 1K Libraries: Datasets pandas Croissant + 1 Dataset card Data Studio Files Files and versions xet Community 3 The user wants me to implement a recommendation system using collaborative filtering in Python. I'll create a comprehensive implementation that covers:\n\n1. **User-Based Collaborative Filtering** - Find similar users and recommend items they liked\n2. **Item-Based Collaborative Filtering** - Find similar items and recommend based on user's past preferences\n3. **Matrix Factorization (SVD)** - Latent factor model\n\nI'll include:\n- Data preprocessing\n- Similarity calculations (cosine, Pearson correlation)\n- Neighborhood-based methods\n- Matrix factorization using SVD\n- Evaluation metrics (RMSE, MAE, Precision, Recall)\n- A practical example with movie ratings\n\nLet me create a well-structured, comprehensive implementation. # Collaborative Filtering Recommendation System in Python\n\nI'll create a comprehensive recommendation system implementation covering user-based, item-based collaborative filtering, and matrix factorization approaches.\n\n## Project Structure\n\n```\nrecommendation_system/\n├── recommendation_system.py # Core implementation\n├── matrix_factorization.py # SVD-based approach\n├── evaluation.py # Evaluation metrics\n├── demo.py # Usage examples\n└── requirements.txt # Dependencies\n```\n\n## 1. Core Recommendation System\n\n```python\n\"\"\"\nrecommendation_system.py\nCollaborative Filtering Recommendation System Implementation\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom abc import ABC, abstractmethod\nfrom collections import defaultdict\nfrom typing import List, Tuple, Dict, Optional, Set\nfrom scipy.sparse import csr_matrix\nfrom scipy.spatial.distance import cosine\nimport warnings\nwarnings.filterwarnings('ignore')\n\n\nclass SimilarityMetrics:\n \"\"\"Collection of similarity metrics for collaborative filtering.\"\"\"\n \n @staticmethod\n def cosine_similarity(vec1: np.ndarray, vec2: np.ndarray) -> float:\n \"\"\"Calculate cosine similarity between two vectors.\"\"\"\n # Handle zero vectors\n norm1 = np.linalg.norm(vec1)\n norm2 = np.linalg.norm(vec2)\n \n if norm1 == 0 or norm2 == 0:\n return 0.0\n \n return np.dot(vec1, vec2) / (norm1 * norm2)\n \n @staticmethod\n def pearson_correlation(vec1: np.ndarray, vec2: np.ndarray) -> float:\n \"\"\"Calculate Pearson correlation coefficient.\"\"\"\n # Find common rated items\n mask = (vec1 != 0) & (vec2 != 0)\n \n if mask.sum() float:\n \"\"\"Calculate adjusted cosine similarity (for item-based CF).\"\"\"\n mask = (vec1 != 0) & (vec2 != 0)\n \n if mask.sum() == 0:\n return 0.0\n \n adj_vec1 = vec1[mask] - item_means[mask]\n adj_vec2 = vec2[mask] - item_means[mask]\n \n norm1 = np.linalg.norm(adj_vec1)\n norm2 = np.linalg.norm(adj_vec2)\n \n if norm1 == 0 or norm2 == 0:\n return 0.0\n \n return np.dot(adj_vec1, adj_vec2) / (norm1 * norm2)\n \n @staticmethod\n def jaccard_similarity(set1: Set, set2: Set) -> float:\n \"\"\"Calculate Jaccard similarity for binary ratings.\"\"\"\n if not set1 or not set2:\n return 0.0\n \n intersection = len(set1 & set2)\n union = len(set1 | set2)\n \n return intersection / union if union > 0 else 0.0\n\n\nclass BaseRecommender(ABC):\n \"\"\"Abstract base class for recommendation systems.\"\"\"\n \n def __init__(self, n_neighbors: int = 20, min_support: int = 3,\n similarity_metric: str = 'cosine'):\n self.n_neighbors = n_neighbors\n self.min_support = min_support\n self.similarity_metric = similarity_metric\n self.ratings_matrix = None\n self.user_id_map = {}\n self.item_id_map = {}\n self.reverse_user_map = {}\n self.reverse_item_map = {}\n \n def _create_mappings(self, user_ids: np.ndarray, item_ids: np.ndarray):\n \"\"\"Create ID mappings for internal indexing.\"\"\"\n unique_users = np.unique(user_ids)\n unique_items = np.unique(item_ids)\n \n self.user_id_map = {uid: idx for idx, uid in enumerate(unique_users)}\n self.item_id_map = {iid: idx for idx, iid in enumerate(unique_items)}\n self.reverse_user_map = {idx: uid for uid, idx in self.user_id_map.items()}\n self.reverse_item_map = {idx: iid for iid, idx in self.item_id_map.items()}\n \n def _build_ratings_matrix(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray) -> np.ndarray:\n \"\"\"Build user-item ratings matrix.\"\"\"\n n_users = len(self.user_id_map)\n n_items = len(self.item_id_map)\n \n matrix = np.zeros((n_users, n_items))\n \n for uid, iid, rating in zip(user_ids, item_ids, ratings):\n user_idx = self.user_id_map[uid]\n item_idx = self.item_id_map[iid]\n matrix[user_idx, item_idx] = rating\n \n return matrix\n \n @abstractmethod\n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray, \n ratings: np.ndarray):\n \"\"\"Fit the recommender to the data.\"\"\"\n pass\n \n @abstractmethod\n def predict(self, user_id, item_id) -> float:\n \"\"\"Predict rating for a user-item pair.\"\"\"\n pass\n \n @abstractmethod\n def recommend(self, user_id, n_recommendations: int = 10) -> List[Tuple]:\n \"\"\"Generate recommendations for a user.\"\"\"\n pass\n\n\nclass UserBasedCF(BaseRecommender):\n \"\"\"\n User-Based Collaborative Filtering\n \n Finds similar users and recommends items that similar users liked.\n \"\"\"\n \n def __init__(self, n_neighbors: int = 20, min_support: int = 3,\n similarity_metric: str = 'pearson'):\n super().__init__(n_neighbors, min_support, similarity_metric)\n self.user_similarities = None\n self.user_means = None\n \n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray):\n \"\"\"\n Fit the user-based CF model.\n \n Args:\n user_ids: Array of user IDs\n item_ids: Array of item IDs\n ratings: Array of ratings\n \"\"\"\n print(\"Fitting User-Based Collaborative Filtering model...\")\n \n self._create_mappings(user_ids, item_ids)\n self.ratings_matrix = self._build_ratings_matrix(user_ids, item_ids, ratings)\n \n # Calculate user means (for bias correction)\n self.user_means = np.zeros(len(self.user_id_map))\n for i in range(len(self.user_id_map)):\n rated_items = self.ratings_matrix[i] != 0\n if rated_items.sum() > 0:\n self.user_means[i] = self.ratings_matrix[i, rated_items].mean()\n \n # Calculate user-user similarities\n self._compute_similarities()\n \n print(f\"Model fitted with {len(self.user_id_map)} users and \"\n f\"{len(self.item_id_map)} items\")\n \n def _compute_similarities(self):\n \"\"\"Compute all user-user similarities.\"\"\"\n n_users = len(self.user_id_map)\n self.user_similarities = np.zeros((n_users, n_users))\n \n metrics = SimilarityMetrics()\n \n for i in range(n_users):\n for j in range(i + 1, n_users):\n if self.similarity_metric == 'cosine':\n sim = metrics.cosine_similarity(\n self.ratings_matrix[i], self.ratings_matrix[j]\n )\n elif self.similarity_metric == 'pearson':\n sim = metrics.pearson_correlation(\n self.ratings_matrix[i], self.ratings_matrix[j]\n )\n else:\n sim = metrics.cosine_similarity(\n self.ratings_matrix[i], self.ratings_matrix[j]\n )\n \n self.user_similarities[i, j] = sim\n self.user_similarities[j, i] = sim\n \n def _get_neighbors(self, user_idx: int, item_idx: int) -> List[Tuple[int, float]]:\n \"\"\"Get neighbors who rated the target item.\"\"\"\n similarities = self.user_similarities[user_idx].copy()\n \n # Find users who rated this item\n rated_mask = self.ratings_matrix[:, item_idx] != 0\n rated_mask[user_idx] = False # Exclude the target user\n \n # Get valid neighbors\n neighbors = []\n for other_idx in np.where(rated_mask)[0]:\n sim = similarities[other_idx]\n if sim > 0: # Only positive correlations\n neighbors.append((other_idx, sim))\n \n # Sort by similarity and take top k\n neighbors.sort(key=lambda x: x[1], reverse=True)\n return neighbors[:self.n_neighbors]\n \n def predict(self, user_id, item_id) -> float:\n \"\"\"\n Predict rating for a user-item pair.\n \n Args:\n user_id: User identifier\n item_id: Item identifier\n \n Returns:\n Predicted rating\n \"\"\"\n if user_id not in self.user_id_map:\n return self.user_means.mean() # Global mean for unknown users\n \n if item_id not in self.item_id_map:\n return self.user_means[self.user_id_map[user_id]] # User mean\n \n user_idx = self.user_id_map[user_id]\n item_idx = self.item_id_map[item_id]\n \n # If user already rated this item\n if self.ratings_matrix[user_idx, item_idx] != 0:\n return self.ratings_matrix[user_idx, item_idx]\n \n neighbors = self._get_neighbors(user_idx, item_idx)\n \n if not neighbors:\n return self.user_means[user_idx]\n \n # Weighted average with bias correction\n numerator = 0.0\n denominator = 0.0\n \n for neighbor_idx, similarity in neighbors:\n neighbor_rating = self.ratings_matrix[neighbor_idx, item_idx]\n neighbor_mean = self.user_means[neighbor_idx]\n \n numerator += similarity * (neighbor_rating - neighbor_mean)\n denominator += abs(similarity)\n \n if denominator == 0:\n return self.user_means[user_idx]\n \n predicted = self.user_means[user_idx] + (numerator / denominator)\n \n # Clip to valid rating range\n return np.clip(predicted, 1.0, 5.0)\n \n def recommend(self, user_id, n_recommendations: int = 10,\n exclude_rated: bool = True) -> List[Tuple]:\n \"\"\"\n Generate top-N recommendations for a user.\n \n Args:\n user_id: User identifier\n n_recommendations: Number of recommendations to generate\n exclude_rated: Whether to exclude already rated items\n \n Returns:\n List of (item_id, predicted_rating) tuples\n \"\"\"\n if user_id not in self.user_id_map:\n # Return popular items for unknown users\n return self._get_popular_items(n_recommendations)\n \n user_idx = self.user_id_map[user_id]\n predictions = []\n \n for item_idx in range(len(self.item_id_map)):\n # Skip rated items if requested\n if exclude_rated and self.ratings_matrix[user_idx, item_idx] != 0:\n continue\n \n item_id = self.reverse_item_map[item_idx]\n predicted_rating = self.predict(user_id, item_id)\n predictions.append((item_id, predicted_rating))\n \n # Sort by predicted rating\n predictions.sort(key=lambda x: x[1], reverse=True)\n \n return predictions[:n_recommendations]\n \n def _get_popular_items(self, n: int) -> List[Tuple]:\n \"\"\"Get most popular items (for cold start).\"\"\"\n item_ratings = []\n \n for item_idx in range(len(self.item_id_map)):\n ratings = self.ratings_matrix[:, item_idx]\n rated = ratings[ratings != 0]\n if len(rated) >= self.min_support:\n avg_rating = rated.mean()\n item_id = self.reverse_item_map[item_idx]\n item_ratings.append((item_id, avg_rating))\n \n item_ratings.sort(key=lambda x: x[1], reverse=True)\n return item_ratings[:n]\n \n def get_similar_users(self, user_id, n: int = 10) -> List[Tuple]:\n \"\"\"Get most similar users to a given user.\"\"\"\n if user_id not in self.user_id_map:\n return []\n \n user_idx = self.user_id_map[user_id]\n similarities = self.user_similarities[user_idx]\n \n similar_users = []\n for other_idx, sim in enumerate(similarities):\n if other_idx != user_idx and sim > 0:\n other_id = self.reverse_user_map[other_idx]\n similar_users.append((other_id, sim))\n \n similar_users.sort(key=lambda x: x[1], reverse=True)\n return similar_users[:n]\n\n\nclass ItemBasedCF(BaseRecommender):\n \"\"\"\n Item-Based Collaborative Filtering\n \n Finds similar items and recommends based on user's past preferences.\n \"\"\"\n \n def __init__(self, n_neighbors: int = 20, min_support: int = 3,\n similarity_metric: str = 'adjusted_cosine'):\n super().__init__(n_neighbors, min_support, similarity_metric)\n self.item_similarities = None\n self.item_means = None\n \n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray):\n \"\"\"Fit the item-based CF model.\"\"\"\n print(\"Fitting Item-Based Collaborative Filtering model...\")\n \n self._create_mappings(user_ids, item_ids)\n self.ratings_matrix = self._build_ratings_matrix(user_ids, item_ids, ratings)\n \n # Calculate item means\n self.item_means = np.zeros(len(self.item_id_map))\n for i in range(len(self.item_id_map)):\n rated_users = self.ratings_matrix[:, i] != 0\n if rated_users.sum() > 0:\n self.item_means[i] = self.ratings_matrix[rated_users, i].mean()\n \n # Calculate user means for adjusted cosine\n self.user_means = np.zeros(len(self.user_id_map))\n for i in range(len(self.user_id_map)):\n rated_items = self.ratings_matrix[i] != 0\n if rated_items.sum() > 0:\n self.user_means[i] = self.ratings_matrix[i, rated_items].mean()\n \n # Compute item-item similarities\n self._compute_similarities()\n \n print(f\"Model fitted with {len(self.user_id_map)} users and \"\n f\"{len(self.item_id_map)} items\")\n \n def _compute_similarities(self):\n \"\"\"Compute all item-item similarities.\"\"\"\n n_items = len(self.item_id_map)\n self.item_similarities = np.zeros((n_items, n_items))\n \n metrics = SimilarityMetrics()\n \n # Transpose for item-centric view\n item_vectors = self.ratings_matrix.T\n \n for i in range(n_items):\n for j in range(i + 1, n_items):\n if self.similarity_metric == 'adjusted_cosine':\n sim = self._adjusted_cosine_items(i, j)\n elif self.similarity_metric == 'cosine':\n sim = metrics.cosine_similarity(item_vectors[i], item_vectors[j])\n else:\n sim = self._adjusted_cosine_items(i, j)\n \n self.item_similarities[i, j] = sim\n self.item_similarities[j, i] = sim\n \n def _adjusted_cosine_items(self, item_i: int, item_j: int) -> float:\n \"\"\"Calculate adjusted cosine similarity between items.\"\"\"\n # Find users who rated both items\n rated_i = self.ratings_matrix[:, item_i] != 0\n rated_j = self.ratings_matrix[:, item_j] != 0\n common_users = rated_i & rated_j\n \n if common_users.sum() List[Tuple[int, float]]:\n \"\"\"Get similar items rated by the user.\"\"\"\n similarities = self.item_similarities[item_idx].copy()\n \n # Find items rated by this user\n rated_mask = self.ratings_matrix[user_idx] != 0\n rated_mask[item_idx] = False # Exclude target item\n \n neighbors = []\n for other_idx in np.where(rated_mask)[0]:\n sim = similarities[other_idx]\n if sim > 0:\n neighbors.append((other_idx, sim))\n \n neighbors.sort(key=lambda x: x[1], reverse=True)\n return neighbors[:self.n_neighbors]\n \n def predict(self, user_id, item_id) -> float:\n \"\"\"Predict rating for a user-item pair.\"\"\"\n if user_id not in self.user_id_map:\n if item_id in self.item_id_map:\n return self.item_means[self.item_id_map[item_id]]\n return self.item_means.mean()\n \n if item_id not in self.item_id_map:\n return self.user_means[self.user_id_map[user_id]]\n \n user_idx = self.user_id_map[user_id]\n item_idx = self.item_id_map[item_id]\n \n if self.ratings_matrix[user_idx, item_idx] != 0:\n return self.ratings_matrix[user_idx, item_idx]\n \n neighbors = self._get_neighbors(user_idx, item_idx)\n \n if not neighbors:\n return self.item_means[item_idx] if self.item_means[item_idx] > 0 else 3.0\n \n # Weighted average\n numerator = 0.0\n denominator = 0.0\n \n for neighbor_idx, similarity in neighbors:\n rating = self.ratings_matrix[user_idx, neighbor_idx]\n numerator += similarity * rating\n denominator += abs(similarity)\n \n if denominator == 0:\n return self.item_means[item_idx]\n \n return np.clip(numerator / denominator, 1.0, 5.0)\n \n def recommend(self, user_id, n_recommendations: int = 10,\n exclude_rated: bool = True) -> List[Tuple]:\n \"\"\"Generate top-N recommendations for a user.\"\"\"\n if user_id not in self.user_id_map:\n return self._get_popular_items(n_recommendations)\n \n user_idx = self.user_id_map[user_id]\n predictions = []\n \n for item_idx in range(len(self.item_id_map)):\n if exclude_rated and self.ratings_matrix[user_idx, item_idx] != 0:\n continue\n \n item_id = self.reverse_item_map[item_idx]\n predicted_rating = self.predict(user_id, item_id)\n predictions.append((item_id, predicted_rating))\n \n predictions.sort(key=lambda x: x[1], reverse=True)\n return predictions[:n_recommendations]\n \n def _get_popular_items(self, n: int) -> List[Tuple]:\n \"\"\"Get most popular items.\"\"\"\n item_ratings = []\n \n for item_idx in range(len(self.item_id_map)):\n ratings = self.ratings_matrix[:, item_idx]\n rated = ratings[ratings != 0]\n if len(rated) >= self.min_support:\n avg_rating = rated.mean()\n item_id = self.reverse_item_map[item_idx]\n item_ratings.append((item_id, avg_rating))\n \n item_ratings.sort(key=lambda x: x[1], reverse=True)\n return item_ratings[:n]\n \n def get_similar_items(self, item_id, n: int = 10) -> List[Tuple]:\n \"\"\"Get most similar items to a given item.\"\"\"\n if item_id not in self.item_id_map:\n return []\n \n item_idx = self.item_id_map[item_id]\n similarities = self.item_similarities[item_idx]\n \n similar_items = []\n for other_idx, sim in enumerate(similarities):\n if other_idx != item_idx and sim > 0:\n other_id = self.reverse_item_map[other_idx]\n similar_items.append((other_id, sim))\n \n similar_items.sort(key=lambda x: x[1], reverse=True)\n return similar_items[:n]\n\n\nclass HybridCF:\n \"\"\"\n Hybrid Collaborative Filtering\n \n Combines user-based and item-based approaches.\n \"\"\"\n \n def __init__(self, user_weight: float = 0.5, n_neighbors: int = 20):\n self.user_weight = user_weight\n self.item_weight = 1 - user_weight\n self.user_cf = UserBasedCF(n_neighbors=n_neighbors)\n self.item_cf = ItemBasedCF(n_neighbors=n_neighbors)\n self.is_fitted = False\n \n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray):\n \"\"\"Fit both user and item-based models.\"\"\"\n print(\"Fitting Hybrid Collaborative Filtering model...\")\n self.user_cf.fit(user_ids, item_ids, ratings)\n self.item_cf.fit(user_ids, item_ids, ratings)\n self.is_fitted = True\n \n def predict(self, user_id, item_id) -> float:\n \"\"\"Predict using weighted combination of both methods.\"\"\"\n user_pred = self.user_cf.predict(user_id, item_id)\n item_pred = self.item_cf.predict(user_id, item_id)\n \n return self.user_weight * user_pred + self.item_weight * item_pred\n \n def recommend(self, user_id, n_recommendations: int = 10) -> List[Tuple]:\n \"\"\"Generate recommendations using hybrid approach.\"\"\"\n if user_id not in self.user_cf.user_id_map:\n return self.item_cf._get_popular_items(n_recommendations)\n \n user_idx = self.user_cf.user_id_map[user_id]\n predictions = []\n \n for item_idx in range(len(self.user_cf.item_id_map)):\n if self.user_cf.ratings_matrix[user_idx, item_idx] != 0:\n continue\n \n item_id = self.user_cf.reverse_item_map[item_idx]\n predicted_rating = self.predict(user_id, item_id)\n predictions.append((item_id, predicted_rating))\n \n predictions.sort(key=lambda x: x[1], reverse=True)\n return predictions[:n_recommendations]\n```\n\n## 2. Matrix Factorization (SVD)\n\n```python\n\"\"\"\nmatrix_factorization.py\nMatrix Factorization approaches for recommendations\n\"\"\"\n\nimport numpy as np\nfrom typing import List, Tuple, Optional\nfrom scipy.sparse import csr_matrix\nfrom scipy.sparse.linalg import svds\n\n\nclass SVDRecommender:\n \"\"\"\n Matrix Factorization using Singular Value Decomposition\n \n Decomposes the user-item matrix into latent factors.\n \"\"\"\n \n def __init__(self, n_factors: int = 50, n_epochs: int = 100,\n learning_rate: float = 0.005, regularization: float = 0.02):\n self.n_factors = n_factors\n self.n_epochs = n_epochs\n self.lr = learning_rate\n self.reg = regularization\n \n self.user_factors = None\n self.item_factors = None\n self.user_biases = None\n self.item_biases = None\n self.global_mean = None\n \n self.user_id_map = {}\n self.item_id_map = {}\n self.reverse_user_map = {}\n self.reverse_item_map = {}\n \n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray, verbose: bool = True):\n \"\"\"\n Train the SVD model using stochastic gradient descent.\n \n Args:\n user_ids: Array of user IDs\n item_ids: Array of item IDs\n ratings: Array of ratings\n verbose: Print training progress\n \"\"\"\n # Create mappings\n unique_users = np.unique(user_ids)\n unique_items = np.unique(item_ids)\n \n self.user_id_map = {uid: idx for idx, uid in enumerate(unique_users)}\n self.item_id_map = {iid: idx for idx, iid in enumerate(unique_items)}\n self.reverse_user_map = {idx: uid for uid, idx in self.user_id_map.items()}\n self.reverse_item_map = {idx: iid for iid, idx in self.item_id_map.items()}\n \n n_users = len(unique_users)\n n_items = len(unique_items)\n \n # Initialize factors randomly\n np.random.seed(42)\n self.user_factors = np.random.normal(0, 0.1, (n_users, self.n_factors))\n self.item_factors = np.random.normal(0, 0.1, (n_items, self.n_factors))\n self.user_biases = np.zeros(n_users)\n self.item_biases = np.zeros(n_items)\n self.global_mean = np.mean(ratings)\n \n # Convert IDs to indices\n user_indices = np.array([self.user_id_map[uid] for uid in user_ids])\n item_indices = np.array([self.item_id_map[iid] for iid in item_ids])\n \n if verbose:\n print(f\"Training SVD model with {n_users} users, {n_items} items, \"\n f\"{len(ratings)} ratings\")\n print(f\"Parameters: {self.n_factors} factors, {self.n_epochs} epochs\")\n \n # SGD training\n for epoch in range(self.n_epochs):\n # Shuffle data\n indices = np.random.permutation(len(ratings))\n \n total_error = 0.0\n \n for idx in indices:\n u = user_indices[idx]\n i = item_indices[idx]\n r = ratings[idx]\n \n # Predict\n pred = self.global_mean + self.user_biases[u] + self.item_biases[i]\n pred += np.dot(self.user_factors[u], self.item_factors[i])\n \n # Error\n error = r - pred\n total_error += error ** 2\n \n # Update biases\n self.user_biases[u] += self.lr * (error - self.reg * self.user_biases[u])\n self.item_biases[i] += self.lr * (error - self.reg * self.item_biases[i])\n \n # Update factors\n user_factor = self.user_factors[u].copy()\n self.user_factors[u] += self.lr * (\n error * self.item_factors[i] - self.reg * self.user_factors[u]\n )\n self.item_factors[i] += self.lr * (\n error * user_factor - self.reg * self.item_factors[i]\n )\n \n rmse = np.sqrt(total_error / len(ratings))\n \n if verbose and (epoch + 1) % 10 == 0:\n print(f\" Epoch {epoch + 1}/{self.n_epochs} - RMSE: {rmse:.4f}\")\n \n if verbose:\n print(\"Training completed!\")\n \n def predict(self, user_id, item_id) -> float:\n \"\"\"Predict rating for a user-item pair.\"\"\"\n if user_id not in self.user_id_map:\n if item_id in self.item_id_map:\n item_idx = self.item_id_map[item_id]\n return self.global_mean + self.item_biases[item_idx]\n return self.global_mean\n \n if item_id not in self.item_id_map:\n user_idx = self.user_id_map[user_id]\n return self.global_mean + self.user_biases[user_idx]\n \n user_idx = self.user_id_map[user_id]\n item_idx = self.item_id_map[item_id]\n \n pred = self.global_mean\n pred += self.user_biases[user_idx]\n pred += self.item_biases[item_idx]\n pred += np.dot(self.user_factors[user_idx], self.item_factors[item_idx])\n \n return np.clip(pred, 1.0, 5.0)\n \n def recommend(self, user_id, n_recommendations: int = 10,\n rated_items: Optional[set] = None) -> List[Tuple]:\n \"\"\"Generate top-N recommendations for a user.\"\"\"\n if user_id not in self.user_id_map:\n return self._get_popular_items(n_recommendations)\n \n user_idx = self.user_id_map[user_id]\n \n # Calculate scores for all items\n scores = (self.global_mean + \n self.user_biases[user_idx] + \n self.item_biases + \n np.dot(self.item_factors, self.user_factors[user_idx]))\n \n # Get top items\n predictions = []\n for item_idx in np.argsort(scores)[::-1]:\n item_id = self.reverse_item_map[item_idx]\n \n if rated_items and item_id in rated_items:\n continue\n \n predictions.append((item_id, float(np.clip(scores[item_idx], 1.0, 5.0))))\n \n if len(predictions) >= n_recommendations:\n break\n \n return predictions\n \n def _get_popular_items(self, n: int) -> List[Tuple]:\n \"\"\"Get items with highest biases (popularity proxy).\"\"\"\n item_scores = self.global_mean + self.item_biases\n top_indices = np.argsort(item_scores)[::-1][:n]\n \n return [(self.reverse_item_map[idx], float(np.clip(item_scores[idx], 1.0, 5.0)))\n for idx in top_indices]\n \n def get_user_embedding(self, user_id) -> Optional[np.ndarray]:\n \"\"\"Get latent factor representation for a user.\"\"\"\n if user_id not in self.user_id_map:\n return None\n return self.user_factors[self.user_id_map[user_id]]\n \n def get_item_embedding(self, item_id) -> Optional[np.ndarray]:\n \"\"\"Get latent factor representation for an item.\"\"\"\n if item_id not in self.item_id_map:\n return None\n return self.item_factors[self.item_id_map[item_id]]\n\n\nclass ALSRecommender:\n \"\"\"\n Alternating Least Squares Matrix Factorization\n \n Efficient for implicit feedback or large datasets.\n \"\"\"\n \n def __init__(self, n_factors: int = 50, n_iterations: int = 20,\n regularization: float = 0.1):\n self.n_factors = n_factors\n self.n_iterations = n_iterations\n self.reg = regularization\n \n self.user_factors = None\n self.item_factors = None\n self.ratings_matrix = None\n \n self.user_id_map = {}\n self.item_id_map = {}\n self.reverse_user_map = {}\n self.reverse_item_map = {}\n \n def fit(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray, verbose: bool = True):\n \"\"\"\n Train using Alternating Least Squares.\n \"\"\"\n # Create mappings\n unique_users = np.unique(user_ids)\n unique_items = np.unique(item_ids)\n \n self.user_id_map = {uid: idx for idx, uid in enumerate(unique_users)}\n self.item_id_map = {iid: idx for idx, iid in enumerate(unique_items)}\n self.reverse_user_map = {idx: uid for uid, idx in self.user_id_map.items()}\n self.reverse_item_map = {idx: iid for iid, idx in self.item_id_map.items()}\n \n n_users = len(unique_users)\n n_items = len(unique_items)\n \n # Build ratings matrix\n self.ratings_matrix = np.zeros((n_users, n_items))\n for uid, iid, r in zip(user_ids, item_ids, ratings):\n self.ratings_matrix[self.user_id_map[uid], self.item_id_map[iid]] = r\n \n # Initialize factors\n np.random.seed(42)\n self.user_factors = np.random.rand(n_users, self.n_factors)\n self.item_factors = np.random.rand(n_items, self.n_factors)\n \n if verbose:\n print(f\"Training ALS model with {n_users} users, {n_items} items\")\n \n # Alternating optimization\n for iteration in range(self.n_iterations):\n # Fix item factors, update user factors\n self._update_user_factors()\n \n # Fix user factors, update item factors\n self._update_item_factors()\n \n if verbose and (iteration + 1) % 5 == 0:\n rmse = self._calculate_rmse()\n print(f\" Iteration {iteration + 1}/{self.n_iterations} - RMSE: {rmse:.4f}\")\n \n if verbose:\n print(\"Training completed!\")\n \n def _update_user_factors(self):\n \"\"\"Update user factors while holding item factors fixed.\"\"\"\n item_gram = self.item_factors.T @ self.item_factors\n reg_matrix = self.reg * np.eye(self.n_factors)\n \n for user_idx in range(len(self.user_factors)):\n rated_mask = self.ratings_matrix[user_idx] != 0\n if not rated_mask.any():\n continue\n \n item_factors_rated = self.item_factors[rated_mask]\n ratings_user = self.ratings_matrix[user_idx, rated_mask]\n \n A = item_factors_rated.T @ item_factors_rated + reg_matrix\n b = item_factors_rated.T @ ratings_user\n \n self.user_factors[user_idx] = np.linalg.solve(A, b)\n \n def _update_item_factors(self):\n \"\"\"Update item factors while holding user factors fixed.\"\"\"\n user_gram = self.user_factors.T @ self.user_factors\n reg_matrix = self.reg * np.eye(self.n_factors)\n \n for item_idx in range(len(self.item_factors)):\n rated_mask = self.ratings_matrix[:, item_idx] != 0\n if not rated_mask.any():\n continue\n \n user_factors_rated = self.user_factors[rated_mask]\n ratings_item = self.ratings_matrix[rated_mask, item_idx]\n \n A = user_factors_rated.T @ user_factors_rated + reg_matrix\n b = user_factors_rated.T @ ratings_item\n \n self.item_factors[item_idx] = np.linalg.solve(A, b)\n \n def _calculate_rmse(self) -> float:\n \"\"\"Calculate RMSE on observed ratings.\"\"\"\n predictions = self.user_factors @ self.item_factors.T\n mask = self.ratings_matrix != 0\n \n errors = (self.ratings_matrix[mask] - predictions[mask]) ** 2\n return np.sqrt(np.mean(errors))\n \n def predict(self, user_id, item_id) -> float:\n \"\"\"Predict rating for a user-item pair.\"\"\"\n if user_id not in self.user_id_map or item_id not in self.item_id_map:\n return 3.0 # Default rating\n \n user_idx = self.user_id_map[user_id]\n item_idx = self.item_id_map[item_id]\n \n pred = np.dot(self.user_factors[user_idx], self.item_factors[item_idx])\n return np.clip(pred, 1.0, 5.0)\n \n def recommend(self, user_id, n_recommendations: int = 10) -> List[Tuple]:\n \"\"\"Generate top-N recommendations.\"\"\"\n if user_id not in self.user_id_map:\n return []\n \n user_idx = self.user_id_map[user_id]\n scores = self.user_factors[user_idx] @ self.item_factors.T\n \n # Get unrated items\n rated_mask = self.ratings_matrix[user_idx] != 0\n scores[rated_mask] = -np.inf\n \n top_indices = np.argsort(scores)[::-1][:n_recommendations]\n \n return [(self.reverse_item_map[idx], float(np.clip(scores[idx], 1.0, 5.0)))\n for idx in top_indices]\n```\n\n## 3. Evaluation Metrics\n\n```python\n\"\"\"\nevaluation.py\nEvaluation metrics for recommendation systems\n\"\"\"\n\nimport numpy as np\nfrom typing import List, Dict, Tuple, Set\nfrom collections import defaultdict\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import train_test_split\n\n\nclass RecommenderEvaluator:\n \"\"\"Evaluation metrics for recommendation systems.\"\"\"\n \n @staticmethod\n def rmse(predictions: List[float], actuals: List[float]) -> float:\n \"\"\"Root Mean Square Error.\"\"\"\n predictions = np.array(predictions)\n actuals = np.array(actuals)\n return np.sqrt(np.mean((predictions - actuals) ** 2))\n \n @staticmethod\n def mae(predictions: List[float], actuals: List[float]) -> float:\n \"\"\"Mean Absolute Error.\"\"\"\n predictions = np.array(predictions)\n actuals = np.array(actuals)\n return np.mean(np.abs(predictions - actuals))\n \n @staticmethod\n def precision_at_k(recommended: List, relevant: Set, k: int) -> float:\n \"\"\"\n Precision@k - Fraction of recommended items that are relevant.\n \n Args:\n recommended: List of recommended items (ordered)\n relevant: Set of relevant items\n k: Number of top recommendations to consider\n \"\"\"\n if k == 0:\n return 0.0\n \n recommended_at_k = set(recommended[:k])\n relevant_and_recommended = len(recommended_at_k & relevant)\n \n return relevant_and_recommended / k\n \n @staticmethod\n def recall_at_k(recommended: List, relevant: Set, k: int) -> float:\n \"\"\"\n Recall@k - Fraction of relevant items that are recommended.\n \"\"\"\n if len(relevant) == 0:\n return 0.0\n \n recommended_at_k = set(recommended[:k])\n relevant_and_recommended = len(recommended_at_k & relevant)\n \n return relevant_and_recommended / len(relevant)\n \n @staticmethod\n def f1_at_k(recommended: List, relevant: Set, k: int) -> float:\n \"\"\"F1 Score at k.\"\"\"\n precision = RecommenderEvaluator.precision_at_k(recommended, relevant, k)\n recall = RecommenderEvaluator.recall_at_k(recommended, relevant, k)\n \n if precision + recall == 0:\n return 0.0\n \n return 2 * precision * recall / (precision + recall)\n \n @staticmethod\n def ndcg_at_k(recommended: List, relevant: Set, k: int,\n relevance_scores: Dict = None) -> float:\n \"\"\"\n Normalized Discounted Cumulative Gain at k.\n \n Args:\n recommended: List of recommended items\n relevant: Set of relevant items\n k: Number of recommendations\n relevance_scores: Optional dict of item -> relevance score\n \"\"\"\n if len(relevant) == 0:\n return 0.0\n \n # Calculate DCG\n dcg = 0.0\n for i, item in enumerate(recommended[:k]):\n if item in relevant:\n rel = relevance_scores.get(item, 1.0) if relevance_scores else 1.0\n dcg += rel / np.log2(i + 2) # i+2 because i starts at 0\n \n # Calculate ideal DCG\n ideal_relevances = []\n for item in relevant:\n rel = relevance_scores.get(item, 1.0) if relevance_scores else 1.0\n ideal_relevances.append(rel)\n \n ideal_relevances.sort(reverse=True)\n \n idcg = 0.0\n for i, rel in enumerate(ideal_relevances[:k]):\n idcg += rel / np.log2(i + 2)\n \n if idcg == 0:\n return 0.0\n \n return dcg / idcg\n \n @staticmethod\n def mean_average_precision(recommendations: Dict[int, List],\n relevants: Dict[int, Set]) -> float:\n \"\"\"\n Mean Average Precision across all users.\n \n Args:\n recommendations: Dict of user_id -> recommended items list\n relevants: Dict of user_id -> relevant items set\n \"\"\"\n average_precisions = []\n \n for user_id in recommendations:\n if user_id not in relevants or len(relevants[user_id]) == 0:\n continue\n \n recommended = recommendations[user_id]\n relevant = relevants[user_id]\n \n hits = 0\n sum_precisions = 0.0\n \n for i, item in enumerate(recommended):\n if item in relevant:\n hits += 1\n sum_precisions += hits / (i + 1)\n \n if hits > 0:\n average_precisions.append(sum_precisions / len(relevant))\n \n return np.mean(average_precisions) if average_precisions else 0.0\n \n @staticmethod\n def hit_rate(recommended: List, relevant: Set) -> float:\n \"\"\"Whether at least one relevant item is recommended.\"\"\"\n return 1.0 if set(recommended) & relevant else 0.0\n \n @staticmethod\n def coverage(all_recommendations: List[List], n_items: int) -> float:\n \"\"\"Fraction of items that were recommended to at least one user.\"\"\"\n recommended_items = set()\n for recs in all_recommendations:\n recommended_items.update(recs)\n \n return len(recommended_items) / n_items if n_items > 0 else 0.0\n \n @staticmethod\n def diversity(recommended: List, item_similarities: np.ndarray,\n item_id_map: Dict) -> float:\n \"\"\"\n Intra-list diversity - average dissimilarity between recommended items.\n \"\"\"\n if len(recommended) 0 else 0.0\n\n\nclass CrossValidator:\n \"\"\"Cross-validation for recommender systems.\"\"\"\n \n def __init__(self, n_folds: int = 5):\n self.n_folds = n_folds\n \n def split_data(self, user_ids: np.ndarray, item_ids: np.ndarray,\n ratings: np.ndarray, test_size: float = 0.2):\n \"\"\"Split data into train and test sets.\"\"\"\n indices = np.arange(len(ratings))\n train_idx, test_idx = train_test_split(\n indices, test_size=test_size, random_state=42\n )\n \n return (\n user_ids[train_idx], item_ids[train_idx], ratings[train_idx],\n user_ids[test_idx], item_ids[test_idx], ratings[test_idx]\n )\n \n def evaluate_recommender(self, recommender, train_data: Tuple,\n test_data: Tuple, k_values: List[int] = [5, 10, 20]):\n \"\"\"\n Comprehensive evaluation of a recommender system.\n \"\"\"\n train_users, train_items, train_ratings = train_data\n test_users, test_items, test_ratings = test_data\n \n # Fit the model\n recommender.fit(train_users, train_items, train_ratings)\n \n results = {\n 'rating_prediction': {},\n 'ranking_metrics': defaultdict(list)\n }\n \n # Rating prediction metrics\n predictions = []\n actuals = []\n \n for user_id, item_id, rating in zip(test_users, test_items, test_ratings):\n pred = recommender.predict(user_id, item_id)\n predictions.append(pred)\n actuals.append(rating)\n \n evaluator = RecommenderEvaluator()\n results['rating_prediction']['RMSE'] = evaluator.rmse(predictions, actuals)\n results['rating_prediction']['MAE'] = evaluator.mae(predictions, actuals)\n \n # Build test set per user\n user_test_items = defaultdict(set)\n for user_id, item_id, rating in zip(test_users, test_items, test_ratings):\n if rating >= 4.0: # Consider high ratings as relevant\n user_test_items[user_id].add(item_id)\n \n # Ranking metrics\n for k in k_values:\n precisions = []\n recalls = []\n ndcgs = []\n \n for user_id in set(test_users):\n if user_id not in user_test_items:\n continue\n \n relevant = user_test_items[user_id]\n recommendations = recommender.recommend(user_id, n_recommendations=k)\n recommended = [r[0] for r in recommendations]\n \n precisions.append(evaluator.precision_at_k(recommended, relevant, k))\n recalls.append(evaluator.recall_at_k(recommended, relevant, k))\n ndcgs.append(evaluator.ndcg_at_k(recommended, relevant, k))\n \n results['ranking_metrics'][f'Precision@{k}'] = np.mean(precisions)\n results['ranking_metrics'][f'Recall@{k}'] = np.mean(recalls)\n results['ranking_metrics'][f'NDCG@{k}'] = np.mean(ndcgs)\n \n return results\n\n\ndef plot_evaluation_results(results: Dict, title: str = \"Evaluation Results\"):\n \"\"\"Visualize evaluation results.\"\"\"\n fig, axes = plt.subplots(1, 2, figsize=(14, 5))\n \n # Rating prediction metrics\n rating_metrics = results['rating_prediction']\n ax1 = axes[0]\n bars = ax1.bar(rating_metrics.keys(), rating_metrics.values(), color=['#3498db', '#e74c3c'])\n ax1.set_ylabel('Error')\n ax1.set_title('Rating Prediction Metrics')\n for bar, val in zip(bars, rating_metrics.values()):\n ax1.text(bar.get_x() + bar.get_width()/2, bar.get_height() + 0.01,\n f'{val:.3f}', ha='center', va='bottom')\n \n # Ranking metrics\n ranking_metrics = dict(results['ranking_metrics'])\n ax2 = axes[1]\n x = np.arange(len(ranking_metrics))\n bars = ax2.bar(x, ranking_metrics.values(), color='#2ecc71')\n ax2.set_xticks(x)\n ax2.set_xticklabels(ranking_metrics.keys(), rotation=45, ha='right')\n ax2.set_ylabel('Score')\n ax2.set_title('Ranking Metrics')\n for bar, val in zip(bars, ranking_metrics.values()):\n ax2.text(bar.get_x() + bar.get_width()/2, bar.get_height() + 0.01,\n f'{val:.3f}', ha='center', va='bottom', fontsize=8)\n \n plt.suptitle(title)\n plt.tight_layout()\n plt.savefig('evaluation_results.png', dpi=150, bbox_inches='tight')\n plt.show()\n```\n\n## 4. Demo and Usage Examples\n\n```python\n\"\"\"\ndemo.py\nDemonstration of the recommendation system\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom recommendation_system import UserBasedCF, ItemBasedCF, HybridCF\nfrom matrix_factorization import SVDRecommender, ALSRecommender\nfrom evaluation import CrossValidator, RecommenderEvaluator, plot_evaluation_results\n\n\ndef generate_sample_data(n_users: int = 500, n_items: int = 200,\n n_ratings: int = 10000, seed: int = 42):\n \"\"\"Generate sample rating data for demonstration.\"\"\"\n np.random.seed(seed)\n \n # Generate user preferences (latent factors)\n n_genres = 5\n user_preferences = np.random.rand(n_users, n_genres)\n item_genres = np.random.rand(n_items, n_genres)\n \n # Generate ratings based on preferences\n user_ids = []\n item_ids = []\n ratings = []\n \n user_item_pairs = set()\n \n while len(ratings) =1.21.0\npandas>=1.3.0\nscipy>=1.7.0\nscikit-learn>=0.24.0\nmatplotlib>=3.4.0\n```\n\n## Visualization of the Recommendation Process\n\n```\n┌─────────────────────────────────────────────────────────────────────┐\n│ COLLABORATIVE FILTERING APPROACHES │\n└─────────────────────────────────────────────────────────────────────┘\n\n┌─────────────────────────────────────────────────────────────────────┐\n│ USER-BASED CF │\n│ │\n│ User-Item Matrix Similar Users Recommendations │\n│ ┌───────────────┐ ┌─────────────┐ ┌──────────────┐ │\n│ │ I1 I2 I3 I4 │ │ User A │ │ Recommend │ │\n│ │U1 5 3 - 4 │ ───► │ similarity │ ──► │ items that │ │\n│ │U2 4 - 2 5 │ │ with U1 │ │ similar │ │\n│ │U3 - 4 4 - │ │ = 0.89 │ │ users liked │ │\n│ │... │ └─────────────┘ └──────────────┘ │\n│ └───────────────┘ │\n└─────────────────────────────────────────────────────────────────────┘\n\n┌─────────────────────────────────────────────────────────────────────┐\n│ ITEM-BASED CF │\n│ │\n│ User-Item Matrix Similar Items Recommendations │\n│ ┌───────────────┐ ┌─────────────┐ ┌──────────────┐ │\n│ │ I1 I2 I3 I4 │ │ Item A │ │ Recommend │ │\n│ │U1 5 3 - 4 │ ───► │ similar to │ ──► │ items │ │\n│ │U2 4 - 2 5 │ │ user's │ │ similar to │ │\n│ │U3 - 4 4 - │ │ rated items│ │ past likes │ │\n│ │... │ └─────────────┘ └──────────────┘ │\n│ └───────────────┘ │\n└─────────────────────────────────────────────────────────────────────┘\n\n┌─────────────────────────────────────────────────────────────────────┐\n│ MATRIX FACTORIZATION (SVD) │\n│ │\n│ User-Item Matrix = User Factors × Item Factors │\n│ ┌───────────────┐ ┌─────────┐ ┌─────────────┐ │\n│ │ R(n×m) │ = │ U(n×k) │ × │ V(k×m) │ │\n│ │ │ │ │ │ │ │\n│ │ Rating Matrix │ │ Latent │ │ Latent │ │\n│ │ │ │ User │ │ Item │ │\n│ │ │ │ Features│ │ Features │ │\n│ └───────────────┘ └─────────┘ └─────────────┘ │\n│ │\n│ Prediction: r̂(u,i) = μ + bᵤ + bᵢ + uᵀ × vᵢ │\n└─────────────────────────────────────────────────────────────────────┘\n\n┌─────────────────────────────────────────────────────────────────────┐\n│ EVALUATION METRICS │\n├─────────────────┬───────────────────────────────────────────────────┤\n│ RMSE │ √(Σ(actual - predicted)² / n) - Rating accuracy │\n│ MAE │ Σ|actual - predicted| / n - Absolute error │\n│ Precision@k │ Relevant∩Recommended / k - Recommendation quality │\n│ Recall@k │ Relevant∩Recommended / |Relevant| - Coverage │\n│ NDCG@k │ DCG/IDCG - Ranking quality with position weights │\n└─────────────────┴───────────────────────────────────────────────────┘\n```\n\n## Key Features\n\n| Feature | Description |\n|---------|-------------|\n| **User-Based CF** | Find similar users, recommend their liked items |\n| **Item-Based CF** | Find similar items to user's preferences |\n| **SVD** | Matrix factorization with gradient descent |\n| **ALS** | Alternating Least Squares optimization |\n| **Hybrid** | Combines user and item-based approaches |\n| **Evaluation** | RMSE, MAE, Precision, Recall, NDCG |\n\nThis implementation provides a complete collaborative filtering system suitable for building production-ready recommendation engines!"}],"string":"[\n {\n \"role\": \"system\",\n \"content\": \"\"\n },\n {\n \"role\": \"user\",\n \"content\": \"Develop Python code to implement a recommendation system using collaborative filtering.\"\n },\n {\n \"role\": \"assistant\",\n \"content\": \" The user wants me to implement a recommendation system using collaborative filtering in Python. I'll create a comprehensive implementation that covers:\\n\\n1. **User-Based Collaborative Filtering** - Find similar users and recommend items they liked\\n2. **Item-Based Collaborative Filtering** - Find similar items and recommend based on user's past preferences\\n3. **Matrix Factorization (SVD)** - Latent factor model\\n\\nI'll include:\\n- Data preprocessing\\n- Similarity calculations (cosine, Pearson correlation)\\n- Neighborhood-based methods\\n- Matrix factorization using SVD\\n- Evaluation metrics (RMSE, MAE, Precision, Recall)\\n- A practical example with movie ratings\\n\\nLet me create a well-structured, comprehensive implementation. # Collaborative Filtering Recommendation System in Python\\n\\nI'll create a comprehensive recommendation system implementation covering user-based, item-based collaborative filtering, and matrix factorization approaches.\\n\\n## Project Structure\\n\\n```\\nrecommendation_system/\\n├── recommendation_system.py # Core implementation\\n├── matrix_factorization.py # SVD-based approach\\n├── evaluation.py # Evaluation metrics\\n├── demo.py # Usage examples\\n└── requirements.txt # Dependencies\\n```\\n\\n## 1. Core Recommendation System\\n\\n```python\\n\\\"\\\"\\\"\\nrecommendation_system.py\\nCollaborative Filtering Recommendation System Implementation\\n\\\"\\\"\\\"\\n\\nimport numpy as np\\nimport pandas as pd\\nfrom abc import ABC, abstractmethod\\nfrom collections import defaultdict\\nfrom typing import List, Tuple, Dict, Optional, Set\\nfrom scipy.sparse import csr_matrix\\nfrom scipy.spatial.distance import cosine\\nimport warnings\\nwarnings.filterwarnings('ignore')\\n\\n\\nclass SimilarityMetrics:\\n \\\"\\\"\\\"Collection of similarity metrics for collaborative filtering.\\\"\\\"\\\"\\n \\n @staticmethod\\n def cosine_similarity(vec1: np.ndarray, vec2: np.ndarray) -> float:\\n \\\"\\\"\\\"Calculate cosine similarity between two vectors.\\\"\\\"\\\"\\n # Handle zero vectors\\n norm1 = np.linalg.norm(vec1)\\n norm2 = np.linalg.norm(vec2)\\n \\n if norm1 == 0 or norm2 == 0:\\n return 0.0\\n \\n return np.dot(vec1, vec2) / (norm1 * norm2)\\n \\n @staticmethod\\n def pearson_correlation(vec1: np.ndarray, vec2: np.ndarray) -> float:\\n \\\"\\\"\\\"Calculate Pearson correlation coefficient.\\\"\\\"\\\"\\n # Find common rated items\\n mask = (vec1 != 0) & (vec2 != 0)\\n \\n if mask.sum() float:\\n \\\"\\\"\\\"Calculate adjusted cosine similarity (for item-based CF).\\\"\\\"\\\"\\n mask = (vec1 != 0) & (vec2 != 0)\\n \\n if mask.sum() == 0:\\n return 0.0\\n \\n adj_vec1 = vec1[mask] - item_means[mask]\\n adj_vec2 = vec2[mask] - item_means[mask]\\n \\n norm1 = np.linalg.norm(adj_vec1)\\n norm2 = np.linalg.norm(adj_vec2)\\n \\n if norm1 == 0 or norm2 == 0:\\n return 0.0\\n \\n return np.dot(adj_vec1, adj_vec2) / (norm1 * norm2)\\n \\n @staticmethod\\n def jaccard_similarity(set1: Set, set2: Set) -> float:\\n \\\"\\\"\\\"Calculate Jaccard simil
2026-01-13T09:29:19
https://git-scm.com/book/pl/v2/Rozproszony-Git-Rozproszone-przep%c5%82ywy-pracy
Git - Rozproszone przepływy pracy About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pierwsze kroki 1.1 Wprowadzenie do kontroli wersji 1.2 Krótka historia Git 1.3 Podstawy Git 1.4 Linia poleceń 1.5 Instalacja Git 1.6 Wstępna konfiguracja Git 1.7 Uzyskiwanie pomocy 1.8 Podsumowanie 2. Podstawy Gita 2.1 Pierwsze repozytorium Gita 2.2 Rejestrowanie zmian w repozytorium 2.3 Podgląd historii rewizji 2.4 Cofanie zmian 2.5 Praca ze zdalnym repozytorium 2.6 Tagowanie 2.7 Aliasy 2.8 Podsumowanie 3. Gałęzie Gita 3.1 Czym jest gałąź 3.2 Podstawy rozgałęziania i scalania 3.3 Zarządzanie gałęziami 3.4 Sposoby pracy z gałęziami 3.5 Gałęzie zdalne 3.6 Zmiana bazy 3.7 Podsumowanie 4. Git na serwerze 4.1 Protokoły 4.2 Uruchomienie Git na serwerze 4.3 Generowanie Twojego publicznego klucza SSH 4.4 Konfigurowanie serwera 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Inne opcje hostowania przez podmioty zewnętrzne 4.10 Podsumowanie 5. Rozproszony Git 5.1 Rozproszone przepływy pracy 5.2 Wgrywanie zmian do projektu 5.3 Utrzymywanie projektu 5.4 Podsumowanie 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Narzędzia Gita 7.1 Wskazywanie rewizji 7.2 Interaktywne używanie przechowali 7.3 Schowek i czyszczenie 7.4 Signing Your Work 7.5 Searching 7.6 Przepisywanie historii 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugowanie z Gitem 7.11 Moduły zależne 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Podsumowanie 8. Dostosowywanie Gita 8.1 Konfiguracja Gita 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git i inne systemy 9.1 Git jako klient 9.2 Migracja do Gita 9.3 Podsumowanie 10. Mechanizmy wewnętrzne w Git 10.1 Komendy typu plumbing i porcelain 10.2 Obiekty Gita 10.3 Referencje w Git 10.4 Spakowane pliki (packfiles) 10.5 Refspec 10.6 Protokoły transferu 10.7 Konserwacja i odzyskiwanie danych 10.8 Environment Variables 10.9 Podsumowanie A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 5.1 Rozproszony Git - Rozproszone przepływy pracy Teraz, gdy masz już skonfigurowane zdalne repozytorium, które służy do wymiany pracy między programistami w projekcie oraz jesteś zaznajomiony z podstawowymi komendami pozwalającymi na pracę z lokalnym repozytorium Git, zobaczysz jak wykorzystać jego możliwości w rozproszonym trybie pracy, który Git umożliwia. W tym rozdziale, zobaczysz jak pracować z Gitem w rozproszonym środowisku jako współpracownik lub integrator zmian. Nauczysz się jak udostępniać wprowadzone zmiany oraz jak zrobić to najprościej jak tylko się da dla siebie i opiekuna projektu, oraz jak zarządzać projektem w którym uczestniczy wielu programistów. Rozproszone przepływy pracy Odmiennie do scentralizowanych systemów kontroli wersji (ang. Centralized Version Control Systems , CVCSs), rozproszona natura systemu Git pozwala na dużo bardziej elastyczne podejście do tego w jaki sposób przebiega współpraca między programistami. W scentralizowanych systemach każdy programista jest osobnym elementem pracującym na centralnym serwerze. W Gitcie każdy programista posiada zarówno swoje oddzielne repozytorium, które może zostać udostępnione dla innych, jak również centralny serwer do którego inni mogą wgrywać swoje zmiany. Umożliwia to szerokie możliwości współpracy dla Twojego projektu i/lub zespołu, dlatego opiszę kilka często używanych zachować które z tego korzystają. Pokażemy zalety i wady każdego z rozwiązań; możesz wybrać jeden odpowiadający tobie, lub możesz je połączyć i zmieszać ze sobą. Scentralizowany przepływ pracy W scentralizowanych systemach, zazwyczaj jest stosowany model centralnego przepływu. W jednym centralnym punkcie znajduje się repozytorium, do którego wgrywane są zmiany, a pozostali współpracownicy synchronizują swoją pracę z nim. Wszyscy programiści uczestniczący w projekcie są końcówkami, łączącymi się z centralnym serwerem – oraz synchronizującymi się z nim Figure 54. Scentralizowany przepływ pracy. Oznacza to tyle, że w sytuacji w której dwóch niezależnych programistów korzystających z tego centralnego repozytorium będzie próbowało wgrać swoje zmiany, tylko pierwszemu z nich uda się tego dokonać bezproblemowo. Drugi przed wgraniem, będzie musiał najpierw pobrać i zintegrować zmiany wprowadzone przez pierwszego programistę, a dopiero później ponowić próbę wysłania swoich na serwer. Taki rodzaj współpracy sprawdza się doskonale w Gitcie, tak samo jak funkcjonuje on w Subversion (lub każdym innym CVCS). Jeżeli masz mały zespół, lub dobrze znacie pracę z jednym centralnym repozytorium w firmie lub zespole, możesz bez problemów kontynuować ten rodzaj pracy z Gitem. Po prostu załóż nowe repozytorium, nadaj każdej osobie z zespołu uprawnienia do wgrywania zmian (za pomocą komendy push ); Git nie pozwoli na nadpisanie pracy jednego programisty przez innego. Powiedzmy, że John i Jessica zaczynają pracować w tym samym czasie. John kończy wprowadzać swoje zmiany i przesyła je na serwer. Następnie Jessica próbuje wysłać swoje zmiany, ale serwer je odrzuca. Dostaje informację zwrotną, że próbuje wysłać zmiany nie będące w trybie przewijania do przodu ( not-fast-forward ), i że nie będzie mogła tego zrobić, dopóki nie pobierze ( fetch ) i nie połączy ( merge ) zmian. Nie jest to również ograniczone do małych zespołów. Dzięki modelowi rozgałęziania Gita, możliwe jest, aby setki programistów z powodzeniem pracowało jednocześnie nad jednym projektem poprzez dziesiątki gałęzi. Przepływ pracy z osobą integrującą zmiany Ponieważ Git powala na posiadanie wielu zdalnych repozytoriów, możliwy jest schemat pracy w którym każdy programista ma uprawnienia do zapisu do swojego własnego repozytorium oraz uprawnienia do odczytu do repozytorium innych osób w zespole. Ten scenariusz często zawiera jedno centralne "oficjalne" repozytorium projektu. Aby wgrać zmiany do projektu, należy stworzyć publiczną kopię tego repozytorium i wgrać ( push ) zmiany do niego. Następnie należy wysłać prośbę do opiekuna aby pobrał zmiany do głównego repozytorium. Może on dodać Twoje repozytorium jako zdalne, przetestować Twoje zmiany lokalnie, włączyć je do nowej gałęzi i następnie wgrać do repozytorium. Proces ten wygląda następująco (por. Przepływ pracy z osobą integrującą zmiany. ): Opiekun projektu wgrywa zmiany do publicznego repozytorium. Programiści klonują to repozytorium i wprowadzają zmiany. Programista wgrywa zmiany do swojego publicznego repozytorium. Programista wysyła prośbę do opiekuna projektu, aby pobrał zmiany z jego repozytorium. Opiekun dodaje repozytorium programisty jako repozytorium zdalne i pobiera zmiany. Opiekun wgrywa włączone zmiany do głównego repozytorium. Figure 55. Przepływ pracy z osobą integrującą zmiany. To jest bardzo popularne podejście podczas współpracy przy pomocy stron takich jak GitHub lub GitLab, gdzie bardzo łatwo można stworzyć kopię repozytorium i wgrywać zmiany do niego aby każdy mógł je zobaczyć. Jedną z głównych zalet takiego podejścia jest to, że możesz kontynuować pracę, a opiekun może pobrać Twoje zmiany w dowolnym czasie. Programiści nie muszą czekać na opiekuna, aż ten włączy ich zmiany – każdy z nich może pracować oddzielnie. Przepływ pracy z dyktatorem i porucznikami To jest wariant przepływu z wieloma repozytoriami. Zazwyczaj jest on używany w bardzo dużych projektach, z setkami programistów; najbardziej znanym przykładem może być jądro Linuksa. Kilkoro opiekunów jest wydelegowanych do obsługi wydzielonych części repozytorium; nazwijmy ich porucznikami. Wszyscy z nich mają jedną, główną osobę integrującą zmiany – znaną jako miłościwy dyktator. Repozytorium dyktatora jest wzorcowym, z którego wszyscy programiści pobierają zmiany. Cały proces działa następująco (por. Przepływ pracy z miłościwym dyktatorem. ): Programiści pracują nad swoimi gałęziami tematycznymi, oraz wykonują rebase na gałęzi master . Gałąź master jest tą pobraną od dyktatora. Porucznicy włączają ( merge ) zmiany programistów do swojej gałęzi master . Dyktator włącza ( merge ) gałęzie master udostępnione przez poruczników do swojej gałęzi master . Dyktator wypycha ( push ) swoją gałąź master do głównego repozytorium, tak aby inni programiści mogli na niej pracować. Figure 56. Przepływ pracy z miłościwym dyktatorem. Ten rodzaj współpracy nie jest częsty w użyciu, ale może być użyteczny w bardzo dużych projektach, lub bardzo rozbudowanych strukturach zespołów. Pozwala liderowi projektu (dyktatorowi) na delegowanie dużej części pracy i zbieranie dużych części kodu w wielu punktach czasu przed ich zintegrowaniem. Podsumowanie przepływu pracy To są najczęściej stosowane przepływy pracy możliwe przy użyciu rozproszonego systemu takiego jak Git, jednak możesz zauważyć że istnieje w tym względzie duża dowolność, tak abyś mógł dostosować go do używanego przez siebie tryby pracy. Teraz gdy (miejmy nadzieję) możesz już wybrać sposób pracy który jest dla Ciebie odpowiedni, pokażemy kilka konkretnych przykładów w jaki sposób osiągnąć odpowiedni podział ról dla każdego z opisanych przepływów. W następnej części dowiesz się o kilku popularnych wzorcach współtworzenia projektu. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://doc.rust-lang.org/cargo/reference/config.html#http
Configuration - The Cargo Book Keyboard shortcuts Press ← or → to navigate between chapters Press S or / to search in the book Press ? to show this help Press Esc to hide this help Auto Light Rust Coal Navy Ayu The Cargo Book Configuration This document explains how Cargo’s configuration system works, as well as available keys or configuration. For configuration of a package through its manifest, see the manifest format . Hierarchical structure Cargo allows local configuration for a particular package as well as global configuration. It looks for configuration files in the current directory and all parent directories. If, for example, Cargo were invoked in /projects/foo/bar/baz , then the following configuration files would be probed for and unified in this order: /projects/foo/bar/baz/.cargo/config.toml /projects/foo/bar/.cargo/config.toml /projects/foo/.cargo/config.toml /projects/.cargo/config.toml /.cargo/config.toml $CARGO_HOME/config.toml which defaults to: Windows: %USERPROFILE%\.cargo\config.toml Unix: $HOME/.cargo/config.toml With this structure, you can specify configuration per-package, and even possibly check it into version control. You can also specify personal defaults with a configuration file in your home directory. If a key is specified in multiple config files, the values will get merged together. Numbers, strings, and booleans will use the value in the deeper config directory taking precedence over ancestor directories, where the home directory is the lowest priority. Arrays will be joined together with higher precedence items being placed later in the merged array. At present, when being invoked from a workspace, Cargo does not read config files from crates within the workspace. i.e. if a workspace has two crates in it, named /projects/foo/bar/baz/mylib and /projects/foo/bar/baz/mybin , and there are Cargo configs at /projects/foo/bar/baz/mylib/.cargo/config.toml and /projects/foo/bar/baz/mybin/.cargo/config.toml , Cargo does not read those configuration files if it is invoked from the workspace root ( /projects/foo/bar/baz/ ). Note: Cargo also reads config files without the .toml extension, such as .cargo/config . Support for the .toml extension was added in version 1.39 and is the preferred form. If both files exist, Cargo will use the file without the extension. Configuration format Configuration files are written in the TOML format (like the manifest), with simple key-value pairs inside of sections (tables). The following is a quick overview of all settings, with detailed descriptions found below. paths = ["/path/to/override"] # path dependency overrides [alias] # command aliases b = "build" c = "check" t = "test" r = "run" rr = "run --release" recursive_example = "rr --example recursions" space_example = ["run", "--release", "--", "\"command list\""] [build] jobs = 1 # number of parallel jobs, defaults to # of CPUs rustc = "rustc" # the rust compiler tool rustc-wrapper = "…" # run this wrapper instead of `rustc` rustc-workspace-wrapper = "…" # run this wrapper instead of `rustc` for workspace members rustdoc = "rustdoc" # the doc generator tool target = "triple" # build for the target triple (ignored by `cargo install`) target-dir = "target" # path of where to place generated artifacts build-dir = "target" # path of where to place intermediate build artifacts rustflags = ["…", "…"] # custom flags to pass to all compiler invocations rustdocflags = ["…", "…"] # custom flags to pass to rustdoc incremental = true # whether or not to enable incremental compilation dep-info-basedir = "…" # path for the base directory for targets in depfiles [credential-alias] # Provides a way to define aliases for credential providers. my-alias = ["/usr/bin/cargo-credential-example", "--argument", "value", "--flag"] [doc] browser = "chromium" # browser to use with `cargo doc --open`, # overrides the `BROWSER` environment variable [env] # Set ENV_VAR_NAME=value for any process run by Cargo ENV_VAR_NAME = "value" # Set even if already present in environment ENV_VAR_NAME_2 = { value = "value", force = true } # `value` is relative to the parent of `.cargo/config.toml`, env var will be the full absolute path ENV_VAR_NAME_3 = { value = "relative/path", relative = true } [future-incompat-report] frequency = 'always' # when to display a notification about a future incompat report [cache] auto-clean-frequency = "1 day" # How often to perform automatic cache cleaning [cargo-new] vcs = "none" # VCS to use ('git', 'hg', 'pijul', 'fossil', 'none') [http] debug = false # HTTP debugging proxy = "host:port" # HTTP proxy in libcurl format ssl-version = "tlsv1.3" # TLS version to use ssl-version.max = "tlsv1.3" # maximum TLS version ssl-version.min = "tlsv1.1" # minimum TLS version timeout = 30 # timeout for each HTTP request, in seconds low-speed-limit = 10 # network timeout threshold (bytes/sec) cainfo = "cert.pem" # path to Certificate Authority (CA) bundle proxy-cainfo = "cert.pem" # path to proxy Certificate Authority (CA) bundle check-revoke = true # check for SSL certificate revocation multiplexing = true # HTTP/2 multiplexing user-agent = "…" # the user-agent header [install] root = "/some/path" # `cargo install` destination directory [net] retry = 3 # network retries git-fetch-with-cli = true # use the `git` executable for git operations offline = true # do not access the network [net.ssh] known-hosts = ["..."] # known SSH host keys [patch.<registry>] # Same keys as for [patch] in Cargo.toml [profile.<name>] # Modify profile settings via config. inherits = "dev" # Inherits settings from [profile.dev]. opt-level = 0 # Optimization level. debug = true # Include debug info. split-debuginfo = '...' # Debug info splitting behavior. strip = "none" # Removes symbols or debuginfo. debug-assertions = true # Enables debug assertions. overflow-checks = true # Enables runtime integer overflow checks. lto = false # Sets link-time optimization. panic = 'unwind' # The panic strategy. incremental = true # Incremental compilation. codegen-units = 16 # Number of code generation units. rpath = false # Sets the rpath linking option. [profile.<name>.build-override] # Overrides build-script settings. # Same keys for a normal profile. [profile.<name>.package.<name>] # Override profile for a package. # Same keys for a normal profile (minus `panic`, `lto`, and `rpath`). [resolver] incompatible-rust-versions = "allow" # Specifies how resolver reacts to these [registries.<name>] # registries other than crates.io index = "…" # URL of the registry index token = "…" # authentication token for the registry credential-provider = "cargo:token" # The credential provider for this registry. [registries.crates-io] protocol = "sparse" # The protocol to use to access crates.io. [registry] default = "…" # name of the default registry token = "…" # authentication token for crates.io credential-provider = "cargo:token" # The credential provider for crates.io. global-credential-providers = ["cargo:token"] # The credential providers to use by default. [source.<name>] # source definition and replacement replace-with = "…" # replace this source with the given named source directory = "…" # path to a directory source registry = "…" # URL to a registry source local-registry = "…" # path to a local registry source git = "…" # URL of a git repository source branch = "…" # branch name for the git repository tag = "…" # tag name for the git repository rev = "…" # revision for the git repository [target.<triple>] linker = "…" # linker to use runner = "…" # wrapper to run executables rustflags = ["…", "…"] # custom flags for `rustc` rustdocflags = ["…", "…"] # custom flags for `rustdoc` [target.<cfg>] linker = "…" # linker to use runner = "…" # wrapper to run executables rustflags = ["…", "…"] # custom flags for `rustc` [target.<triple>.<links>] # `links` build script override rustc-link-lib = ["foo"] rustc-link-search = ["/path/to/foo"] rustc-flags = "-L /some/path" rustc-cfg = ['key="value"'] rustc-env = {key = "value"} rustc-cdylib-link-arg = ["…"] metadata_key1 = "value" metadata_key2 = "value" [term] quiet = false # whether cargo output is quiet verbose = false # whether cargo provides verbose output color = 'auto' # whether cargo colorizes output hyperlinks = true # whether cargo inserts links into output unicode = true # whether cargo can render output using non-ASCII unicode characters progress.when = 'auto' # whether cargo shows progress bar progress.width = 80 # width of progress bar progress.term-integration = true # whether cargo reports progress to terminal emulator Environment variables Cargo can also be configured through environment variables in addition to the TOML configuration files. For each configuration key of the form foo.bar the environment variable CARGO_FOO_BAR can also be used to define the value. Keys are converted to uppercase, dots and dashes are converted to underscores. For example the target.x86_64-unknown-linux-gnu.runner key can also be defined by the CARGO_TARGET_X86_64_UNKNOWN_LINUX_GNU_RUNNER environment variable. Environment variables will take precedence over TOML configuration files. Currently only integer, boolean, string and some array values are supported to be defined by environment variables. Descriptions below indicate which keys support environment variables and otherwise they are not supported due to technical issues . In addition to the system above, Cargo recognizes a few other specific environment variables . Command-line overrides Cargo also accepts arbitrary configuration overrides through the --config command-line option. The argument should be in TOML syntax of KEY=VALUE or provided as a path to an extra configuration file: # With `KEY=VALUE` in TOML syntax cargo --config net.git-fetch-with-cli=true fetch # With a path to a configuration file cargo --config ./path/to/my/extra-config.toml fetch The --config option may be specified multiple times, in which case the values are merged in left-to-right order, using the same merging logic that is used when multiple configuration files apply. Configuration values specified this way take precedence over environment variables, which take precedence over configuration files. When the --config option is provided as an extra configuration file, The configuration file loaded this way follow the same precedence rules as other options specified directly with --config . Some examples of what it looks like using Bourne shell syntax: # Most shells will require escaping. cargo --config http.proxy=\"http://example.com\" … # Spaces may be used. cargo --config "net.git-fetch-with-cli = true" … # TOML array example. Single quotes make it easier to read and write. cargo --config 'build.rustdocflags = ["--html-in-header", "header.html"]' … # Example of a complex TOML key. cargo --config "target.'cfg(all(target_arch = \"arm\", target_os = \"none\"))'.runner = 'my-runner'" … # Example of overriding a profile setting. cargo --config profile.dev.package.image.opt-level=3 … Config-relative paths Paths in config files may be absolute, relative, or a bare name without any path separators. Paths for executables without a path separator will use the PATH environment variable to search for the executable. Paths for non-executables will be relative to where the config value is defined. In particular, rules are: For environment variables, paths are relative to the current working directory. For config values loaded directly from the --config KEY=VALUE option, paths are relative to the current working directory. For config files, paths are relative to the parent directory of the directory where the config files were defined, no matter those files are from either the hierarchical probing or the --config <path> option. Note: To maintain consistency with existing .cargo/config.toml probing behavior, it is by design that a path in a config file passed via --config <path> is also relative to two levels up from the config file itself. To avoid unexpected results, the rule of thumb is putting your extra config files at the same level of discovered .cargo/config.toml in your project. For instance, given a project /my/project , it is recommended to put config files under /my/project/.cargo or a new directory at the same level, such as /my/project/.config . # Relative path examples. [target.x86_64-unknown-linux-gnu] runner = "foo" # Searches `PATH` for `foo`. [source.vendored-sources] # Directory is relative to the parent where `.cargo/config.toml` is located. # For example, `/my/project/.cargo/config.toml` would result in `/my/project/vendor`. directory = "vendor" Executable paths with arguments Some Cargo commands invoke external programs, which can be configured as a path and some number of arguments. The value may be an array of strings like ['/path/to/program', 'somearg'] or a space-separated string like '/path/to/program somearg' . If the path to the executable contains a space, the list form must be used. If Cargo is passing other arguments to the program such as a path to open or run, they will be passed after the last specified argument in the value of an option of this format. If the specified program does not have path separators, Cargo will search PATH for its executable. Credentials Configuration values with sensitive information are stored in the $CARGO_HOME/credentials.toml file. This file is automatically created and updated by cargo login and cargo logout when using the cargo:token credential provider. Tokens are used by some Cargo commands such as cargo publish for authenticating with remote registries. Care should be taken to protect the tokens and to keep them secret. It follows the same format as Cargo config files. [registry] token = "…" # Access token for crates.io [registries.<name>] token = "…" # Access token for the named registry As with most other config values, tokens may be specified with environment variables. The token for crates.io may be specified with the CARGO_REGISTRY_TOKEN environment variable. Tokens for other registries may be specified with environment variables of the form CARGO_REGISTRIES_<name>_TOKEN where <name> is the name of the registry in all capital letters. Note: Cargo also reads and writes credential files without the .toml extension, such as .cargo/credentials . Support for the .toml extension was added in version 1.39. In version 1.68, Cargo writes to the file with the extension by default. However, for backward compatibility reason, when both files exist, Cargo will read and write the file without the extension. Configuration keys This section documents all configuration keys. The description for keys with variable parts are annotated with angled brackets like target.<triple> where the <triple> part can be any target triple like target.x86_64-pc-windows-msvc . paths Type: array of strings (paths) Default: none Environment: not supported An array of paths to local packages which are to be used as overrides for dependencies. For more information see the Overriding Dependencies guide . [alias] Type: string or array of strings Default: see below Environment: CARGO_ALIAS_<name> The [alias] table defines CLI command aliases. For example, running cargo b is an alias for running cargo build . Each key in the table is the subcommand, and the value is the actual command to run. The value may be an array of strings, where the first element is the command and the following are arguments. It may also be a string, which will be split on spaces into subcommand and arguments. The following aliases are built-in to Cargo: [alias] b = "build" c = "check" d = "doc" t = "test" r = "run" rm = "remove" Aliases are not allowed to redefine existing built-in commands. Aliases are recursive: [alias] rr = "run --release" recursive_example = "rr --example recursions" [build] The [build] table controls build-time operations and compiler settings. build.jobs Type: integer or string Default: number of logical CPUs Environment: CARGO_BUILD_JOBS Sets the maximum number of compiler processes to run in parallel. If negative, it sets the maximum number of compiler processes to the number of logical CPUs plus provided value. Should not be 0. If a string default is provided, it sets the value back to defaults. Can be overridden with the --jobs CLI option. build.rustc Type: string (program path) Default: "rustc" Environment: CARGO_BUILD_RUSTC or RUSTC Sets the executable to use for rustc . build.rustc-wrapper Type: string (program path) Default: none Environment: CARGO_BUILD_RUSTC_WRAPPER or RUSTC_WRAPPER Sets a wrapper to execute instead of rustc . The first argument passed to the wrapper is the path to the actual executable to use (i.e., build.rustc , if that is set, or "rustc" otherwise). build.rustc-workspace-wrapper Type: string (program path) Default: none Environment: CARGO_BUILD_RUSTC_WORKSPACE_WRAPPER or RUSTC_WORKSPACE_WRAPPER Sets a wrapper to execute instead of rustc , for workspace members only. When building a single-package project without workspaces, that package is considered to be the workspace. The first argument passed to the wrapper is the path to the actual executable to use (i.e., build.rustc , if that is set, or "rustc" otherwise). It affects the filename hash so that artifacts produced by the wrapper are cached separately. If both rustc-wrapper and rustc-workspace-wrapper are set, then they will be nested: the final invocation is $RUSTC_WRAPPER $RUSTC_WORKSPACE_WRAPPER $RUSTC . build.rustdoc Type: string (program path) Default: "rustdoc" Environment: CARGO_BUILD_RUSTDOC or RUSTDOC Sets the executable to use for rustdoc . build.target Type: string or array of strings Default: host platform Environment: CARGO_BUILD_TARGET The default target platform triples to compile to. Possible values: Any supported target in rustc --print target-list . "host-tuple" , which will internally be substituted by the host’s target. This can be particularly useful if you’re cross-compiling some crates, and don’t want to specify your host’s machine as a target (for instance, an xtask in a shared project that may be worked on by many hosts). A path to a custom target specification. See Custom Target Lookup Path for more information. Can be overridden with the --target CLI option. [build] target = ["x86_64-unknown-linux-gnu", "i686-unknown-linux-gnu"] build.target-dir Type: string (path) Default: "target" Environment: CARGO_BUILD_TARGET_DIR or CARGO_TARGET_DIR The path to where all compiler output is placed. The default if not specified is a directory named target located at the root of the workspace. Can be overridden with the --target-dir CLI option. For more information see the build cache documentation . build.build-dir Type: string (path) Default: Defaults to the value of build.target-dir Environment: CARGO_BUILD_BUILD_DIR The directory where intermediate build artifacts will be stored. Intermediate artifacts are produced by Rustc/Cargo during the build process. This option supports path templating. Available template variables: {workspace-root} resolves to root of the current workspace. {cargo-cache-home} resolves to CARGO_HOME {workspace-path-hash} resolves to a hash of the manifest path For more information see the build cache documentation . build.rustflags Type: string or array of strings Default: none Environment: CARGO_BUILD_RUSTFLAGS or CARGO_ENCODED_RUSTFLAGS or RUSTFLAGS Extra command-line flags to pass to rustc . The value may be an array of strings or a space-separated string. There are four mutually exclusive sources of extra flags. They are checked in order, with the first one being used: CARGO_ENCODED_RUSTFLAGS environment variable. RUSTFLAGS environment variable. All matching target.<triple>.rustflags and target.<cfg>.rustflags config entries joined together. build.rustflags config value. Additional flags may also be passed with the cargo rustc command. If the --target flag (or build.target ) is used, then the flags will only be passed to the compiler for the target. Things being built for the host, such as build scripts or proc macros, will not receive the args. Without --target , the flags will be passed to all compiler invocations (including build scripts and proc macros) because dependencies are shared. If you have args that you do not want to pass to build scripts or proc macros and are building for the host, pass --target with the host triple . It is not recommended to pass in flags that Cargo itself usually manages. For example, the flags driven by profiles are best handled by setting the appropriate profile setting. Caution : Due to the low-level nature of passing flags directly to the compiler, this may cause a conflict with future versions of Cargo which may issue the same or similar flags on its own which may interfere with the flags you specify. This is an area where Cargo may not always be backwards compatible. build.rustdocflags Type: string or array of strings Default: none Environment: CARGO_BUILD_RUSTDOCFLAGS or CARGO_ENCODED_RUSTDOCFLAGS or RUSTDOCFLAGS Extra command-line flags to pass to rustdoc . The value may be an array of strings or a space-separated string. There are four mutually exclusive sources of extra flags. They are checked in order, with the first one being used: CARGO_ENCODED_RUSTDOCFLAGS environment variable. RUSTDOCFLAGS environment variable. All matching target.<triple>.rustdocflags config entries joined together. build.rustdocflags config value. Additional flags may also be passed with the cargo rustdoc command. Caution : Due to the low-level nature of passing flags directly to the compiler, this may cause a conflict with future versions of Cargo which may issue the same or similar flags on its own which may interfere with the flags you specify. This is an area where Cargo may not always be backwards compatible. build.incremental Type: bool Default: from profile Environment: CARGO_BUILD_INCREMENTAL or CARGO_INCREMENTAL Whether or not to perform incremental compilation . The default if not set is to use the value from the profile . Otherwise this overrides the setting of all profiles. The CARGO_INCREMENTAL environment variable can be set to 1 to force enable incremental compilation for all profiles, or 0 to disable it. This env var overrides the config setting. build.dep-info-basedir Type: string (path) Default: none Environment: CARGO_BUILD_DEP_INFO_BASEDIR Strips the given path prefix from dep info file paths. This config setting is intended to convert absolute paths to relative paths for tools that require relative paths. The setting itself is a config-relative path. So, for example, a value of "." would strip all paths starting with the parent directory of the .cargo directory. build.pipelining This option is deprecated and unused. Cargo always has pipelining enabled. [credential-alias] Type: string or array of strings Default: empty Environment: CARGO_CREDENTIAL_ALIAS_<name> The [credential-alias] table defines credential provider aliases. These aliases can be referenced as an element of the registry.global-credential-providers array, or as a credential provider for a specific registry under registries.<NAME>.credential-provider . If specified as a string, the value will be split on spaces into path and arguments. For example, to define an alias called my-alias : [credential-alias] my-alias = ["/usr/bin/cargo-credential-example", "--argument", "value", "--flag"] See Registry Authentication for more information. [doc] The [doc] table defines options for the cargo doc command. doc.browser Type: string or array of strings ( program path with args ) Default: BROWSER environment variable, or, if that is missing, opening the link in a system specific way This option sets the browser to be used by cargo doc , overriding the BROWSER environment variable when opening documentation with the --open option. [cargo-new] The [cargo-new] table defines defaults for the cargo new command. cargo-new.name This option is deprecated and unused. cargo-new.email This option is deprecated and unused. cargo-new.vcs Type: string Default: "git" or "none" Environment: CARGO_CARGO_NEW_VCS Specifies the source control system to use for initializing a new repository. Valid values are git , hg (for Mercurial), pijul , fossil or none to disable this behavior. Defaults to git , or none if already inside a VCS repository. Can be overridden with the --vcs CLI option. [env] The [env] section allows you to set additional environment variables for build scripts, rustc invocations, cargo run and cargo build . [env] OPENSSL_DIR = "/opt/openssl" By default, the variables specified will not override values that already exist in the environment. This behavior can be changed by setting the force flag. Setting the relative flag evaluates the value as a config-relative path that is relative to the parent directory of the .cargo directory that contains the config.toml file. The value of the environment variable will be the full absolute path. [env] TMPDIR = { value = "/home/tmp", force = true } OPENSSL_DIR = { value = "vendor/openssl", relative = true } [future-incompat-report] The [future-incompat-report] table controls setting for future incompat reporting future-incompat-report.frequency Type: string Default: "always" Environment: CARGO_FUTURE_INCOMPAT_REPORT_FREQUENCY Controls how often we display a notification to the terminal when a future incompat report is available. Possible values: always (default): Always display a notification when a command (e.g. cargo build ) produces a future incompat report never : Never display a notification [cache] The [cache] table defines settings for cargo’s caches. Global caches When running cargo commands, Cargo will automatically track which files you are using within the global cache. Periodically, Cargo will delete files that have not been used for some period of time. It will delete files that have to be downloaded from the network if they have not been used in 3 months. Files that can be generated without network access will be deleted if they have not been used in 1 month. The automatic deletion of files only occurs when running commands that are already doing a significant amount of work, such as all of the build commands ( cargo build , cargo test , cargo check , etc.), and cargo fetch . Automatic deletion is disabled if cargo is offline such as with --offline or --frozen to avoid deleting artifacts that may need to be used if you are offline for a long period of time. Note : This tracking is currently only implemented for the global cache in Cargo’s home directory. This includes registry indexes and source files downloaded from registries and git dependencies. Support for tracking build artifacts is not yet implemented, and tracked in cargo#13136 . Additionally, there is an unstable feature to support manually triggering cache cleaning, and to further customize the configuration options. See the Unstable chapter for more information. cache.auto-clean-frequency Type: string Default: "1 day" Environment: CARGO_CACHE_AUTO_CLEAN_FREQUENCY This option defines how often Cargo will automatically delete unused files in the global cache. This does not define how old the files must be, those thresholds are described above . It supports the following settings: "never" — Never deletes old files. "always" — Checks to delete old files every time Cargo runs. An integer followed by “seconds”, “minutes”, “hours”, “days”, “weeks”, or “months” — Checks to delete old files at most the given time frame. [http] The [http] table defines settings for HTTP behavior. This includes fetching crate dependencies and accessing remote git repositories. http.debug Type: boolean Default: false Environment: CARGO_HTTP_DEBUG If true , enables debugging of HTTP requests. The debug information can be seen by setting the CARGO_LOG=network=debug environment variable (or use network=trace for even more information). Be wary when posting logs from this output in a public location. The output may include headers with authentication tokens which you don’t want to leak! Be sure to review logs before posting them. http.proxy Type: string Default: none Environment: CARGO_HTTP_PROXY or HTTPS_PROXY or https_proxy or http_proxy Sets an HTTP and HTTPS proxy to use. The format is in libcurl format as in [protocol://]host[:port] . If not set, Cargo will also check the http.proxy setting in your global git configuration. If none of those are set, the HTTPS_PROXY or https_proxy environment variables set the proxy for HTTPS requests, and http_proxy sets it for HTTP requests. http.timeout Type: integer Default: 30 Environment: CARGO_HTTP_TIMEOUT or HTTP_TIMEOUT Sets the timeout for each HTTP request, in seconds. http.cainfo Type: string (path) Default: none Environment: CARGO_HTTP_CAINFO Path to a Certificate Authority (CA) bundle file, used to verify TLS certificates. If not specified, Cargo attempts to use the system certificates. http.proxy-cainfo Type: string (path) Default: falls back to http.cainfo if not set Environment: CARGO_HTTP_PROXY_CAINFO Path to a Certificate Authority (CA) bundle file, used to verify proxy TLS certificates. http.check-revoke Type: boolean Default: true (Windows) false (all others) Environment: CARGO_HTTP_CHECK_REVOKE This determines whether or not TLS certificate revocation checks should be performed. This only works on Windows. http.ssl-version Type: string or min/max table Default: none Environment: CARGO_HTTP_SSL_VERSION This sets the minimum TLS version to use. It takes a string, with one of the possible values of "default" , "tlsv1" , "tlsv1.0" , "tlsv1.1" , "tlsv1.2" , or "tlsv1.3" . This may alternatively take a table with two keys, min and max , which each take a string value of the same kind that specifies the minimum and maximum range of TLS versions to use. The default is a minimum version of "tlsv1.0" and a max of the newest version supported on your platform, typically "tlsv1.3" . http.low-speed-limit Type: integer Default: 10 Environment: CARGO_HTTP_LOW_SPEED_LIMIT This setting controls timeout behavior for slow connections. If the average transfer speed in bytes per second is below the given value for http.timeout seconds (default 30 seconds), then the connection is considered too slow and Cargo will abort and retry. http.multiplexing Type: boolean Default: true Environment: CARGO_HTTP_MULTIPLEXING When true , Cargo will attempt to use the HTTP2 protocol with multiplexing. This allows multiple requests to use the same connection, usually improving performance when fetching multiple files. If false , Cargo will use HTTP 1.1 without pipelining. http.user-agent Type: string Default: Cargo’s version Environment: CARGO_HTTP_USER_AGENT Specifies a custom user-agent header to use. The default if not specified is a string that includes Cargo’s version. [install] The [install] table defines defaults for the cargo install command. install.root Type: string (path) Default: Cargo’s home directory Environment: CARGO_INSTALL_ROOT Sets the path to the root directory for installing executables for cargo install . Executables go into a bin directory underneath the root. To track information of installed executables, some extra files, such as .crates.toml and .crates2.json , are also created under this root. The default if not specified is Cargo’s home directory (default .cargo in your home directory). Can be overridden with the --root command-line option. [net] The [net] table controls networking configuration. net.retry Type: integer Default: 3 Environment: CARGO_NET_RETRY Number of times to retry possibly spurious network errors. net.git-fetch-with-cli Type: boolean Default: false Environment: CARGO_NET_GIT_FETCH_WITH_CLI If this is true , then Cargo will use the git executable to fetch registry indexes and git dependencies. If false , then it uses a built-in git library. Setting this to true can be helpful if you have special authentication requirements that Cargo does not support. See Git Authentication for more information about setting up git authentication. net.offline Type: boolean Default: false Environment: CARGO_NET_OFFLINE If this is true , then Cargo will avoid accessing the network, and attempt to proceed with locally cached data. If false , Cargo will access the network as needed, and generate an error if it encounters a network error. Can be overridden with the --offline command-line option. net.ssh The [net.ssh] table contains settings for SSH connections. net.ssh.known-hosts Type: array of strings Default: see description Environment: not supported The known-hosts array contains a list of SSH host keys that should be accepted as valid when connecting to an SSH server (such as for SSH git dependencies). Each entry should be a string in a format similar to OpenSSH known_hosts files. Each string should start with one or more hostnames separated by commas, a space, the key type name, a space, and the base64-encoded key. For example: [net.ssh] known-hosts = [ "example.com ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFO4Q5T0UV0SQevair9PFwoxY9dl4pQl3u5phoqJH3cF" ] Cargo will attempt to load known hosts keys from common locations supported in OpenSSH, and will join those with any listed in a Cargo configuration file. If any matching entry has the correct key, the connection will be allowed. Cargo comes with the host keys for github.com built-in. If those ever change, you can add the new keys to the config or known_hosts file. See Git Authentication for more details. [patch] Just as you can override dependencies using [patch] in Cargo.toml , you can override them in the cargo configuration file to apply those patches to any affected build. The format is identical to the one used in Cargo.toml . Since .cargo/config.toml files are not usually checked into source control, you should prefer patching using Cargo.toml where possible to ensure that other developers can compile your crate in their own environments. Patching through cargo configuration files is generally only appropriate when the patch section is automatically generated by an external build tool. If a given dependency is patched both in a cargo configuration file and a Cargo.toml file, the patch in the configuration file is used. If multiple configuration files patch the same dependency, standard cargo configuration merging is used, which prefers the value defined closest to the current directory, with $HOME/.cargo/config.toml taking the lowest precedence. Relative path dependencies in such a [patch] section are resolved relative to the configuration file they appear in. [profile] The [profile] table can be used to globally change profile settings, and override settings specified in Cargo.toml . It has the same syntax and options as profiles specified in Cargo.toml . See the Profiles chapter for details about the options. [profile.<name>.build-override] Environment: CARGO_PROFILE_<name>_BUILD_OVERRIDE_<key> The build-override table overrides settings for build scripts, proc macros, and their dependencies. It has the same keys as a normal profile. See the overrides section for more details. [profile.<name>.package.<name>] Environment: not supported The package table overrides settings for specific packages. It has the same keys as a normal profile, minus the panic , lto , and rpath settings. See the overrides section for more details. profile.<name>.codegen-units Type: integer Default: See profile docs. Environment: CARGO_PROFILE_<name>_CODEGEN_UNITS See codegen-units . profile.<name>.debug Type: integer or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_DEBUG See debug . profile.<name>.split-debuginfo Type: string Default: See profile docs. Environment: CARGO_PROFILE_<name>_SPLIT_DEBUGINFO See split-debuginfo . profile.<name>.debug-assertions Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_DEBUG_ASSERTIONS See debug-assertions . profile.<name>.incremental Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_INCREMENTAL See incremental . profile.<name>.lto Type: string or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_LTO See lto . profile.<name>.overflow-checks Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_OVERFLOW_CHECKS See overflow-checks . profile.<name>.opt-level Type: integer or string Default: See profile docs. Environment: CARGO_PROFILE_<name>_OPT_LEVEL See opt-level . profile.<name>.panic Type: string Default: See profile docs. Environment: CARGO_PROFILE_<name>_PANIC See panic . profile.<name>.rpath Type: boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_RPATH See rpath . profile.<name>.strip Type: string or boolean Default: See profile docs. Environment: CARGO_PROFILE_<name>_STRIP See strip . [resolver] The [resolver] table overrides dependency resolution behavior for local development (e.g. excludes cargo install ). resolver.incompatible-rust-versions Type: string Default: See resolver docs Environment: CARGO_RESOLVER_INCOMPATIBLE_RUST_VERSIONS When resolving which version of a dependency to use, select how versions with incompatible package.rust-version s are treated. Values include: allow : treat rust-version -incompatible versions like any other version fallback : only consider rust-version -incompatible versions if no other version matched Can be overridden with --ignore-rust-version CLI option Setting the dependency’s version requirement higher than any version with a compatible rust-version Specifying the version to cargo update with --precise See the resolver chapter for more details. MSRV: allow is supported on any version fallback is respected as of 1.84 [registries] The [registries] table is used for specifying additional registries . It consists of a sub-table for each named registry. registries.<name>.index Type: string (url) Default: none Environment: CARGO_REGISTRIES_<name>_INDEX Specifies the URL of the index for the registry. registries.<name>.token Type: string Default: none Environment: CARGO_REGISTRIES_<name>_TOKEN Specifies the authentication token for the given registry. This value should only appear in the credentials file. This is used for registry commands like cargo publish that require authentication. Can be overridden with the --token command-line option. registries.<name>.credential-provider Type: string or array of path and arguments Default: none Environment: CARGO_REGISTRIES_<name>_CREDENTIAL_PROVIDER Specifies the credential provider for the given registry. If not set, the providers in registry.global-credential-providers will be used. If specified as a string, path and arguments will be split on spaces. For paths or arguments that contain spaces, use an array. If the value exists in the [credential-alias] table, the alias will be used. See Registry Authentication for more information. registries.crates-io.protocol Type: string Default: "sparse" Environment: CARGO_REGISTRIES_CRATES_IO_PROTOCOL Specifies the protocol used to access crates.io. Allowed values are git or sparse . git causes Cargo to clone the entire index of all packages ever published to crates.io from https://github.com/rust-lang/crates.io-index/ . This can have performance implications due to the size of the index. sparse is a newer protocol which uses HTTPS to download only what is necessary from https://index.crates.io/ . This can result in a significant performance improvement for resolving new dependencies in most situations. More information about registry protocols may be found in the Registries chapter . [registry] The [registry] table controls the default registry used when one is not specified. registry.index This value is no longer accepted and should not be used. registry.default Type: string Default: "crates-io" Environment: CARGO_REGISTRY_DEFAULT The name of the registry (from the registries table ) to use by default for registry commands like cargo publish . Can be overridden with the --registry command-line option. registry.credential-provider Type: string or array of path and arguments Default: none Environment: CARGO_REGISTRY_CREDENTIAL_PROVIDER Specifies the credential provider for crates.io . If not set, the providers in registry.global-credential-providers will be used. If specified as a string, path and arguments will be split on spaces. For paths or arguments that contain spaces, use an array. If the value exists in the [credential-alias] table, the alias will be used. See Registry Authentication for more information. registry.token Type: string Default: none Environment: CARGO_REGISTRY_TOKEN Specifies the authentication token for crates.io . This value should only appear in the credentials file. This is used for registry commands like cargo publish that require authentication. Can be overridden with the --token command-line option. registry.global-credential-providers Type: array Default: ["cargo:token"] Environment: CARGO_REGISTRY_GLOBAL_CREDENTIAL_PROVIDERS Specifies the list of global credential providers. If credential provider is not set for a specific registry using registries.<name>.credential-provider , Cargo will use the credential providers in this list. Providers toward the end of the list have precedence. Path and arguments are split on spaces. If the path or arguments contains spaces, the credential provider should be defined in the [credential-alias] table and referenced here by its alias. See Registry Authentication for more information. [source] The [source] table defines the registry sources available. See Source Replacement for more information. It consists of a sub-table for each named source. A source should only define one kind (directory, registry, local-registry, or git). source.<name>.replace-with Type: string Default: none Environment: not supported If set, replace this source with the given named source or named registry. source.<name>.directory Type: string (path) Default: none Environment: not supported Sets the path to a directory to use as a directory source. source.<name>.registry Type: string (url) Default: none Environment: not supported Sets the URL to use for a registry source. source.<name>.local-registry Type: string (path) Default: none Environment: not supported Sets the path to a directory to use as a local registry source. source.<name>.git Type: string (url) Default: none Environment: not supported Sets the URL to use for a git repository source. source.<name>.branch Type: string Default: none Environment: not supported Sets the branch name to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. source.<name>.tag Type: string Default: none Environment: not supported Sets the tag name to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. source.<name>.rev Type: string Default: none Environment: not supported Sets the revision to use for a git repository. If none of branch , tag , or rev is set, defaults to the master branch. [target] The [target] table is used for specifying settings for specific platform targets. It consists of a sub-table which is either a platform triple or a cfg() expression . The given values will be used if the target platform matches either the <triple> value or the <cfg> expression. [target.thumbv7m-none-eabi] linker = "arm-none-eabi-gcc" runner = "my-emulator" rustflags = ["…", "…"] [target.'cfg(all(target_arch = "arm", target_os = "none"))'] runner = "my-arm-wrapper" rustflags = ["…", "…"] cfg values come from those built-in to the compiler (run rustc --print=cfg to view) and extra --cfg flags passed to rustc (such as those defined in RUSTFLAGS ). Do not try to match on debug_assertions , test , Cargo features like feature="foo" , or values set by build scripts . If using a target spec JSON file, the <triple> value is the filename stem. For example --target foo/bar.json would match [target.bar] . target.<triple>.ar This option is deprecated and unused. target.<triple>.linker Type: string (program path) Default: none Environment: CARGO_TARGET_<triple>_LINKER Specifies the linker which is passed to rustc (via -C linker ) when the <triple> is being compiled for. By default, the linker is not overridden. target.<cfg>.linker This is similar to the target linker , but using a cfg() expression . If both a <triple> and <cfg> runner match, the <triple> will take precedence. It is an error if more than one <cfg> runner matches the current target. target.<triple>.runner Type: string or array of strings ( program path with args ) Default: none Environment: CARGO_TARGET_<triple>_RUNNER If a runner is provided, executables for the target <triple> will be executed by invoking the specified runner with the actual executable passed as an argument. This applies to cargo run , cargo test and cargo bench commands. By default, compiled executables are executed directly. target.<cfg>.runner This is similar to the target runner , but using a cfg() expression . If both a <triple> and <cfg> runner match, the <triple> will take precedence. It is an error if more than one <cfg> runner matches the current target. target.<triple>.rustflags Type: string or array of strings Default: none Environment: CARGO_TARGET_<triple>_RUSTFLAGS Passes a set of custom flags to the compiler for this <triple> . The value may be an array of strings or a space-separated string. See build.rustflags for more details on the different ways to specific extra flags. target.<cfg>.rustflags This is similar to the target rustflags , but using a cfg() expression . If several <cfg> and <triple> entries match the current target, the flags are joined together. target.<triple>.rustdocflags Type: string or array of strings Default: none Environment: CARGO_TARGET_<triple>_RUSTDOCFLAGS Passes a set of custom flags to the compiler for this <triple> . The value may be an array of strings or a space-separated string. See build.rustdocflags for more details on the different ways to specific extra flags. target.<triple>.<links> The links sub-table provides a way to override a build script . When specified, the build script for the given links library will not be run, and the given values will be used instead. [target.x86_64-unknown-linux-gnu.foo] rustc-link-lib = ["foo"] rustc-link-search = ["/path/to/foo"] rustc-flags = "-L /some/path" rustc-cfg = ['key="value"'] rustc-env = {key = "value"} rustc-cdylib-link-arg = ["…"] metadata_key1 = "value" metadata_key2 = "value" [term] The [term] table controls terminal output and interaction. term.quiet Type: boolean Default: false Environment: CARGO_TERM_QUIET Controls whether or not log messages are displayed by Cargo. Specifying the --quiet flag will override and force quiet output. Specifying the --verbose flag will override and disable quiet output. term.verbose Type: boolean Default: false Environment: CARGO_TERM_VERBOSE Controls whether or not extra detailed messages are displayed by Cargo. Specifying the --quiet flag will override and disable verbose output. Specifying the --verbose flag will override and force verbose output. term.color Type: string Default: "auto" Environment: CARGO_TERM_COLOR Controls whether or not colored output is used in the terminal. Possible values: auto (default): Automatically detect if color support is available on the terminal. always : Always display colors. never : Never display colors. Can be overridden with the --color command-line option. term.hyperlinks Type: bool Default: auto-detect Environment: CARGO_TERM_HYPERLINKS Controls whether or not hyperlinks are used in the terminal. term.unicode Type: bool Default: auto-detect Environment: CARGO_TERM_UNICODE Control whether output can be rendered using non-ASCII unicode characters. term.progress.when Type: string Default: "auto" Environment: CARGO_TERM_PROGRESS_WHEN Controls whether or not progress bar is shown in the terminal. Possible values: auto (default): Intelligently guess whether to show progress bar. always : Always show progress bar. never : Never show progress bar. term.progress.width Type: integer Default: none Environment: CARGO_TERM_PROGRESS_WIDTH Sets the width for progress bar. term.progress.term-integration Type: bool Default: auto-detect Environment: CARGO_TERM_PROGRESS_TERM_INTEGRATION Report progress to the terminal emulator for display in places like the task bar.
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Branchen-in-Git-Branches-op-afstand-Remote-branches
Git - Branches op afstand (Remote branches) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 3.5 Branchen in Git - Branches op afstand (Remote branches) Branches op afstand (Remote branches) Remote branches zijn referenties (pointers) naar jouw remote repositories, inclusief branches, tags, enz. Je kunt een volledige lijst van remote referenties expliciet krijgen met git ls-remote [remote] of ook git remote show [remote] voor zowel remote branches als voor meer informatie. Niettemin is heth gebruikelijker om te profiteren van remote-tracking branches. Remote-tracking branches zijn referenties naar de staat van remote branches. Het zijn lokale referenties die je niet kunt verplaatsen; Git verplaatst ze automatisch voor je op de momenten dat je een vorm van netwerk communicatie uitvoert, dit om te verzekeren dat ze een accurate representatie zijn van de staat van de remote repository. Remote-tracking branches gedragen zich als boekenleggers om je eraan te helpen herinneren wat de staat van de branches was van je remote repositories toen je voor het laatst met ze in contact was. Remote-tracking branch namen hebben de vorm <remote>/<branch> . Bijvoorbeeld, als je wil zien hoe de master -branch op je origin remote er uitzag de laatste keer dat je er mee communiceerde, dan zal je de origin/master -branch moeten bekijken. Als je samen met een partner aan het werk bent met een probleem en zij heeft een iss53 -branch gepusht, is het niet onmogelijk dat je zelf een eigen lokale iss53 hebt, maar de branch op de server zal vertegenwoordigd zijn door de remote-tracking branch origin/iss53 . Dit kan wat verwarrend zijn, dus laten we eens naar een voorbeeld kijken. Stel dat je een Git-server in je netwerk hebt op git.ourcompany.com . Als je hiervan kloont dan wordt die door het clone commando van Git automatisch origin voor je genoemd, Git haalt alle gegevens binnen, maakt een verwijzing naar waar de master -branch is en noemt dat lokaal origin/master . Git geeft je ook een eigen lokale master -branch, beginnend op dezelfde plaats als de master -branch van origin, zodat je iets hebt om vanaf te werken. Noot “origin” is niets speciaal Net als de branch naam “master” geen enkele speciale betekenis heeft in Git, heeft “origin” dat ook niet. Waar “master” de standaard naam is voor een branch die dient als beginpunt als je git init aanroept, wat de enige reden is waarom het zo vaak wordt gebruikt, is “origin” de standaard naam voor een remote als je git clone aanroept. Als je git clone -o booyah gebruikt, krijg je booyah/master als je standaard remote branch. Figuur 30. Server en lokale repositories na klonen Als je wat werk doet op je lokale master -branch, en in de tussentijd pusht iemand anders iets naar git.ourcompany.com waardoor die master -branch wordt vernieuwd, dan zijn jullie histories verschillend vooruit geschoven. En zolang je geen contact hebt met de origin server, zal jouw origin/master verwijzing niet verplaatsen. Figuur 31. Lokaal en remote werk kan gaan afwijken Om je werk te synchroniseren met een gegeven remote, voer je een git fetch <remote> commando uit (in ons geval, git fetch origin ). Dit commando bekijkt welke server “origin” is (in dit geval is het git.ourcompany.com ), haalt gegevens er vanaf die je nog niet hebt en vernieuwt je lokale database, waarbij je origin/master -verwijzing naar zijn nieuwe positie verplaatst wordt die meer up-to-date is. Figuur 32. git fetch vernieuwt je remote referenties Om het hebben van meerdere remote servers te demonstreren en hoe remote branches voor die remote projecten er uitzien, zullen we aannemen dat je nog een interne Git-server hebt die alleen wordt gebruikt voor ontwikkelingen gedaan door een van je sprint teams. Deze server bevindt zich op git.team1.ourcompany.com . Je kunt het als een nieuwe remote referentie toevoegen aan het project waar je nu aan werkt door het git remote add commando uit te voeren, zoals we behandeld hebben in Git Basics . Noem deze remote teamone , wat jouw afkorting voor die hele URL wordt. Figuur 33. Een andere server toevoegen als remote Nu kun je git fetch teamone uitvoeren om alles op te halen dat wat de teamone remote server heeft en jij nog niet. Omdat die server een subset heeft van de gegevens die jouw origin server op dit moment heeft, haalt Git geen gegevens op maar maakt een remote-tracking branch genaamd teamone/master aan en laat die wijzen naar de commit die teamone heeft als zijn master -branch. Figuur 34. Remote tracking branch voor teamone/master Pushen Als je een branch wil delen met de rest van de wereld, dan moet je het naar een remote terugzetten waar je schrijftoegang op hebt. Je lokale branches worden niet automatisch gesynchroniseerd met de remotes waar je naar schrijft — je moet de branches die je wilt delen expliciet pushen. Op die manier kun je privé branches gebruiken voor het werk dat je niet wil delen, en alleen die topic branches pushen waar je op wilt samenwerken. Als je een branch genaamd serverfix hebt waar je met anderen aan wilt werken, dan kun je die op dezelfde manier pushen als waarop je dat voor de eerste branch hebt gedaan. Voer git push <remote> <branch< uit: $ git push origin serverfix Counting objects: 24, done. Delta compression using up to 8 threads. Compressing objects: 100% (15/15), done. Writing objects: 100% (24/24), 1.91 KiB | 0 bytes/s, done. Total 24 (delta 2), reused 0 (delta 0) To https://github.com/schacon/simplegit * [new branch] serverfix -> serverfix Dit is wel een beetje de bocht afsnijden. Git zal de serverfix -branchnaam automatisch expanderen naar refs/heads/serverfix:refs/heads/serverfix , wat staat voor “Neem mijn lokale serverfix branch en push die om de serverfix branch van de remote te vernieuwen.”. We zullen het refs/heads gedeelte gedetaileerd behandelen in Git Binnenwerk , maar je kunt het normaalgesproken weglaten. Je kun ook git push origin serverfix:serverfix doen, wat hetzelfde doet. Dit staat voor “Neem mijn serverfix en maak het de serverfix van de remote.” Je kunt dit formaat gebruiken om een lokale branch te pushen naar een remote branch die anders heet. Als je niet wil dat het serverfix heet aan de remote kant, kan je in plaats daarvan git push origin serverfix:awesomebranch gebruiken om je lokale serverfix -branch naar de awesomebranch op het remote project te pushen. Noot Type niet elke keer je wachtwoord Als je een HTTPS URL gebruikt om mee te pushen, zal de Git server je elke keer vragen naar je usernaam en wachtwoord voor authenticatie. Standaard zal het je via de terminal vragen (prompten) om deze informatie zodat de server kan vaststellen of je mag pushen. Als je dit niet elke keer wilt intypen als je pusht, kan je een “credential cache” opzetten. Het eenvoudigste is om het gewoon in het geheugen te houden voor een aantal minuten. Dit kan je simpel opzetten door git config --global credential.helper cache aan te roepen. Voor meer informatie over de verschillende beschikbare credential caching opties, zie Het opslaan van inloggegevens . De volgende keer dat één van je medewerkers van de server fetcht zal deze een referentie krijgen naar de versie van serverfix op de server, onder de remote branch origin/serverfix : $ git fetch origin remote: Counting objects: 7, done. remote: Compressing objects: 100% (2/2), done. remote: Total 3 (delta 0), reused 3 (delta 0) Unpacking objects: 100% (3/3), done. From https://github.com/schacon/simplegit * [new branch] serverfix -> origin/serverfix Het is belangrijk om op te merken dat wanneer je een fetch doet die nieuwe remote-tracking branches ophaalt, je niet automatisch lokale aanpasbare kopieën daarvan hebt. In andere woorden, in dit geval heb je geen nieuwe serverfix -branch — je hebt alleen een origin/serverfix verwijzing die je niet kunt aanpassen. Om dit werk in je huidige werk branch te mergen, kun je git merge origin/serverfix uitvoeren. Als je een eigen serverfix -branch wilt waar je op kunt werken, dan kun je deze op je remote-tracking branch baseren: $ git checkout -b serverfix origin/serverfix Branch serverfix set up to track remote branch serverfix from origin. Switched to a new branch 'serverfix' Dit maakt een lokale branch aan waar je op kunt werken, die begint waar origin/serverfix is. Tracking branches Een lokale branch uitchecken van een remote-tracking branch creëert automatisch een zogenoemde “tracking branch” (“volg branch”, en de branch die hij volgt heet een “upstream branch”). Tracking branches zijn lokale branches die een directe releatie met een remote branch hebben. Als je op een tracking branch zit en git pull typt, dat weet Git automatisch naar welke server moet gaan om de wijzigingen op te halen en in welke branch deze moeten worden gemerged. Als je een repository kloont, zal het over het algemeen automatisch een master -branch aanmaken die origin/master trackt. Maar je kan ook andere tracking branches aanmaken als je dat wilt — andere die branches op andere remotes tracken, of niet de master -branch tracken. Een eenvoudig voorbeeld is wat je zojuist gezien hebt: git checkout -b <branch> <remotenaam>/<branch> uitvoeren. Deze operatie komt dusdanig vaak voor dat Git de --track afkorting levert: $ git checkout --track origin/serverfix Branch serverfix set up to track remote branch serverfix from origin. Switched to a new branch 'serverfix' Sterker nog, dit is zo gewoon dat er zelfs een afkorting voor de afkorting is. Als de branch naam die je uit wilt checken (a) niet bestaat en (b) dezelfde naam maar op een remote voorkomt, zal Git een tracking branch voor je aanmaken: $ git checkout serverfix Branch serverfix set up to track remote branch serverfix from origin. Switched to a new branch 'serverfix' Om een lokale branch te maken met een andere naam dan de remote branch, kun je simpelweg de eerste variant met een andere lokale branchnaam gebruiken: $ git checkout -b sf origin/serverfix Branch sf set up to track remote branch serverfix from origin. Switched to a new branch 'sf' Nu zal je lokale sf -branch automatisch pullen van origin/serverfix . Als je al een lokale branch hebt en je wilt deze koppelen aan een remote branch die je zojuist gepulld hebt, of de stroomopwaartse branch die je trackt wijzigen, kan je de -u of --set-upstream-to optie gebruiken bij de git branch om het expliciet te zetten op de momenten dat jij het wilt. $ git branch -u origin/serverfix Branch serverfix set up to track remote branch serverfix from origin. Noot Upstream verwijzing Als je een tracking branch ingericht hebt, kan je hiernaar refereren met de @{upstream} of @{u} verwijzing. Dus als je op de master -branch zit en deze origin/master trackt, kan je iets als git merge @{u} opgeven in plaats van git merge origin/master als je zou willen. Als je wilt zien welke tracking branches je ingericht hebt, kan je de -vv optie aan git branch meegeven. Dit zal jouw lokale branches afdrukken met meer informatie, inclusief wat elk van de branches trackt en of je lokale branch voorloopt, acherloopt of beide. $ git branch -vv iss53 7e424c3 [origin/iss53: ahead 2] forgot the brackets master 1ae2a45 [origin/master] deploying index fix * serverfix f8674d9 [teamone/server-fix-good: ahead 3, behind 1] this should do it testing 5ea463a trying something new Dus hier kunnen we zien dat onze iss53 origin/iss53 trackt en twee “voorloopt”, wat inhoudt dat we lokaal twee commits hebben die nog niet naar de server zijn gepusht. We kunnen ook zien dat onze master -branch origin/master trackt en up-to-date is. Vervolgens zien we dat onze serverfix -branch de server-fix-good -branch trackt op onze teamone server en drie voorloopt en een achterloopt. Dit betekent dat er een commit op de server staat die we nog niet hebben gemerged en er drie commits lokaal staan die we nog niet gepusht hebben. Tenslotte kunnen we zien dat onze testing -branch geen enkele remote branch trackt. Het is belangrijk om op te merken dat deze getallen accuraat zijn op het moment dat je voor het laatst gefetcht hebt van elke server. Dit commando maakt geen contact met de servers, het vertelt je wat het van deze servers lokaal heeft opgeslagen. Als je volledige up-to-date gegevens wilt hebben over voorsprong en achterstand zal je van alle remotes moeten fetchen voordat je dit aanroept. Je zou iets als volgt kunnen doen: $ git fetch --all; git branch -vv Pullen Waar het git fetch commando alle wijzigingen van de server zal ophalen die je nog niet hebt, zal het je werk directory helemaal niet wijzigen. Het haalt simpelweg de gegevens voor je op en laat het mergen aan jou over. Er is echter een commando git pull geheten die feitelijk een git fetch onmiddelijk gevolgd door een git merge is in de meeste gevallen. Als je een tracking branch opgezet hebt zoals in de vorige paragraaf getoond, expliciet opgezet of voor je opgezet door de clone of checkout commando’s, zal git pull kijken welke server en branch je huidige branch trackt, van die server fetchen en de betreffende remote branch daarin mergen. Over het algemeen is het beter om de fetch en merge commando’s expliciet te gebruiken omdat de magie van git pull vaak verwarrend kan zijn. Remote branches verwijderen Stel dat je klaar bent met een remote branch - zeg maar, jij en je medewerkers zijn klaar met een feature en hebben het gemerged in de master -branch van de remote (of welke branch jullie stabiele code ook in zit). Je kunt een remote branch verwijderen met de --delete optie bij git push . Als je de serverfix -branch van de server wilt verwijderen, dan voer je het volgende uit: $ git push origin --delete serverfix To https://github.com/schacon/simplegit - [deleted] serverfix Alles wat dit doet is de pointer van de server verwijderen. De Git server zal over het algemeen de gegevens nog een poos behouden totdat de garbage collection draait, dus als het per ongeluk verwijderd is, is het vaak eenvoudig terug te halen. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/pulse/2025-highlights-wins-proof-points-momentum-bmc-software-wkosc?trk=organization_guest_main-feed-card_feed-article-content
BMC 2025 | Orchestration Wins + Mainframe Modernization ROI Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Sign in to view more content Create your free account or sign in to continue your search Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Skip to main content LinkedIn Top Content People Learning Jobs Games Join now Sign in 2 min read A year-end snapshot of measurable outcomes across orchestration and mainframe modernization—plus resources to support 2026 planning. 2025 Highlights: Wins, Proof Points, and Momentum Report this article BMC Software BMC Software Published Dec 29, 2025 + Follow As 2025 wraps up, one theme stands out: the systems the world relies on don’t slow down, and neither do the teams running them. This edition is a quick, shareable look at the milestones you helped make possible, plus a few ways to keep the momentum rolling into 2026.  BMC Wrapped 2025: The year in highlights From earning the #1 spot in workload automation and orchestration (again) to supporting the enterprises that keep the global economy moving, 2025 was about helping customers orchestrate, automate, and modernize what matters most.  Mainframe momentum: ROI + recognition   Modernization isn’t just a technology initiative; it’s a business and talent strategy. Here are three quick ways to see the proof, support the community, and explore what’s possible.  Analyst spotlight: The ROI of mainframe modernization  | This analyst report examines outcomes from six enterprises that modernized their mainframe application development environments, demonstrating measurable ROI and productivity gains.  A few highlights:  217% ROI over three years   Payback in under six months   15x development velocity with intelligent automation  240% increase in junior developer hiring   50% reduction in onboarding time and 99% reduction in application downtime with instant rollbacks  Download the analyst report     Last call: DevOps Dozen voting closes soon | Our community continues to raise the bar, and BMC AMI is honored to be recognized again. If BMC AMI has helped you improve performance, reliability, or modernization outcomes this year, we’d be grateful for your vote before voting closes.  Explore in minutes: BMC AMI Cloud demos | Want a quick “show me” view of what modern resilience can look like ? These on-demand demos are a great starting point.   Orchestration momentum: customer proof + try it yourself    Orchestration is where strategy meets execution of connecting systems, teams, and data, so work moves reliably at scale. Here’s one customer example, plus a fast way to explore the platform.    Customer story spotlight: Raymond James scales orchestration with Control-M   | When growth accelerates, orchestration becomes the difference between “keeping up” and “moving ahead.” Raymond James relies on Control-M to keep critical workflows running smoothly across the business. Helping teams scale operations, reduce friction, and deliver reliable outcomes.  Raymond James scales orchestration with Control-M . Raymond James Customer Story Spotlight Explore in minutes: Control-M demo library  If you’re mapping priorities for 2026, these quick demos are a great place to start — no long setup, just practical “show me” value.  Explore orchestration use cases across hybrid environments from visibility and control to modern data pipeline workflows.    Thanks for being part of our 2025. We’re honored to help customers orchestrate, automate, and modernize the systems the world relies on every day, and we’re excited for what we’ll build together in 2026.    The BMC Briefing The BMC Briefing 68,329 followers + Subscribe Like Like Celebrate Support Love Insightful Funny Comment Copy LinkedIn Facebook X Share 113 2 Comments Noel Watson 2w Report this comment Thanks for sharing Like Reply 1 Reaction Bernard Denke 2w Report this comment Congrats! Like Reply 1 Reaction See more comments To view or add a comment, sign in More articles by BMC Software Nov 25, 2025 Control-M Ranks #1 for the 8th Consecutive Year. What EMA and Gartner® Say About Orchestration IT leaders are accelerating workload automation and service orchestration to ship faster and reduce risk across hybrid… 153 10 Comments Oct 29, 2025 BMC Named a Leader in the 2025 Gartner® Magic Quadrant™ for SOAP We’re proud to announce that BMC has been recognized as a Leader in the 2025 Gartner® Magic Quadrant™ for SOAP. In our… 126 6 Comments Aug 26, 2025 Reimagining Enterprise IT for a Connected World BMC powers the world with solutions designed to meet today’s toughest IT challenges and tomorrow’s biggest… 79 Jun 24, 2025 Whether You’re Exploring or Expanding—Take a Guided Tour of Agentic AI-Powered BMC Helix Inside the BMC Helix Blueprint Service Modeling product tour We understand you don’t have time to sift through endless… 94 May 9, 2025 Where AI innovations meet measurable business outcomes Enterprise tech isn't about buzzwords; it's about bottom lines. BMC delivers AI and automation solutions that cut costs… 160 4 Comments May 7, 2025 Redefining ServiceOps with Agentic AI At BMC Helix, we believe the future of IT is more than automation— it’s about intent-based intelligence. Today, we’re… 154 2 Comments Apr 23, 2025 Reimagining What’s Possible: Highlights from Our Annual Sales Kickoff in Las Vegas Last week, we gathered in Las Vegas for our Annual Sales Kickoff—a powerful three days packed with energy, innovation… 177 3 Comments Feb 27, 2025 AI, Data, and Automation in 2025: Insights from BMC and Industry Reports As the enterprise technology landscape evolves, several groundbreaking developments reshape how organizations approach… 54 Oct 30, 2024 Exploring the Future of Business with Generative AI, AIOps, DataOps, and More at BMC Connect In a world where technology shapes the business landscape, generative AI (GenAI) is transforming operations… 105 3 Comments Oct 18, 2024 Discovering the Future at BMC Connect 2024: Highlights and Key Insights BMC Connect 2024 has officially wrapped, and what an incredible event it has been! With over 1,000 attendees—including… 206 6 Comments Show more See all articles Explore content categories Career Productivity Finance Soft Skills & Emotional Intelligence Project Management Education Technology Leadership Ecommerce User Experience Recruitment & HR Customer Experience Real Estate Marketing Sales Retail & Merchandising Science Supply Chain Management Future Of Work Consulting Writing Economics Artificial Intelligence Employee Experience Workplace Trends Fundraising Networking Corporate Social Responsibility Negotiation Communication Engineering Hospitality & Tourism Business Strategy Change Management Organizational Culture Design Innovation Event Planning Training & Development Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language
2026-01-13T09:29:19
https://git-scm.com/book/de/v2/Git-Grundlagen-Taggen
Git - Taggen About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Erste Schritte 1.1 Was ist Versionsverwaltung? 1.2 Kurzer Überblick über die Historie von Git 1.3 Was ist Git? 1.4 Die Kommandozeile 1.5 Git installieren 1.6 Git Basis-Konfiguration 1.7 Hilfe finden 1.8 Zusammenfassung 2. Git Grundlagen 2.1 Ein Git-Repository anlegen 2.2 Änderungen nachverfolgen und im Repository speichern 2.3 Anzeigen der Commit-Historie 2.4 Ungewollte Änderungen rückgängig machen 2.5 Mit Remotes arbeiten 2.6 Taggen 2.7 Git Aliases 2.8 Zusammenfassung 3. Git Branching 3.1 Branches auf einen Blick 3.2 Einfaches Branching und Merging 3.3 Branch-Management 3.4 Branching-Workflows 3.5 Remote-Branches 3.6 Rebasing 3.7 Zusammenfassung 4. Git auf dem Server 4.1 Die Protokolle 4.2 Git auf einem Server einrichten 4.3 Erstellung eines SSH-Public-Keys 4.4 Einrichten des Servers 4.5 Git-Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Von Drittanbietern gehostete Optionen 4.10 Zusammenfassung 5. Verteiltes Git 5.1 Verteilter Arbeitsablauf 5.2 An einem Projekt mitwirken 5.3 Ein Projekt verwalten 5.4 Zusammenfassung 6. GitHub 6.1 Einrichten und Konfigurieren eines Kontos 6.2 Mitwirken an einem Projekt 6.3 Ein Projekt betreuen 6.4 Verwalten einer Organisation 6.5 Skripte mit GitHub 6.6 Zusammenfassung 7. Git Tools 7.1 Revisions-Auswahl 7.2 Interaktives Stagen 7.3 Stashen und Bereinigen 7.4 Deine Arbeit signieren 7.5 Suchen 7.6 Den Verlauf umschreiben 7.7 Reset entzaubert 7.8 Fortgeschrittenes Merging 7.9 Rerere 7.10 Debuggen mit Git 7.11 Submodule 7.12 Bundling 7.13 Replace (Ersetzen) 7.14 Anmeldeinformationen speichern 7.15 Zusammenfassung 8. Git einrichten 8.1 Git Konfiguration 8.2 Git-Attribute 8.3 Git Hooks 8.4 Beispiel für Git-forcierte Regeln 8.5 Zusammenfassung 9. Git und andere VCS-Systeme 9.1 Git als Client 9.2 Migration zu Git 9.3 Zusammenfassung 10. Git Interna 10.1 Basisbefehle und Standardbefehle (Plumbing and Porcelain) 10.2 Git Objekte 10.3 Git Referenzen 10.4 Packdateien (engl. Packfiles) 10.5 Die Referenzspezifikation (engl. Refspec) 10.6 Transfer Protokolle 10.7 Wartung und Datenwiederherstellung 10.8 Umgebungsvariablen 10.9 Zusammenfassung A1. Anhang A: Git in anderen Umgebungen A1.1 Grafische Schnittstellen A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Zusammenfassung A2. Anhang B: Git in deine Anwendungen einbetten A2.1 Die Git-Kommandozeile A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Anhang C: Git Kommandos A3.1 Setup und Konfiguration A3.2 Projekte importieren und erstellen A3.3 Einfache Snapshot-Funktionen A3.4 Branching und Merging A3.5 Projekte gemeinsam nutzen und aktualisieren A3.6 Kontrollieren und Vergleichen A3.7 Debugging A3.8 Patchen bzw. Fehlerkorrektur A3.9 E-mails A3.10 Externe Systeme A3.11 Administration A3.12 Basisbefehle 2nd Edition 2.6 Git Grundlagen - Taggen Taggen Wie die meisten VCSs hat Git die Möglichkeit, bestimmte Punkte in der Historie eines Repositorys als wichtig zu markieren. Normalerweise verwenden Leute diese Funktionalität, um Releases zu markieren ( v1.0 , v2.0 usw). In diesem Abschnitt erfährst du, wie bestehende Tags aufgelistet, Tags erstellt und gelöscht werden können sowie was die unterschiedlichen Tag-Typen sind. Deine Tags auflisten Die Auflistung der vorhandenen Tags in Git ist unkompliziert. Gib einfach git tag (mit optionalem -l oder --list ) ein: $ git tag v1.0 v2.0 Dieser Befehl listet die Tags in alphabetischer Reihenfolge auf. Die Reihenfolge, in der sie angezeigt werden, hat keine wirkliche Bedeutung. Du kannst auch nach Tags suchen, die einer bestimmten Zeichenfolge entsprechen. Das Git-Source-Repo zum Beispiel enthält mehr als 500 Tags. Wenn du nur daran interessiert bist, dir die 1.8.5-Serie anzusehen, kannst du Folgendes ausführen: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Anmerkung Das Auflisten von Tag-Wildcards erfordert die Option -l oder --list Wenn du lediglich die gesamte Liste der Tags wünschst, geht die Ausführung des Befehls git tag implizit davon aus, dass du eine Auflistung haben willst und gibt sie aus; die Verwendung von -l oder --list ist in diesem Fall optional. Wenn du jedoch ein Platzhaltermuster angibst, das mit den Tag-Namen übereinstimmt, ist die Verwendung von -l oder --list obligatorisch. Erstellen von Tags Git unterstützt zwei Arten von Tags: lightweight (d.h. nicht-annotiert) und annotated . Ein nicht-annotiertes Tag ist sehr ähnlich eines Branches, der sich nicht ändert – es ist nur ein Zeiger auf einen bestimmten Commit. Annotierte Tags werden dagegen als vollständige Objekte in der Git-Datenbank gespeichert. Sie werden mit einer Prüfsumme versehen, enthalten den Tagger-Namen, die E-Mail-Adresse und das Datum, haben eine Tagging-Nachricht und können mit GNU Privacy Guard (GPG) signiert und überprüft werden. Es wird allgemein empfohlen annotierte Tags zu erstellen, damit all diese Informationen gespeichert werden; aber wenn du ein temporäres Tag wünschst oder aus irgendwelchen Gründen die anderen Informationen nicht speichern willst, sind auch nicht-annotierte Tags möglich. Annotated Tags Das Erstellen eines annotierten Tags in Git ist einfach. Der einfachste Weg ist die Eingabe von -a , beim Ausführen des tag Befehls: $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 Ein -m spezifiziert eine Tagging-Nachricht, die mit dem Tag gespeichert wird. Wenn du keine Nachricht für ein annotierten Tag angibst, startet Git deinen Editor, damit du eine Nachricht eingeben kannst. Du kannst die Tag-Daten zusammen mit dem getaggten Commit sehen, indem du den Befehl git show verwendest: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Es werden die Tagger-Informationen, das Datum an dem der Commit getaggt wurde, und die Annotationsnachricht angezeigt, gefolgt von den Commit-Informationen. Lightweight Tags Eine weitere Möglichkeit, Commits zu markieren, ist ein leichtgewichtiger, nicht-annotierter Tag. Das ist im Grunde genommen die in einer Datei gespeicherte Commit-Prüfsumme – es werden keine weiteren Informationen gespeichert. Um einen leichtgewichtigen Tag zu erstellen, gib keine der Optionen -a , -s oder -m an, sondern nur einen Tag-Namen: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Wenn du diesmal git show auf dem Tag ausführst, siehst du keine zusätzlichen Tag-Informationen. Der Befehl zeigt nur den Commit an: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Nachträgliches Tagging Du kannst auch ältere Commits markieren. Angenommen, dein Commit-Verlauf sieht so aus: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme Nehmen wir an, du hast vergessen, das Projekt mit v1.2 beim Commit von „Update rakefile“ zu taggen. Du kannst ihn nachträglich hinzufügen. Um diesen Commit zu markieren, gib am Ende des Befehls die Commit-Prüfsumme (oder einen Teil davon) an: $ git tag -a v1.2 9fceb02 Du siehst, dass du den Commit getaggt hast: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... Tags teilen Normalerweise überträgt der Befehl git push keine Tags an den Remote-Server. Du musst Tags explizit auf einen Server verschieben, nachdem du sie erstellt hast. Dieser Prozess funktioniert genauso wie das Teilen von Remote-Branches – Du musst dazu git push origin <tagname> ausführen. $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Wenn du viele Tags hast, die du auf einmal pushen willst, kannst du auch die Option --tags mit dem Befehl git push verwenden. Dadurch werden alle deine Tags auf den Remote-Server übertragen, die sich noch nicht auf dem Server befinden. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Wenn jetzt jemand anderes aus deinem Repository klont oder pullt, erhält er auch alle deine Tags. Anmerkung git push pusht beide Arten von Tags git push <remote> --tags wird sowohl Lightweight- als auch Annotated-Tags pushen. Es gibt zur Zeit keine Möglichkeit, nur Lightweight-Tags zu pushen, aber wenn du git push <remote> --follow-tags verwendest, werden nur annotierte Tags an den Remote gepusht. Tags löschen Um einen Tag aus dem lokalen Repository zu löschen, verwende git tag -d <tagname> . Wir könnten beispielsweise den leichtgewichtigen Tag wie folgt entfernen: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Beachte, dass dadurch der Tag nicht von Remote-Servern entfernt wird. Es gibt zwei gängige Varianten, um ein Tag von einem remote Server zu löschen. Die erste Möglichkeit ist git push <remote> :refs/tags/<tagname> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw Zur Erklärung des obigen Befehls: Vor dem Doppelpunkt ist nichts und somit als Nullwert zu lesen. Darauf wird der Remote-Tag-Name gepusht, wodurch dieser gelöscht wird. Der zweite, intuitivere Weg, ein Remote-Tag zu löschen, ist mit: $ git push origin --delete <tagname> Tags auschecken Wenn du die Dateiversion anzeigen möchtest, auf die ein bestimmter Tag zeigt, kannst du git checkout auf dieses Tag durchführen. Dies versetzt dein Repository in den Zustand „detached HEAD“ (dt. losgelöst), was einige negative Nebenwirkungen hat: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image Wenn du im Zustand „detached HEAD“ Änderungen wornimmst und dann einen Commit erstellst, bleibt der Tag gleich, aber dein neuer Commit gehört zu keinem Branch und ist unzugänglich, außer mit dem genauen Commit-Hash. Wenn du also Änderungen vornehmen musst – z.B. wenn du einen Fehler in einer älteren Version behebst – wirst du normalerweise einen Branch erstellen: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Wenn du das tust und einen Commit erstellst, wird sich dein Branch version2 leicht von deinem Tag v2.0.0 unterscheiden, da er mit deinen neuen Änderungen fortschreitet, sei also vorsichtig. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.facebook.com/login/?next=https%3A%2F%2Fl.facebook.com%2Fl.php%3Fu%3Dhttps%253A%252F%252Fwww.instagram.com%252F%26amp%253Bh%3DAT2jynv_wZlhhh0Kc6i7-3C09sWVWSFDLQsCv-oL323p5C3FDdb4_-W1Vj3y34PJEC89ywL62_PxmBRaMAr9S-DSfPKk_u5S67oaU3b74ETfMROUGkfvPlPCP9NB_XMRkVyiX1SVZSj_jsq4
Facebook Facebook 이메일 또는 휴대폰 비밀번호 계정을 잊으셨나요? 새 계정 만들기 일시적으로 차단됨 일시적으로 차단됨 회원님의 이 기능 사용 속도가 너무 빠른 것 같습니다. 이 기능 사용에서 일시적으로 차단되었습니다. Back 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch 가입하기 로그인 Messenger Facebook Lite 동영상 Meta Pay Meta 스토어 Meta Quest Ray-Ban Meta Meta AI Meta AI 콘텐츠 더 보기 Instagram Threads 투표 정보 센터 개인정보처리방침 개인정보 보호 센터 정보 광고 만들기 페이지 만들기 개발자 채용 정보 쿠키 AdChoices 이용 약관 고객 센터 연락처 업로드 및 비사용자 설정 활동 로그 Meta © 2026
2026-01-13T09:29:19
https://git-scm.com/book/zh/v2/Git-%e5%9f%ba%e7%a1%80-%e8%bf%9c%e7%a8%8b%e4%bb%93%e5%ba%93%e7%9a%84%e4%bd%bf%e7%94%a8
Git - 远程仓库的使用 About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. 起步 1.1 关于版本控制 1.2 Git 简史 1.3 Git 是什么? 1.4 命令行 1.5 安装 Git 1.6 初次运行 Git 前的配置 1.7 获取帮助 1.8 总结 2. Git 基础 2.1 获取 Git 仓库 2.2 记录每次更新到仓库 2.3 查看提交历史 2.4 撤消操作 2.5 远程仓库的使用 2.6 打标签 2.7 Git 别名 2.8 总结 3. Git 分支 3.1 分支简介 3.2 分支的新建与合并 3.3 分支管理 3.4 分支开发工作流 3.5 远程分支 3.6 变基 3.7 总结 4. 服务器上的 Git 4.1 协议 4.2 在服务器上搭建 Git 4.3 生成 SSH 公钥 4.4 配置服务器 4.5 Git 守护进程 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 第三方托管的选择 4.10 总结 5. 分布式 Git 5.1 分布式工作流程 5.2 向一个项目贡献 5.3 维护项目 5.4 总结 6. GitHub 6.1 账户的创建和配置 6.2 对项目做出贡献 6.3 维护项目 6.4 管理组织 6.5 脚本 GitHub 6.6 总结 7. Git 工具 7.1 选择修订版本 7.2 交互式暂存 7.3 贮藏与清理 7.4 签署工作 7.5 搜索 7.6 重写历史 7.7 重置揭密 7.8 高级合并 7.9 Rerere 7.10 使用 Git 调试 7.11 子模块 7.12 打包 7.13 替换 7.14 凭证存储 7.15 总结 8. 自定义 Git 8.1 配置 Git 8.2 Git 属性 8.3 Git 钩子 8.4 使用强制策略的一个例子 8.5 总结 9. Git 与其他系统 9.1 作为客户端的 Git 9.2 迁移到 Git 9.3 总结 10. Git 内部原理 10.1 底层命令与上层命令 10.2 Git 对象 10.3 Git 引用 10.4 包文件 10.5 引用规范 10.6 传输协议 10.7 维护与数据恢复 10.8 环境变量 10.9 总结 A1. 附录 A: 在其它环境中使用 Git A1.1 图形界面 A1.2 Visual Studio 中的 Git A1.3 Visual Studio Code 中的 Git A1.4 IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine 中的 Git A1.5 Sublime Text 中的 Git A1.6 Bash 中的 Git A1.7 Zsh 中的 Git A1.8 PowerShell 中的 Git A1.9 总结 A2. 附录 B: 在你的应用中嵌入 Git A2.1 命令行 Git 方式 A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. 附录 C: Git 命令 A3.1 设置与配置 A3.2 获取与创建项目 A3.3 快照基础 A3.4 分支与合并 A3.5 项目分享与更新 A3.6 检查与比较 A3.7 调试 A3.8 补丁 A3.9 邮件 A3.10 外部系统 A3.11 管理 A3.12 底层命令 2nd Edition 2.5 Git 基础 - 远程仓库的使用 远程仓库的使用 为了能在任意 Git 项目上协作,你需要知道如何管理自己的远程仓库。 远程仓库是指托管在因特网或其他网络中的你的项目的版本库。 你可以有好几个远程仓库,通常有些仓库对你只读,有些则可以读写。 与他人协作涉及管理远程仓库以及根据需要推送或拉取数据。 管理远程仓库包括了解如何添加远程仓库、移除无效的远程仓库、管理不同的远程分支并定义它们是否被跟踪等等。 在本节中,我们将介绍一部分远程管理的技能。 Note 远程仓库可以在你的本地主机上 你完全可以在一个“远程”仓库上工作,而实际上它在你本地的主机上。 词语“远程”未必表示仓库在网络或互联网上的其它位置,而只是表示它在别处。 在这样的远程仓库上工作,仍然需要和其它远程仓库上一样的标准推送、拉取和抓取操作。 查看远程仓库 如果想查看你已经配置的远程仓库服务器,可以运行 git remote 命令。 它会列出你指定的每一个远程服务器的简写。 如果你已经克隆了自己的仓库,那么至少应该能看到 origin ——这是 Git 给你克隆的仓库服务器的默认名字: $ git clone https://github.com/schacon/ticgit Cloning into 'ticgit'... remote: Reusing existing pack: 1857, done. remote: Total 1857 (delta 0), reused 0 (delta 0) Receiving objects: 100% (1857/1857), 374.35 KiB | 268.00 KiB/s, done. Resolving deltas: 100% (772/772), done. Checking connectivity... done. $ cd ticgit $ git remote origin 你也可以指定选项 -v ,会显示需要读写远程仓库使用的 Git 保存的简写与其对应的 URL。 $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) 如果你的远程仓库不止一个,该命令会将它们全部列出。 例如,与几个协作者合作的,拥有多个远程仓库的仓库看起来像下面这样: $ cd grit $ git remote -v bakkdoor https://github.com/bakkdoor/grit (fetch) bakkdoor https://github.com/bakkdoor/grit (push) cho45 https://github.com/cho45/grit (fetch) cho45 https://github.com/cho45/grit (push) defunkt https://github.com/defunkt/grit (fetch) defunkt https://github.com/defunkt/grit (push) koke git://github.com/koke/grit.git (fetch) koke git://github.com/koke/grit.git (push) origin git@github.com:mojombo/grit.git (fetch) origin git@github.com:mojombo/grit.git (push) 这表示我们能非常方便地拉取其它用户的贡献。我们还可以拥有向他们推送的权限,这里暂不详述。 注意这些远程仓库使用了不同的协议。我们将会在 在服务器上搭建 Git 中了解关于它们的更多信息。 添加远程仓库 我们在之前的章节中已经提到并展示了 git clone 命令是如何自行添加远程仓库的, 不过这里将告诉你如何自己来添加它。 运行 git remote add <shortname> <url> 添加一个新的远程 Git 仓库,同时指定一个方便使用的简写: $ git remote origin $ git remote add pb https://github.com/paulboone/ticgit $ git remote -v origin https://github.com/schacon/ticgit (fetch) origin https://github.com/schacon/ticgit (push) pb https://github.com/paulboone/ticgit (fetch) pb https://github.com/paulboone/ticgit (push) 现在你可以在命令行中使用字符串 pb 来代替整个 URL。 例如,如果你想拉取 Paul 的仓库中有但你没有的信息,可以运行 git fetch pb : $ git fetch pb remote: Counting objects: 43, done. remote: Compressing objects: 100% (36/36), done. remote: Total 43 (delta 10), reused 31 (delta 5) Unpacking objects: 100% (43/43), done. From https://github.com/paulboone/ticgit * [new branch] master -> pb/master * [new branch] ticgit -> pb/ticgit 现在 Paul 的 master 分支可以在本地通过 pb/master 访问到——你可以将它合并到自己的某个分支中, 或者如果你想要查看它的话,可以检出一个指向该点的本地分支。 (我们将会在 Git 分支 中详细介绍什么是分支以及如何使用分支。) 从远程仓库中抓取与拉取 就如刚才所见,从远程仓库中获得数据,可以执行: $ git fetch <remote> 这个命令会访问远程仓库,从中拉取所有你还没有的数据。 执行完成后,你将会拥有那个远程仓库中所有分支的引用,可以随时合并或查看。 如果你使用 clone 命令克隆了一个仓库,命令会自动将其添加为远程仓库并默认以 “origin” 为简写。 所以, git fetch origin 会抓取克隆(或上一次抓取)后新推送的所有工作。 必须注意 git fetch 命令只会将数据下载到你的本地仓库——它并不会自动合并或修改你当前的工作。 当准备好时你必须手动将其合并入你的工作。 如果你的当前分支设置了跟踪远程分支(阅读下一节和 Git 分支 了解更多信息), 那么可以用 git pull 命令来自动抓取后合并该远程分支到当前分支。 这或许是个更加简单舒服的工作流程。默认情况下, git clone 命令会自动设置本地 master 分支跟踪克隆的远程仓库的 master 分支(或其它名字的默认分支)。 运行 git pull 通常会从最初克隆的服务器上抓取数据并自动尝试合并到当前所在的分支。 推送到远程仓库 当你想分享你的项目时,必须将其推送到上游。 这个命令很简单: git push <remote> <branch> 。 当你想要将 master 分支推送到 origin 服务器时(再次说明,克隆时通常会自动帮你设置好那两个名字), 那么运行这个命令就可以将你所做的备份到服务器: $ git push origin master 只有当你有所克隆服务器的写入权限,并且之前没有人推送过时,这条命令才能生效。 当你和其他人在同一时间克隆,他们先推送到上游然后你再推送到上游,你的推送就会毫无疑问地被拒绝。 你必须先抓取他们的工作并将其合并进你的工作后才能推送。 阅读 Git 分支 了解如何推送到远程仓库服务器的详细信息。 查看某个远程仓库 如果想要查看某一个远程仓库的更多信息,可以使用 git remote show <remote> 命令。 如果想以一个特定的缩写名运行这个命令,例如 origin ,会得到像下面类似的信息: $ git remote show origin * remote origin Fetch URL: https://github.com/schacon/ticgit Push URL: https://github.com/schacon/ticgit HEAD branch: master Remote branches: master tracked dev-branch tracked Local branch configured for 'git pull': master merges with remote master Local ref configured for 'git push': master pushes to master (up to date) 它同样会列出远程仓库的 URL 与跟踪分支的信息。 这些信息非常有用,它告诉你正处于 master 分支,并且如果运行 git pull , 就会抓取所有的远程引用,然后将远程 master 分支合并到本地 master 分支。 它也会列出拉取到的所有远程引用。 这是一个经常遇到的简单例子。 如果你是 Git 的重度使用者,那么还可以通过 git remote show 看到更多的信息。 $ git remote show origin * remote origin URL: https://github.com/my-org/complex-project Fetch URL: https://github.com/my-org/complex-project Push URL: https://github.com/my-org/complex-project HEAD branch: master Remote branches: master tracked dev-branch tracked markdown-strip tracked issue-43 new (next fetch will store in remotes/origin) issue-45 new (next fetch will store in remotes/origin) refs/remotes/origin/issue-11 stale (use 'git remote prune' to remove) Local branches configured for 'git pull': dev-branch merges with remote dev-branch master merges with remote master Local refs configured for 'git push': dev-branch pushes to dev-branch (up to date) markdown-strip pushes to markdown-strip (up to date) master pushes to master (up to date) 这个命令列出了当你在特定的分支上执行 git push 会自动地推送到哪一个远程分支。 它也同样地列出了哪些远程分支不在你的本地,哪些远程分支已经从服务器上移除了, 还有当你执行 git pull 时哪些本地分支可以与它跟踪的远程分支自动合并。 远程仓库的重命名与移除 你可以运行 git remote rename 来修改一个远程仓库的简写名。 例如,想要将 pb 重命名为 paul ,可以用 git remote rename 这样做: $ git remote rename pb paul $ git remote origin paul 值得注意的是这同样也会修改你所有远程跟踪的分支名字。 那些过去引用 pb/master 的现在会引用 paul/master 。 如果因为一些原因想要移除一个远程仓库——你已经从服务器上搬走了或不再想使用某一个特定的镜像了, 又或者某一个贡献者不再贡献了——可以使用 git remote remove 或 git remote rm : $ git remote remove paul $ git remote origin 一旦你使用这种方式删除了一个远程仓库,那么所有和这个远程仓库相关的远程跟踪分支以及配置信息也会一起被删除。 prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-ngeniusone/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_image-click
nGeniusONE | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in nGeniusONE Application Performance Monitoring (APM) Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Real-time Network and Application Performance Visibility The nGeniusONE® solution provides real-time visibility into any infrastructure environment - data center, private & public cloud, remote offices, and co-locations. Leveraging NETSCOUT® Smart Data, nGeniusONE delivers smarter analytics for all their applications including voice, video, data, SaaS, UCaaS, and even custom applications from wherever they are accessed - headquarters, regional locations, or home offices. nGeniusONE is the foundation of NETSCOUT’s nGenius Enterprise Performance Management solution, providing continuous network and application performance monitoring and user experience assurance to support today’s modern enterprises. The solution's insights increase IT team's understanding of consumption patterns, application utilization, and overall digital experience to better support resource optimization and capacity planning. Media Products media viewer No more previous content nGeniusONE Solution Isolate issues, troubleshoot effectively, and reduce MTTR. With the nGeniusONE solution, organizations gain end-through-end visibility to monitor network and application performance. Demo: nGeniusONE Advanced Application Troubleshooting All your applications and services are critical today - disruption to any of these can cripple business productivity. Learn how holistic visibility and analytics can assure you deliver seamless business application performance. Watch the demo. Monitoring TEAMS Voice and Video Collaboration With NETSCOUT nGeniusONE Learn how to identify the root cause of voice and video quality issues. Moving Applications From Legacy Environments to the Cloud Cloud services have never been so essential. You need immediate, actionable visibility to solve performance and availability issues impacting your critical business services. Hybrid Workforce NETSCOUT nGenius Enterprise Performance Management solutions continuously monitor network and application performance to support the hybrid workforce. Ensure quality digital user experiences and availability from anywhere with NETSCOUT. No more next content Similar products Cisco ThousandEyes Cisco ThousandEyes Application Performance Monitoring (APM) Software Datadog Datadog Application Performance Monitoring (APM) Software Splunk Cloud Platform Splunk Cloud Platform Application Performance Monitoring (APM) Software Elastic Observability Elastic Observability Application Performance Monitoring (APM) Software New Relic New Relic Application Performance Monitoring (APM) Software Sumo Logic Sumo Logic Application Performance Monitoring (APM) Software Sign in to see more Show more Show less NETSCOUT products Arbor Edge Defense Arbor Edge Defense DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/redwood-software-jscape-by-redwood/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
JSCAPE by Redwood | LinkedIn Skip to main content LinkedIn Redwood Software in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in JSCAPE by Redwood Managed File Transfer (MFT) Software by Redwood Software See who's skilled in this Add as skill Request demo Report this product About Automate your file transfers while staying secure and compliant with JSCAPE by Redwood, able to support any platform, protocol or environment. Easy to use and set up, IT teams are empowered with a single view into all file transfer operations, with real-time notifications and alerts. With a built-in library of 100+ no-code/low-code workflow automation templates, REST APIs and off-the-shelf integrations, your experts have the tools they need to deliver reliable file transfer operations. Experience the power of scalable file transfer with JSCAPE. Experience the power of secure, seamless and scalable file transfer with JSCAPE. Elevate your enterprise's data management capabilities and unlock new levels of productivity and agility today. Media Products media viewer No more previous content Admin console Network storage Trading partner JSCAPE embedded in RunMyJobs by Redwood No more next content Featured customers of JSCAPE by Redwood SAP Software Development 4,679,808 followers L'Oréal Personal Care Product Manufacturing 6,347,679 followers Starbucks Retail 3,091,714 followers Similar products Axway Managed File Transfer Axway Managed File Transfer Managed File Transfer (MFT) Software Progress MOVEit Progress MOVEit Managed File Transfer (MFT) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software MLADU MLADU Managed File Transfer (MFT) Software dDataBox dDataBox Managed File Transfer (MFT) Software Sign in to see more Show more Show less Redwood Software products ActiveBatch by Redwood ActiveBatch by Redwood Workload Automation Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software Finance Automation by Redwood Finance Automation by Redwood Financial Close Software RunMyJobs by Redwood RunMyJobs by Redwood Workload Automation Software Tidal by Redwood Tidal by Redwood Workload Automation Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/sl/v2/Osnove-Git-Ozna%c4%8devanje
Git - Označevanje About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Začetek 1.1 O nadzoru različic 1.2 Kratka zgodovina Gita 1.3 Kaj je Git? 1.4 Ukazna vrstica 1.5 Namestitev Gita 1.6 Prva nastavitev Gita 1.7 Pridobivanje pomoči 1.8 Povzetek 2. Osnove Git 2.1 Pridobivanje repozitorija Git 2.2 Snemanje sprememb v repozitorij 2.3 Pregled zgodovine potrditev 2.4 Razveljavljanje stvari 2.5 Delo z daljavami 2.6 Označevanje 2.7 Aliasi Git 2.8 Povzetek 3. Veje Git 3.1 Veje na kratko 3.2 Osnove vej in združevanja 3.3 Upravljanje vej 3.4 Poteki dela z vejami 3.5 Oddaljene veje 3.6 Ponovno baziranje 3.7 Povzetek 4. Git na strežniku 4.1 Protokoli 4.2 Pridobitev Gita na strežniku 4.3 Generiranje vaših javnih ključev SSH 4.4 Nastavitev strežnika 4.5 Prikriti proces Git 4.6 Pametni HTTP 4.7 GitWeb 4.8 GitLab 4.9 Možnosti gostovanja pri tretjih ponudnikih 4.10 Povzetek 5. Porazdeljeni Git 5.1 Porazdeljeni poteki dela 5.2 Prispevek k projektu 5.3 Vzdrževanje projekta 5.4 Povzetek 6. GitHub 6.1 Namestitev in konfiguracija računa 6.2 Prispevek k projektu 6.3 Vzdrževanje projekta 6.4 Upravljanje organizacije 6.5 Skriptni GitHub 6.6 Povzetek 7. Orodja Git 7.1 Izbira revizije 7.2 Interaktivno pripravljanje 7.3 Shranjevanje na varno (angl. stashing) in čiščenje 7.4 Podpisovanje vašega dela 7.5 Iskanje 7.6 Prepisovanje zgodovine 7.7 Demistifikacija ponastavitve 7.8 Napredno združevanje 7.9 Rerere 7.10 Razhroščevanje z Gitom 7.11 Podmoduli 7.12 Povezovanje v pakete 7.13 Zamenjava 7.14 Shramba poverilnic 7.15 Povzetek 8. Prilagoditev Gita 8.1 Konfiguracija Git 8.2 Atributi Git 8.3 Kljuke Git 8.4 Primer pravilnika, ki ga uveljavlja Git 8.5 Povzetek 9. Git in ostali sistemi 9.1 Git kot odjemalec 9.2 Migracija na Git 9.3 Povzetek 10. Notranjost Gita 10.1 Napeljava in keramika 10.2 Objekti Git 10.3 Reference Git 10.4 Packfiles (datoteke zmanjšanih podatkov) 10.5 Refspec 10.6 Protokoli prenosa 10.7 Vzdrževanje in obnovitev podatkov 10.8 Spremenljivke okolja 10.9 Povzetek A1. Dodatek A: Git v drugih okoljih A1.1 Grafični vmesniki A1.2 Git v Visual Studio A1.3 Git v Visual Studio Code A1.4 Git v IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git v Sublime Text A1.6 Git v Bashu A1.7 Git v Zsh A1.8 Git v Powershellu A1.9 Povzetek A2. Dodatek B: Vdelava Gita v vašo aplikacijo A2.1 Git v ukazni vrstici A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Dodatek C: Ukazi Git A3.1 Nastavitev in konfiguracija A3.2 Pridobivanje in ustvarjanje projektov A3.3 Osnove posnetkov A3.4 Veje in združevanje A3.5 Deljenje in posodabljanje projektov A3.6 Pregled in primerjava A3.7 Razhroščevanje A3.8 Popravljanje A3.9 E-pošta A3.10 Zunanji sistemi A3.11 Administracija A3.12 Orodja za sisteme napeljave 2nd Edition 2.6 Osnove Git - Označevanje Označevanje Kot večina VCS-jev ima Git zmožnost označevanja določenih točk v zgodovini repozitorija kot pomembne. Običajno ljudje uporabljajo to funkcionalnost za določanje točk izdaj ( v1.0 , v2.0 in tako naprej). V tem razdelku se boste naučili, kako izpisati obstoječe oznake, kako jih ustvariti in izbrisati ter kateri različni tipi oznak so na voljo. Izpisovanje vaših oznak Izpisovanje obstoječih oznak v Gitu je precej enostavno. Samo vpišite git tag (z možnostjo po izbiri -l ali --list ): $ git tag v1.0 v2.0 Ta ukaz izpiše oznake v abecednem vrstnem redu; vrstni red, v katerem se pojavijo, nima neke prave pomembnosti. Oznake, ki se prilegajo določenemu vzorcu, lahko tudi iščete. Izvorni repozitorij Git na primer vsebuje več kot 500 oznak. Če vas zanima pogledati samo serijo 1.8.5, lahko poženete to: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 Opomba Izpis oznak z maskirnim znakom zahteva možnost -l ali --list Če želite samo celoten seznam oznak, bo pogon ukaza git tag posredno predvideval, da želite izpis, in vam ga ponudil; uporaba -l ali --list v tem primeru je neobvezna. Če pa podajate vzorec z maskirnim znakom (nadomestnim znakom, angl. wildcard ), da se prilega imenom oznak, je uporaba -l ali --list obvezna. Ustvarjanje oznak Git podpira dva tipa oznak: enostavne in anotirane . Enostavna oznaka (angl. lightweight tag ) je zelo podobna veji, ki se ne spremeni — je samo kazalec na določeno potrditev. Anotirane oznake so po drugi strani shranjene kot polni objekti v podatkovni bazi Git. Imajo kontrolne vsote; vsebujejo ime, e-pošto in datum označevalca; imajo sporočilo oznake; in lahko so podpisane in preverjene z GNU Privacy Guard (GPG). V splošnem je priporočljivo, da ustvarjate anotirane oznake, da imate lahko vse te informacije; vendar če želite začasno oznako, ali zaradi kakšnega razloga ne želite ohraniti ostalih informacij, so na voljo tudi enostavne oznake. Anotirane oznake Ustvarjanje anotirane oznake v Gitu je enostavno. Najenostavnejši način je določiti -a , ko poženete ukaz tag : $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 -m določa sporočilo označevanja, ki je shranjeno skupaj z oznako. Če ne določite sporočila za anotirano oznako, Git zažene vaš urejevalnik, da ga lahko vpišete vanj. Z uporabo ukaza git show lahko pogledate podatke oznake skupaj s potrditvijo, ki je bila označena: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number To pokaže informacije označitelja, datum potrditve, ko je bila označena, in sporočilo anotacije pred prikazom informacij potrditve. Enostavne oznake Drug način označitve potrditve je enostavna oznaka. To je v osnovi kontrolna vsota potrditve shranjena v datoteki — ne ohrani se nobena druga informacija. Da ustvarite enostavno oznako, ne dodajte možnosti -a , -s ali -m , samo podajte ime oznake: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 Tokrat, če poženete git show na oznaki, ne boste videli dodatnih informacij oznake. Ukaz samo prikazuje potrditev: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number Označevanje kasneje Potrditve lahko označite tudi za tem, ko ste se prestavili preko njih. Predpostavimo, da je vaša zgodovina potrditev videti takole: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme Sedaj predpostavimo, da ste pozabili označiti projekt pri v1.2 , ki je bil pri potrditvi »Update rakefile«. Lahko jo dodate za tem. Da označite to potrditev, na koncu ukaza določite kontrolno vsoto potrditve (ali del nje): $ git tag -a v1.2 9fceb02 Vidite lahko, da ste označili potrditev: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... Deljenje oznak Privzeto, ukaz git push ne prenese oznak na oddaljene strežnike. Morali boste eksplicitno poslati oznake na deljeni strežnik za tem, ko ste jih naredili. Ta proces je enak deljenju oddaljenih vej — lahko poženete git push origin <tagname> . $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 Če imate veliko oznak, ki jih želite poslati naenkrat, lahko uporabite tudi možnost --tags pri ukazu git push . To bo na oddaljeni strežnik preneslo vse vaše oznake, ki še niso tam. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw Sedaj, ko nekdo drug klonira ali prenese iz vašega repozitorija, bo dobil tudi vse vaše oznake. Opomba git push pošlje oba tipa oznak git push <remote> --tags bo poslal tako enostavne kot anotirane oznake. Trenutno ni možnosti, da pošljete samo enostavne oznake, vendar če uporabite git push <remote> --follow-tags bodo na oddaljeni strežnik poslane samo anotirane oznake. Brisanje oznak Da izbrišete oznako na svojem lokalnem repozitoriju, lahko uporabite git tag -d <tagname> . Na primer, našo enostavno oznako zgoraj bi lahko odstranili takole: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) Bodite pozorni, saj to ne odstrani oznake iz nobenega oddaljenega strežnika. Obstajata dve pogosti variaciji za brisanje oznake iz oddaljenega strežnika. Prva variacija je git push <remote> :refs/tags/<tagname> : $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw Način za interpretacijo zgornjega je to prebrati kot, vrednost null pred znakom dvopičja je poslan na oddaljeno ime oznake, kar jo ustrezno izbriše. Drugi (in bolj intuitiven) način za brisanje oddaljene oznake je: $ git push origin --delete <tagname> Izvlečenje oznak Če želite pogledati različice datotek, na katere oznaka kaže, lahko naredite git checkout določene oznake, vendar vam to vaš repozitorij da v »stanje ločene glave« (angl. detached HEAD state ), kar ima določene stranske učinke: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image Če v »stanju ločene glave« (angl. detached HEAD state ) naredite spremembe in nato ustvarite potrditev, bo ostala oznaka enaka, vendar vaša nova potrditev ne bo pripadala nobeni veji in bo nedosegljiva, razen preko točne zgoščene vrednosti potrditve. Torej, če morate narediti spremembe — na primer, da popravljate hrošča na starejši verziji — boste na splošno želeli ustvariti vejo: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' Če to naredite in naredite potrditev, bo vaša veja version2 malenkost drugačna od vaše oznake v2.0.0 , saj ste se premaknili s spremembami naprej, torej bodite previdni. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://l.facebook.com/l.php?u=https%3A%2F%2Fwww.instagram.com%2F&h=AT0JWKYCQBUuUdfar0zJuV4_KDMzyY-EPZWYhnxp2AU0tBxiIw7otKkRgO1rzD9m6zDLsBoSJXyMpqaIrvAucWSpSdYeJ_9ww2mnRshgxwdxaQTaxbzCyG-cVhBWmibMOMzvhh3LyaFyCSKW0KFIPmSyU58RzA
Facebook Facebook 이메일 또는 휴대폰 비밀번호 계정을 잊으셨나요? 새 계정 만들기 일시적으로 차단됨 일시적으로 차단됨 회원님의 이 기능 사용 속도가 너무 빠른 것 같습니다. 이 기능 사용에서 일시적으로 차단되었습니다. Back 한국어 English (US) Tiếng Việt Bahasa Indonesia ภาษาไทย Español 中文(简体) 日本語 Português (Brasil) Français (France) Deutsch 가입하기 로그인 Messenger Facebook Lite 동영상 Meta Pay Meta 스토어 Meta Quest Ray-Ban Meta Meta AI Meta AI 콘텐츠 더 보기 Instagram Threads 투표 정보 센터 개인정보처리방침 개인정보 보호 센터 정보 광고 만들기 페이지 만들기 개발자 채용 정보 쿠키 AdChoices 이용 약관 고객 센터 연락처 업로드 및 비사용자 설정 활동 로그 Meta © 2026
2026-01-13T09:29:19
https://www.linkedin.com/uas/login?fromSignIn=true&session_redirect=https%3A%2F%2Fil.linkedin.com%2Fcompany%2Ffhenix&trk=top-card_ellipsis-menu-semaphore-sign-in-redirect&guestReportContentType=COMPANY&_f=guest-reporting
LinkedIn Login, Sign in | LinkedIn Sign in Sign in with Apple Sign in with a passkey By clicking Continue, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . or Email or phone Password Show Forgot password? Keep me logged in Sign in We’ve emailed a one-time link to your primary email address Click on the link to sign in instantly to your LinkedIn account. If you don’t see the email in your inbox, check your spam folder. Resend email Back New to LinkedIn? Join now Agree & Join LinkedIn By clicking Continue, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . LinkedIn © 2026 User Agreement Privacy Policy Community Guidelines Cookie Policy Copyright Policy Send Feedback Language العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional))
2026-01-13T09:29:19
https://www.linkedin.com/posts/collidervc_tal-s-will-be-in-ethdenver-from-february-activity-7297892248287875072-Yblb
Tal S. will be in ETHDenver from February 26 to March 1—if you're around, reach out and say hi! | Collider Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Skip to main content LinkedIn Top Content People Learning Jobs Games Sign in Join for free Collider’s Post Collider 2,951 followers 10mo Report this post Tal S. will be in ETHDenver from February 26 to March 1—if you're around, reach out and say hi! 28 Like Comment Share Copy LinkedIn Facebook X To view or add a comment, sign in 2,951 followers View Profile Connect Explore content categories Career Productivity Finance Soft Skills & Emotional Intelligence Project Management Education Technology Leadership Ecommerce User Experience Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language Sign in to view more content Create your free account or sign in to continue your search Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy .
2026-01-13T09:29:19
https://git-scm.com/book/uk/v2/%d0%9e%d1%81%d0%bd%d0%be%d0%b2%d0%b8-Git-%d0%a1%d1%82%d0%b2%d0%be%d1%80%d0%b5%d0%bd%d0%bd%d1%8f-Git-%d1%81%d1%85%d0%be%d0%b2%d0%b8%d1%89%d0%b0
Git - Створення Git-сховища About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Вступ 1.1 Про систему контролю версій 1.2 Коротка історія Git 1.3 Основи Git 1.4 Git, зазвичай, тільки додає дані 1.5 Три стани 1.6 Командний рядок 1.7 Інсталяція Git 1.8 Початкове налаштування Git 1.9 Отримання допомоги 1.10 Підсумок 2. Основи Git 2.1 Створення Git-сховища 2.2 Запис змін до репозиторія 2.3 Перегляд історії комітів 2.4 Скасування речей 2.5 Взаємодія з віддаленими сховищами 2.6 Теґування 2.7 Псевдоніми Git 2.8 Підсумок 3. Галуження в git 3.1 Гілки у кількох словах 3.2 Основи галуження та зливання 3.3 Управління гілками 3.4 Процеси роботи з гілками 3.5 Віддалені гілки 3.6 Перебазовування 3.7 Підсумок 4. Git на сервері 4.1 Протоколи 4.2 Отримання Git на сервері 4.3 Генерація вашого публічного ключа SSH 4.4 Налаштування Серверу 4.5 Демон Git 4.6 Розумний HTTP 4.7 GitWeb 4.8 GitLab 4.9 Варіанти стороннього хостингу 4.10 Підсумок 5. Розподілений Git 5.1 Розподілені процеси роботи 5.2 Внесення змін до проекту 5.3 Супроводжування проекту 5.4 Підсумок 6. GitHub 6.1 Створення та налаштування облікового запису 6.2 Як зробити внесок до проекту 6.3 Супроводжування проєкту 6.4 Керування організацією 6.5 Скриптування GitHub 6.6 Підсумок 7. Інструменти Git 7.1 Вибір ревізій 7.2 Інтерактивне індексування 7.3 Ховання та чищення 7.4 Підписання праці 7.5 Пошук 7.6 Переписування історії 7.7 Усвідомлення скидання (reset) 7.8 Складне злиття 7.9 Rerere 7.10 Зневадження з Git 7.11 Підмодулі 7.12 Пакування 7.13 Заміна 7.14 Збереження посвідчення (credential) 7.15 Підсумок 8. Налаштування Git 8.1 Конфігурація Git 8.2 Атрибути Git 8.3 Гаки (hooks) Git 8.4 Приклад політики користування виконуваної Git-ом 8.5 Підсумок 9. Git and Other Systems 9.1 Git як клієнт 9.2 Міграція на Git 9.3 Підсумок 10. Git зсередини 10.1 Кухонні та парадні команди 10.2 Об’єкти Git 10.3 Посилання Git 10.4 Файли пакунки 10.5 Специфікація посилань (refspec) 10.6 Протоколи передачі 10.7 Супроводження та відновлення даних 10.8 Змінні середовища 10.9 Підсумок A1. Додаток A: Git в інших середовищах A1.1 Графічні інтерфейси A1.2 Git у Visual Studio A1.3 Git в Eclipse A1.4 Git у Bash A1.5 Git у Zsh A1.6 Git у Powershell A1.7 Підсумок A2. Додаток B: Вбудовування Git у ваші застосунки A2.1 Git з командного рядка A2.2 Libgit2 A2.3 JGit A2.4 go-git A3. Додаток C: Команди Git A3.1 Налаштування та конфігурація A3.2 Отримання та створення проектів A3.3 Базове збереження відбитків A3.4 Галуження та зливання A3.5 Поширення й оновлення проектів A3.6 Огляд та порівняння A3.7 Зневаджування A3.8 Латання (patching) A3.9 Електронна пошта A3.10 Зовнішні системи A3.11 Адміністрування A3.12 Кухонні команди 2nd Edition 2.1 Основи Git - Створення Git-сховища Якщо ви бажаєте прочитати тільки один розділ, щоб почати працювати з Git, саме цей вам і потрібен. У цьому розділі розглядаються всі основні команди, які потрібні для переважної більшості завдань, що виникають під час роботи з Git. До кінця розділу, ви будете в змозі налаштувати й ініціалізувати репозиторій, починати і зупиняти відстеження файлів, а також готувати і вносити зміни. Ми також покажемо вам, як налаштувати Git ігнорувати певні файли чи шаблони файлів, як швидко і легко скасувати помилки, як переглядати історію своїх проектів або зміни між комітами, а також як відправляти та отримувати зміни з віддалених репозиторіїв. Створення Git-сховища Зазвичай Git репозиторій отримують одним з двох способів: Ви можете взяти локальну директорію, що наразі не під контролем версій, та перетворити її на сховище Git, або Ви можете клонувати існуючий Git репозиторій. У будь-якому разі ви отримаєте на локальній машині готове до роботи Git сховище. Ініціалізація репозиторія в існуючому каталозі Якщо у вас вже є тека з проєктом, що наразі не перебуває під контролем версії, і ви бажаєте почати використовувати з цим проєктом Git, спочатку необхідно перейти до теки цього проєкту. Якщо ви ще ніколи цього не робили, команда може трохи відрізнятися в залежності від вашої системи: для Linux: $ cd /home/user/my_project для macOS: $ cd /Users/user/my_project для Windows: $ cd C:/Users/user/my_project та виконати: $ git init Це створить новий підкаталог .git , який містить всі необхідні файли вашого репозиторія — основу Git-репозиторія. Поки у вашому проєкті ще нічого не відстежується. Див Git зсередини для отримання додаткової інформації про файли, що містяться в каталозі .git , який ви щойно створили. Якщо ви бажаєте додати існуючі файли під версійний контроль (на відміну від порожнього каталогу), ймовірно, вам слід почати відстежувати ці файли і зробити початковий коміт. Ви можете це зробити за допомогою декількох git add команд, що визначать файли, які ви плануєте відстежувати. Далі потрібно виконати git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Початкова версія проєкту' Ми розповімо що саме роблять ці команди за хвилину. Наразі, у вас є Git репозиторій з декількома відстежуваними файлами та першим комітом. Клонування існуючого репозиторія Якщо ви бажаєте отримати копію існуючого Git сховища — наприклад, проєкту, в якому ви хочете прийняти участь — вам потрібна команда git clone . Якщо ви знайомі з іншими СКВ, наприклад Subversion, ви помітите, що команда називається "clone" (клонувати), а не "checkout" (перевірити). Це важлива відмінність — замість отримання просто робочої копії, Git отримує повну копію майже всіх даних, що є у сервера. Кожна версія кожного файлу в історії проєкту витягується автоматично, коли ви виконуєте git clone . Насправді, якщо щось станеться з диском вашого серверу, ви зазвичай можете використати майже будь-який з клонів на будь-якому клієнті щоб повернути сервер до стану на момент клонування (ви можете втратити деякі серверні хуки (hook), проте усі дані під контролем версій повернуться – дивіться Отримання Git на сервері задля детальнішої інформації). Щоб клонувати репозиторій треба використати команду git clone <url> . Наприклад, якщо ви бажаєте зробити клон бібліотеки Git libgit2 , ви можете зробити це так: $ git clone https://github.com/libgit2/libgit2 Це створить теку під назвою libgit2 , проведе ініціалізацію піддиректорії .git , стяне всі дані для цього репозиторія, та приведе директорію до стану останньої версії. Якщо ви зайдете до щойно створеної директорії libgit2 , ви побачите, що всі файли проєкту на місці, готові для використання. Якщо ви бажаєте зробити клон репозиторія в директорію з іншою назвою, ви можете передати її як додатковий параметр: $ git clone https://github.com/libgit2/libgit2 mylibgit Ця команда робить те саме, що й попередня, тільки цільова директорія називається mylibgit . Git має декілька різних протоколів передачі даних, які ви можете використовувати. Попередній приклад використовує протокол https:// , проте ви також можете побачити git:// або user@server:шлях/до/репозиторія.git , що використовує SSH протокол. Отримання Git на сервері познайомить вас з усіма можливими варіантами доступу до Git репозиторія, які може мати сервер, та "за" та "проти" кожного. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Branchen-in-Git-Rebasen
Git - Rebasen About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 3.6 Branchen in Git - Rebasen Rebasen In Git zijn er twee hoofdmanieren om wijzigingen te integreren van de ene branch in een andere: de merge en de rebase . In deze paragraaf ga je leren wat rebasen is, hoe je dat moet doen, waarom het een zeer bijzonder stukje gereedschap is en in welke gevallen je het niet wilt gebruiken. De simpele rebase Als je het eerdere voorbeeld van Eenvoudig mergen (samenvoegen) erop terugslaat, dan zul je zien dat je werk is uiteengelopen en dat je commits hebt gedaan op de twee verschillende branches. Figuur 35. Eenvoudige uiteengelopen historie De simpelste manier om de branches te integreren, zoals we al hebben besproken, is het merge commando. Het voert een drieweg-merge uit tussen de twee laatste snapshots van de branches ( C3 en C4 ), en de meest recente gezamenlijke voorouder van die twee ( C2 ), en maakt een nieuw snapshot (en commit). Figuur 36. Mergen om uiteengelopen werk historie te integreren Maar, er is nog een manier: je kunt de patch van de wijziging die werd geïntroduceerd in C4 pakken en die opnieuw toepassen op C3 . In Git, wordt dit rebasen genoemd. Met het rebase commando kan je alle wijzigingen pakken die zijn gecommit op de ene branch, en ze opnieuw afspelen op een andere. In dit voorbeeld zou je het de branch experiment uitchecken, en dan op de master branch rebasen op de volgende wijze: $ git checkout experiment $ git rebase master First, rewinding head to replay your work on top of it... Applying: added staged command Deze operatie gebeurt door naar de gezamenlijke voorouder van de twee branches te gaan (degene waar je op zit en degene waar je op rebaset), de diff te nemen die geïntroduceerd is door elke losse commit op de branch waar je op zit, die diffs in tijdelijke bestanden te bewaren, de huidige branch terug te zetten naar dezelfde commit als de branch waar je op rebaset, en uiteindelijk elke diff een voor een te applyen. Figuur 37. De wijziging gemaakt in C4 rebasen naar C3 En nu kan je teruggaan naar de master branch en een fast-forward merge uitvoeren. $ git checkout master $ git merge experiment Figuur 38. De master branch fast-forwarden Nu is het snapshot waar C4' naar wijst precies dezelfde als degene waar C5 naar wees in het merge voorbeeld . Er zit geen verschil in het eindresultaat van de integratie, maar rebasen zorgt voor een duidelijkere historie. Als je de log van een branch die gerebased is bekijkt, ziet het eruit als een lineaire historie: het lijkt alsof al het werk volgorderlijk is gebeurd, zelfs wanneer het in werkelijkheid parallel eraan gedaan is. Vaak zal je dit doen om er zeker van te zijn dat je commits netjes toegepast kunnen worden op een remote branch - misschien in een project waar je aan probeert bij te dragen, maar welke je niet beheert. In dit geval zou je het werk in een branch uitvoeren en dan je werk rebasen op origin/master als je klaar ben om je patches in te sturen naar het hoofdproject. Op die manier hoeft de beheerder geen integratiewerk te doen - gewoon een fast-forward of een schone apply. Merk op dat de snapshot waar de laatste commit op het eind naar wijst, of het de laatste van de gerebasede commits voor een rebase is of de laatste merge commit na een merge, detzelfde snapshot is - alleen de historie is verschillend. Rebasen speelt veranderingen van een werklijn opnieuw af op een andere, in de volgorde waarin ze gemaakt zijn, terwijl mergen de eindresultaten pakt en die samenvoegt. Interessantere rebases Je kunt je rebase ook opnieuw laten afspelen op iets anders dan de rebase doel branch. Pak een historie zoals in Een historie met een topic branch vanaf een andere topic branch , bijvoorbeeld. Je hebt een topic branch afgesplitst ( server ) om wat server-kant functionaliteit toe te voegen aan je project en toen een keer gecommit. Daarna heb je daar vanaf gebranched om de client-kant wijzigingen te doen ( client ) en een paar keer gecommit. Als laatste, ben je teruggegaan naar je server branch en hebt nog een paar commits gedaan. Figuur 39. Een historie met een topic branch vanaf een andere topic branch Stel nu, je besluit dat je de client-kant wijzigingen wilt mergen in je hoofdlijn voor een release, maar je wilt de server-kant wijzigingen nog vasthouden totdat het verder getest is. Je kunt de wijzigingen van client pakken, die nog niet op server zitten ( C8 en C9 ) en die opnieuw afspelen op je master -branch door de --onto optie te gebruiken van git rebase : $ git rebase --onto master server client Dit zegt in feite, 'Check de client -branch uit, verzamel de patches van de gezamenlijke voorouder van de client en de server -branches, en speel die opnieuw af in de client -branch alsof deze direct afgeleid was van de master -branch. ' Het is een beetje ingewikkeld, maar het resultaat is best wel gaaf. Figuur 40. Een topic branch rebasen vanaf een andere topic branch Nu kun je een fast-forward doen van je master -branch (zie Je master branch fast-forwarden om de client branch wijzigingen mee te nemen ): $ git checkout master $ git merge client Figuur 41. Je master branch fast-forwarden om de client branch wijzigingen mee te nemen Stel dat je besluit om de server branch ook te pullen. Je kunt de server branch rebasen op de master branch zonder het eerst te hoeven uitchecken door git rebase <basisbranch> <topicbranch> uit te voeren - wat de topic branch voor je uitcheckt (in dit geval, server ) en het opnieuw afspeelt op de basis branch ( master ): $ git rebase master server Dit speelt het server werk opnieuw af op het master werk, zoals getoond in Je server branch op je master branch rebasen . Figuur 42. Je server branch op je master branch rebasen Daarna kan je de basis branch ( master ) fast-forwarden: $ git checkout master $ git merge server Je kunt de client en server -branches verwijderen, omdat al het werk geïntegreerd is en je ze niet meer nodig hebt, en de historie voor het hele proces ziet eruit zoals in Uiteindelijke commit historie : $ git branch -d client $ git branch -d server Figuur 43. Uiteindelijke commit historie De gevaren van rebasen Ahh, maar de zegeningen van rebasen zijn niet geheel zonder nadelen, samengevat in één enkele regel: Rebase geen commits die buiten je repository bekend zijn, en waar anderen werk op gebaseerd hebben. Als je die richtlijn volgt, kan je weinig gebeuren. Als je dat niet doet, zullen mensen je gaan haten en je zult door vrienden en familie uitgehoond worden. Als je spullen rebaset, zet je bestaande commits buitenspel en maak je nieuwe aan die vergelijkbaar zijn maar anders. Wanneer je commits ergens pusht en andere pullen deze en baseren daar werk op, en vervolgens herschrijf je die commits met git rebase en pusht deze weer, dan zullen je medewerkers hun werk opnieuw moeten mergen en zal het allemaal erg vervelend worden als je hun werk probeert te pullen in het jouwe. Laten we eens kijken naar een voorbeeld hoe werk rebasen dat je publiek gemaakt hebt problemen kan veroorzaken. Stel dat je van een centrale server clonet en dan daar wat werk aan doet. Je commit-historie ziet eruit als volgt: Figuur 44. Clone een repository, en baseer er wat werk op Nu doet iemand anders wat meer werk wat een merge bevat, en pusht dat werk naar de centrale server. Je fetcht dat en merget de nieuwe remote branch in jouw werk, zodat je historie eruitziet zoals dit: Figuur 45. Fetch meer commits, en merge ze in jouw werk Daarna, beslist de persoon die het werk gepusht heeft om erop terug te komen en in plaats daarvan zijn werk te gaan rebasen; hij voert een git push --force uit om de historie op de server te herschrijven. Je pullt daarna van die server, waarbij je de nieuwe commits binnen krijgt. Figuur 46. Iemand pusht gerebasede commits, daarbij commits buitenspel zettend waar jij werk op gebaseerd hebt Nu zitten jullie beiden in de penarie. Als jij een git pull doet, ga je een commit merge maken waar beide tijdslijnen in zitten, en je repository zal er zo uit zien: Figuur 47. Je merget hetzelfde werk opnieuw in een nieuwe merge commit Als je een git log uitvoert als je historie er zo uitziet, zie je twee commits die dezelfde auteur, datum en bericht hebben, wat verwarrend is. Daarnaast, als je deze historie naar de server terug pusht, zal je al deze gerebasede commits opnieuw herintroduceren op centrale server, wat weer andere mensen zou kunnen verwarren. Het is redelijk veilig om aan te nemen dat de andere ontwikkelaar C4 en C6 niet in de historie wil, dat is juist de reden waarom ze heeft gerebased. Rebaset spullen rebasen Mocht je in zo’n situatie belanden, heeft Git nog wat tovertrucs in petto die je kunnen helpen. Als iemand of een aantal mensen in jouw team met pushes wijzigingen hebben geforceerd die werk overschrijven waar jij je werk op gebaseerd hebt, is het jouw uitdaging om uit te vinden wat jouw werk is en wat zij herschreven hebben. Het komt zo uit dat naast de SHA-1 checksum van de commit, Git ook een checksum berekent die enkel is gebaseerd op de patch die is geïntroduceerd met de commit. Dit heet een “patch-id”. Als je werk pullt die was herschreven en deze rebased op de nieuwe commits van je partner, kan Git vaak succesvol uitvinden wat specifiek van jou is en deze opnieuw afspelen op de nieuwe branch. Bijvoorbeeld in het vorige scenario, als in plaats van een merge te doen we in een situatie zijn die beschreven is in Iemand pusht gerebasede commits, daarbij commits buitenspel zettend waar jij werk op gebaseerd hebt en we git rebase teamone/master aanroepen, zal Git: Bepalen welk werk uniek is in onze branch (C2, C3, C4, C6, C7) Bepalen welke geen merge commits zijn (C2, C3, C4) Bepalen welke nog niet herschreven zijn in de doel-branch (alleen C2 en C3, omdat C4 dezelfde patch is als C4') Deze commits op teamone/master afspelen Dus in plaats van het resultaat dat we zien in Je merget hetzelfde werk opnieuw in een nieuwe merge commit , zouden we eindigen met iets wat meer lijkt op Rebase op een force-pushed rebase werk. . Figuur 48. Rebase op een force-pushed rebase werk. Dit werkt alleen als de door je partner gemaakte C4 en C4' vrijwel dezelfde patch zijn. Anders kan de rebase niet achterhalen dat het een duplicaat is en zal dan een andere C4-achtige patch toevoegen (die waarschijnlijk niet schoon kan worden toegepast, omdat wijzigingen ongeveer hetzelfde daar al staan). Je kunt dit versimpelen door een git pull --rebase in plaats van een gewone git pull te draaien. Of in dit geval kan je handmatig een git fetch gevolgd door een git rebase teamone/master uitvoeren. Als je git pull gebruikt en --rebase de standaard maken, kan je de pull.rebase configuratie waarde zetten op git config --global pull.rebase true . Als je alleen maar commits rebaset die nooit buiten jouw computer bekend zijn, zou er geen vuiltje aan de lucht moeten zijn. Als je commits rebaset die zijn gepusht, maar niemand nog werk daarop heeft gebaseerd, is er ook geen probleem. Als je commits rebaset die al publiekelijk gepusht zijn, en mensen kunnen hun werk gebaseerd hebben op die commits, dan heb je de basis gelegd voor wat frustrerende problemen, en de hoon van je teamgenoten. Als jij of een partner het nodig vinden op een gegeven moment, verzeker je ervan dat iedereen weet dat ze een git pull --rebase moeten draaien om de pijn te verzachten nadat dit gebeurd is. Rebase vs. Merge Nu we rebasen en mergen in actie hebben laten zien, kan je je afvragen welk van de twee beter is. Voordat we die vraag kunnen beantwoorden, laten we eerst een stapje terug nemen en bespreken wat historie eigenlijk inhoudt. Een standpunt is dat de commit historie van jouw repository een vastlegging is van wat daadwerkelijk gebeurd is . Het is een historisch document, op zichzelf waardevol, waarmee niet mag worden gerommeld. Vanuit dit gezichtspunt, is het wijzigen van de commit historie bijna vloeken in de kerk; je bent aan het liegen over wat er werkelijk gebeurd is. Wat hindert het dat er een slorige reeks merge commits waren? Dat is hoe het gebeurd is, en de repository moet dat bewaren voor het nageslacht. Een ander standpunt is dat de commit historie het verhaal is hoe jouw project tot stand is gekomen. Je puliceert ook niet het eerste manuscript van een boek, en de handleiding hoe je software te onderhouden verdient zorgvuldig samenstellen. Dit is het kamp dat gereedschappen als rebase en filter-branch gebruikt om het verhaal te vertellen dat het beste is voor toekomstige lezers. Nu, terug naar de vraag of mergen of rebasen beter is: hopelijk snap je nu dat het niet zo eenvoudig ligt. Git is een krachtig instrument, en stelt je in staat om veel dingen te doen met en middels je historie, maar elk team en elk project is anders. Nu je weet hoe beide werken, is het aan jou om te besluiten welke het beste is voor jouw specifieke situatie. Om het beste van beide aanpakken te krijgen is het over het algemeen het beste om lokale wijzigingen die je nog niet gedeeld hebt te rebasen voordat je ze pusht zodat je verhaal het schoonste blijft, maar nooit iets te rebasen wat je elders gepusht hebt. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Gedistribueerd-Git-Bijdragen-aan-een-project
Git - Bijdragen aan een project About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 5.2 Gedistribueerd Git - Bijdragen aan een project Bijdragen aan een project De grote moeilijkheid bij het beschrijven van dit proces is dat er een enorm aantal variaties mogelijk zijn in hoe het gebeurt. Om dat Git erg flexibel is, kunnen en zullen mensen op vele manieren samenwerken, en het is lastig om te beschrijven hoe je zou moeten bijdragen aan een project — ieder project is net weer een beetje anders. Een aantal van de betrokken variabelen zijn het aantal actieve bijdragers, gekozen workflow, je commit toegang, en mogelijk de manier waarop externe bijdragen worden gedaan. De eerste variabele is het aantal actieve bijdragers. Hoeveel gebruikers dragen actief code bij aan dit project, en hoe vaak? In veel gevallen zal je twee of drie ontwikkelaars met een paar commits per dag hebben, of misschien minder voor wat meer slapende projecten. Voor zeer grote bedrijven of projecten kan het aantal ontwikkelaars in de duizenden lopen, met tientallen of zelfs honderden patches die iedere dag binnenkomen. Dit is belangrijk omdat met meer en meer ontwikkelaars, je meer en meer problemen tegenkomt bij het je verzekeren dat code netjes gepatched of eenvoudig gemerged kan worden. Wijzigingen die je indient kunnen verouderd of zwaar beschadigd raken door werk dat gemerged is terwijl je ermee aan het werken was, of terwijl je wijzigingen in de wacht stonden voor goedkeuring of toepassing. Hoe kun je jouw code consequent up-to-date en je patches geldig houden? De volgende variabele is de gebruikte workflow in het project. Is het gecentraliseerd, waarbij iedere ontwikkelaar gelijkwaardige schrijftoegang heeft tot de hoofd codebasis? Heeft het project een eigenaar of integrator die alle patches controleert? Worden alle patches ge(peer)reviewed en goedgekeurd? Ben jij betrokken bij dat proces? Is er een luitenanten systeem neergezet, en moet je je werk eerst bij hen inleveren? De volgende variabele is je commit toegang. De benodigde workflow om bij te dragen aan een project is heel anders als je schrijftoegang hebt tot het project dan wanneer je dat niet hebt. Als je geen schrijftoegang hebt, wat is de voorkeur van het project om bijdragen te ontvangen? Is er überhaupt een beleid? Hoeveel werk draag je per keer bij? Hoe vaak draag je bij? Al deze vragen kunnen van invloed zijn op hoe je effectief bijdraagt aan een project en welke workflows de voorkeur hebben of beschikbaar zijn. We zullen een aantal van deze aspecten behandelen in een aantal voorbeelden, waarbij we van eenvoudig tot complexzullen gaan. Je zou in staat moeten zijn om de specifieke workflows die je in jouw praktijk nodig hebt te kunnen herleiden naar deze voorbeelden. Commit richtlijnen Voordat we gaan kijken naar de specifieke gebruiksscenario’s, volgt hier een kort stukje over commit berichten. Het hebben van een goede richtlijn voor het maken commits en je daar aan houden maakt het werken met Git en samenwerken met anderen een stuk makkelijker. Het Git project levert een document waarin een aantal goede tips staan voor het maken van commits waaruit je patches kunt indienen — je kunt het lezen in de Git broncode in het Documentation/SubmittingPatches bestand. Als eerste wil je geen witruimte-fouten indienen. Git geeft je een eenvoudige manier om hierop te controleren — voordat je commit, voer git diff --check uit, wat mogelijke witruimte-fouten identificeert en ze voor je afdrukt. Figuur 57. Uitvoer van git diff --check . Als je dat commando uitvoert alvorens te committen, kun je al zien of je op het punt staat witruimte problemen te committen waaraan andere ontwikkelaars zich zouden kunnen ergeren. Probeer vervolgens om van elke commit een logische set wijzigingen te maken. Probeer, als het je lukt, om je wijzigingen verteerbaar te houden — ga niet het hele weekend zitten coderen op vijf verschillende problemen om dat vervolgens op maandag als één gigantische commit in te dienen. Zelfs als je gedurende het weekend niet commit, gebruik dan het staging gebied op maandag om je werk in ten minste één commit per probleem op te splitsen, met een bruikbaar bericht per commit. Als een paar van de wijzigingen hetzelfde bestand betreffen, probeer dan git add --patch te gebruiken om bestanden gedeeltelijk te stagen (in detail behandeld in Interactief stagen ). De snapshot aan de kop van het project is gelijk of je nu één commit doet of vijf, zolang alle wijzigingen op een gegeven moment maar toegevoegd zijn, probeer dus om het je mede-ontwikkelaars makkelijk te maken als ze je wijzigingen moeten beoordelen. Deze aanpak maakt het ook makkelijker om één wijziging eruit te selecteren of terug te draaien, mocht dat later nodig zijn. Geschiedenis herschrijven beschrijft een aantal handige Git trucs om geschiedenis te herschrijven en bestanden interactief te stagen — gebruik deze instrumenten als hulp om een schone en begrijpelijke historie op te bouwen voordat deze naar iemand anders wordt gestuurd. Het laatste om in gedachten te houden is het commit bericht. Als je er een gewoonte van maakt om commit berichten van goede kwaliteit aan te maken, dan maakt dat het gebruik van en samenwerken in Git een stuk eenvoudiger. In het algemeen zouden je berichten moeten beginnen met een enkele regel die niet langer is dan 50 karakters en die de wijzigingen beknopt omschrijft, gevolgd door een lege regel en daarna een meer gedetailleerde uitleg. Het Git project vereist dat de meer gedetailleerde omschrijving ook je motivatie voor de verandering bevat, en de nieuwe implementatie tegen het oude gedrag afzet — dit is een goede richtlijn om te volgen. Het is ook een goed idee om de gebiedende wijs te gebruiken in deze berichten. Met andere woorden, schrijf: "Los fout op" en niet "Fout opgelost" of "Dit lost fout op". Hier is een sjabloon dat origineel geschreven is door Tim Pope : Capitalized, short (50 chars or less) summary More detailed explanatory text, if necessary. Wrap it to about 72 characters or so. In some contexts, the first line is treated as the subject of an email and the rest of the text as the body. The blank line separating the summary from the body is critical (unless you omit the body entirely); tools like rebase can get confused if you run the two together. Write your commit message in the imperative: "Fix bug" and not "Fixed bug" or "Fixes bug." This convention matches up with commit messages generated by commands like git merge and git revert. Further paragraphs come after blank lines. - Bullet points are okay, too - Typically a hyphen or asterisk is used for the bullet, followed by a single space, with blank lines in between, but conventions vary here - Use a hanging indent Als al je commit berichten er zo uit zien, dan zullen de dingen een stuk eenvoudiger zijn voor jou en de ontwikkelaars waar je mee samenwerkt. Het Git project heeft goed geformatteerde commit berichten - ik raad je aan om git log --no-merges uit te voeren om te zien hoe een goed geformatteerde project-commit historie eruit ziet. Noot Doe wat wij zeggen, niet wat wij doen. In de volgende voorbeelden en verder door de rest van dit boek, zijn omwille van bondigheid, de berichten niet zo netjes geformateerd als dit; in plaats daarvan gebruiken we de -m optie voor git commit . Kortom: doe wat wij zeggen, niet wat wij doen. Besloten klein team De eenvoudigste opzet die je waarschijnlijk zult tegenkomen is een besloten project met één of twee andere ontwikkelaars. Met “besloten” bedoel ik gesloten broncode — zonder leestoegang voor de buitenwereld. Jij en de andere ontwikkelaars hebben allemaal push toegang op de repository. In deze omgeving kan je een workflow aanhouden die vergelijkbaar is met wat je zou doen als je Subversion of een andere gecentraliseerd systeem zou gebruiken. Je hebt nog steeds de voordelen van zaken als offline committen en veel eenvoudiger branchen en mergen, maar de workflow kan erg vergelijkbaar zijn. Het grootste verschil is dat het mergen aan de client-kant gebeurt en niet tijdens het committen aan de server-kant. Laten we eens kijken hoe het er uit zou kunnen zien als twee ontwikkelaars samen beginnen te werken met een gedeelde repository. De eerste ontwikkelaar, John, kloont de repository, maakt een wijziging, en commit lokaal. (De protocol berichten zijn met ... vervangen in deze voorbeelden om ze iets in te korten.) # John's Machine $ git clone john@githost:simplegit.git Cloning into 'simplegit'... ... $ cd simplegit/ $ vim lib/simplegit.rb $ git commit -am 'remove invalid default value' [master 738ee87] remove invalid default value 1 files changed, 1 insertions(+), 1 deletions(-) De tweede ontwikkelaar, Jessica, doet hetzelfde — kloont de repository en commit een wijziging: # Jessica's Machine $ git clone jessica@githost:simplegit.git Cloning into 'simplegit'... ... $ cd simplegit/ $ vim TODO $ git commit -am 'add reset task' [master fbff5bc] add reset task 1 files changed, 1 insertions(+), 0 deletions(-) Nu pusht Jessica haar werk naar de server, en dat werkt prima: # Jessica's Machine $ git push origin master ... To jessica@githost:simplegit.git 1edee6b..fbff5bc master -> master De laatste regel van de uitvoer hierboven laat een bruikbaar bericht zien van de push-operatie. Het basis-formaat is <oldref>..<newref> fromref -> toref , waar oldref de oude reference inhoudt, newref betekent de nieuwe referentie, fromref is de naam van de lokale referentie die gepusht wordt, en toref is de naam van de remote referentie die wordt geupdate. Je zult een vergelijkbare uitvoer zien in de behandeling hieronder, dus een basis inzicht hebben in de betekenins zal helpen met het begrijpen van de verschillende stadia van de repositories. Meer details zijn beschikbaar in de documentie van git-push . We gaan door met het voorbeeld, John maakt wat wijzigingen, commit deze naar zijn lokale repository en probeert ze naar dezelfde server te pushen: # John's Machine $ git push origin master To john@githost:simplegit.git ! [rejected] master -> master (non-fast forward) error: failed to push some refs to 'john@githost:simplegit.git' In dit geval zal de push van John falen, vanwege de eerdere push van Jessica met haar wijzigingen. Dit is belangrijk om te begrijpen als je gewend bent aan Subversion, omdat het je zal opvallen dat de twee ontwikkelaars niet hetzelfde bestand hebben aangepast. Waar Subversion automatisch zo’n merge op de server doet als verschillende bestanden zijn aangepast, moet je in Git de commits eerst lokaal mergen. Met andere wooren: John moet eerst Jessica’s wijzigingen ophalen en ze in zijn lokale repository mergen voor hij mag pushen: Als eerste stap, gaat John Jessica’s werk eerst fetchen (dit zal alleen Jessica’s werk fetchen , het zal het nog niet in John’s werk mergen): $ git fetch origin ... From john@githost:simplegit + 049d078...fbff5bc master -> origin/master Hierna ziet de lokale repository van John er ongeveer zo uit: Figuur 58. De afwijkende historie van John. Nu kan John het werk van Jessica dat hij heeft gefetcht gaan mergen in zijn lokale werk. $ git merge origin/master Merge made by the 'recursive' strategy. TODO | 1 + 1 files changed, 1 insertions(+), 0 deletions(-) Zo lang als de lokale merge gladjes verloopt, zal John’s geupdate history er zo uit zien: Figuur 59. De repository van John na origin/master te hebben gemerged. Nu zou John zijn code moeten testen om er zeker van te zijn dat alles nog steeds goed werkt, en dan kan hij zijn nieuwe gemergede werk pushen naar de server: $ git push origin master ... To john@githost:simplegit.git fbff5bc..72bbc59 master -> master Uiteindelijk ziet de commit historie van John er zo uit: Figuur 60. De historie van John na te hebben gepusht naar de origin server. In de tussentijd heeft Jessica een topic branch genaamd issue54 aangemaakt en daar drie commits op gedaan. Ze heeft John’s wijzigingen nog niet opgehaald, dus haar commit historie ziet er zo uit: Figuur 61. Topic branch van Jessica. Ineens hoort Jessica dat John nieuw werk heeft gepusht op de server en ze wil dit even gaan bekijken, zodat ze alle inhoud van de server kan ophalen dat ze nog niet heeft met: # Jessica's Machine $ git fetch origin ... From jessica@githost:simplegit fbff5bc..72bbc59 master -> origin/master Dit haalt het werk op dat John in de tussentijd gepusht heeft. De historie van Jessica ziet er nu zo uit: Figuur 62. Historie van Jessica na het fetchen van de wijzigingen van John. Jessica denkt dat haar topic branch nu klaar is, maar ze wil weten wat ze in haar werk moet mergen zodat ze kan pushen. Ze voert git log uit om dat uit te zoeken: $ git log --no-merges issue54..origin/master commit 738ee872852dfaa9d6634e0dea7a324040193016 Author: John Smith <jsmith@example.com> Date: Fri May 29 16:01:27 2009 -0700 removed invalid default value De issue54..origin/master syntax is een log filter dat Git vraagt om alleen de lijst van commits te tonen die op de laatstgenoemde branch (in dit geval origin/master ) staan die niet in de eerstegenoemde (in dit geval issue54 ) staan. We zullen deze syntax in detail bespreken in Commit reeksen . Nu zien we in de uitvoer dat er een commit is die John gemaakt heeft die Jessica nog niet gemerged heeft. Als ze origin/master merged, is dat de enige commit die haar lokale werk zal wijzigen. Nu kan Jessica het werk van haar topic branch mergen in haar master branch, het werk van John ( origin/master ) in haar master -branch mergen, en dan naar de server pushen. Eerst (nadat ze al haar werk op de issue54 topic branch heeft gecommit) schakelt Jessica terug naar haar master branch als voorbereiding op het integreren van al dit werk: $ git checkout master Switched to branch 'master' Your branch is behind 'origin/master' by 2 commits, and can be fast-forwarded. Ze kan origin/master of issue54 als eerste mergen — ze zijn beide stroomopwaarts dus de volgorde maakt niet uit. Uiteindelijk zou de snapshot gelijk moeten zijn ongeacht welke volgorde ze kiest; alleen de geschiedenis zal iets verschillen. Ze kiest ervoor om issue54 eerst te mergen: $ git merge issue54 Updating fbff5bc..4af4298 Fast forward README | 1 + lib/simplegit.rb | 6 +++++- 2 files changed, 6 insertions(+), 1 deletions(-) Er doen zich geen problemen voor, zoals je kunt zien was het een eenvoudige fast-forward. Jessica voltooit het lokale merge proces met het mergen van John’s eerder opgehaalde werk die in de origin/master -branch zit: $ git merge origin/master Auto-merging lib/simplegit.rb Merge made by the 'recursive' strategy. lib/simplegit.rb | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) Alles merget netjes, en de historie van Jessica ziet er nu zo uit: Figuur 63. Historie van Jessica na het mergen van de wijzigingen van John. Nu is origin/master bereikbaar vanuit Jessica’s master -branch, dus ze zou in staat moeten zijn om succesvol te pushen (even aangenomen dat John in de tussentijd niet weer iets gepusht heeft): $ git push origin master ... To jessica@githost:simplegit.git 72bbc59..8059c15 master -> master Iedere ontwikkelaar heeft een paar keer gecommit en elkaars werk succesvol gemerged. Figuur 64. Historie van Jessica na al haar wijzigingen naar de server te hebben gepusht. Dit is één van de eenvoudigste workflows. Je werkt een tijdje (over het algemeen in een topic branch) en merget dit in je master -branch als het klaar is om te worden geïntegreerd. Als je dat werk wilt delen, dan fetch merge je je eigen master -branch met de origin/master als die is gewijzigd, en als laatste push je naar de master -branch op de server. De algemene volgorde is als volgt: Figuur 65. Algemene volgorde van gebeurtenissen voor een eenvoudige multi-ontwikkelaar Git workflow. Besloten aangestuurd team In het volgende scenario zul je kijken naar de rol van de bijdragers in een grotere besloten groep. Je zult leren hoe te werken in een omgeving waar kleine groepen samenwerken aan functies, waarna die team-gebaseerde bijdragen worden geïntegreerd door een andere partij. Stel dat John en Jessica samen werken aan een functie (laten we zeggen “featureA”), terwijl Jessica en een derde ontwikkelaar, Josie, aan een tweede (zeg, “featureB”) aan het werken zijn. In dit geval gebruikt het bedrijf een integratie-manager achtige workflow, waarbij het werk van de individuele groepen alleen wordt geïntegreerd door bepaalde ontwikkelaars, en de master -branch van het hoofd repo alleen kan worden vernieuwd door die ontwikkelaars. In dit scenario wordt al het werk gedaan in team-gebaseerde branches en later door de integrators samengevoegd. Laten we Jessica’s workflow volgen terwijl ze aan haar twee features werkt, in parallel met twee verschillende ontwikkelaars in deze omgeving. We nemen even aan dat ze haar repository al gekloond heeft, en dat ze besloten heeft als eerste te werken aan featureA . Ze maakt een nieuwe branch aan voor de functie en doet daar wat werk: # Jessica's Machine $ git checkout -b featureA Switched to a new branch 'featureA' $ vim lib/simplegit.rb $ git commit -am 'add limit to log function' [featureA 3300904] add limit to log function 1 files changed, 1 insertions(+), 1 deletions(-) Op dit punt, moet ze haar werk delen met John, dus ze pusht haar commits naar de featureA -branch op de server. Jessica heeft geen push toegang op de master -branch — alleen de integratoren hebben dat — dus ze moet naar een andere branch pushen om samen te kunnen werken met John: $ git push -u origin featureA ... To jessica@githost:simplegit.git * [new branch] featureA -> featureA Jessica mailt John om hem te zeggen dat ze wat werk gepusht heeft in een branch genaamd featureA en dat hij er nu naar kan kijken. Terwijl ze op terugkoppeling van John wacht, besluit Jessica te beginnen met het werken aan featureB met Josie. Om te beginnen start ze een nieuwe feature branch, gebaseerd op de master -branch van de server: # Jessica's Machine $ git fetch origin $ git checkout -b featureB origin/master Switched to a new branch 'featureB' Nu doet Jessica een paar commits op de featureB -branch: $ vim lib/simplegit.rb $ git commit -am 'made the ls-tree function recursive' [featureB e5b0fdc] made the ls-tree function recursive 1 files changed, 1 insertions(+), 1 deletions(-) $ vim lib/simplegit.rb $ git commit -am 'add ls-files' [featureB 8512791] add ls-files 1 files changed, 5 insertions(+), 0 deletions(-) Jessica’s repository ziet eruit als volgt: Figuur 66. Initiële commit historie van Jessica. Ze is klaar om haar werk te pushen, maar ze krijgt een mail van Josie dat een branch met wat initieel werk voor featureB erin al gepusht is naar de server in de featureBee -branch. Jessica moet die wijzigingen eerst mergen met die van haar voordat ze kan pushen naar de server. Ze kan dan Josie’s wijzigingen ophalen met git fetch : $ git fetch origin ... From jessica@githost:simplegit * [new branch] featureBee -> origin/featureBee Aangenomen dat Jessica nog op haar uitgecheckte featureB -branch zit, kan Jessica kan dit nu mergen in het werk wat zij gedaan heeft met git merge : $ git merge origin/featureBee Auto-merging lib/simplegit.rb Merge made by the 'recursive' strategy. lib/simplegit.rb | 4 ++++ 1 files changed, 4 insertions(+), 0 deletions(-) Nu wil Jessica al dit gemergede “featureB” werk terug pushen naar de server, maar ze wil niet eenvoudigweg haar eigen featureB -branch pushen. Omdat Josie al een featureBee -branch gemaakt heeft, zou Jessica op die branch willen pushen en dat doet ze met: $ git push -u origin featureB:featureBee ... To jessica@githost:simplegit.git fba9af8..cd685d1 featureB -> featureBee Dit wordt een refspec genoemd. Zie De Refspec voor een meer gedetailleerde behandeling van Git refspecs en de verschillende dingen die je daarmee kan doen. Merk ook de -u vlag op; dit is een verkorte notatie voor -set-upstream , wat de branches voor eenvoudigere pushen en pullen op een later moment inricht. Plotseling mailt John naar Jessica om te zeggen dat hij wat wijzigingen naar de featureA -branch waar ze op samenwerken gepusht heeft, om haar te vragen die te bekijken. Wederom voert Jessica een eenvoudige git fetch uit om alle nieuwe wijzigingen van de server op te halen, inclusief (uiteraard) het werk van John: $ git fetch origin ... From jessica@githost:simplegit 3300904..aad881d featureA -> origin/featureA Jessica kan de log bekijken van John’s nieuwe werk door de inhoud van de zojuist opgehaalde featureA -branch met haar lokale kopie van dezelfde branch: $ git log featureA..origin/featureA commit aad881d154acdaeb2b6b18ea0e827ed8a6d671e6 Author: John Smith <jsmith@example.com> Date: Fri May 29 19:57:33 2009 -0700 changed log output to 30 from 25 Als wat Jessica ziet haar bevalt, kan ze het nieuwe werk van John in haar lokale featureA -branch mergen met: $ git checkout featureA Switched to branch 'featureA' $ git merge origin/featureA Updating 3300904..aad881d Fast forward lib/simplegit.rb | 10 +++++++++- 1 files changed, 9 insertions(+), 1 deletions(-) Jessica kan tot slot nog wat kleine wijzigingen aanbrengen in de gemergede inhoud, dus ze het staat haar vrij om dat te doen, deze wijzigingen committen in haar lokale featureA -branch en dan het eindresultaat terug pushen op de server. $ git commit -am 'small tweak' [featureA 774b3ed] small tweak 1 files changed, 1 insertions(+), 1 deletions(-) $ git push ... To jessica@githost:simplegit.git 3300904..774b3ed featureA -> featureA De commit historie van Jessica ziet er nu zo uit: Figuur 67. De historie van Jessica na het committen op een feature branch. Jessica, Josie en John informeren de integrators nu dat de featureA en featureBee -branches op de server klaar zijn voor integratie in de hoofdlijn. Nadat zij die branches in de hoofdlijn geïntegreerd hebben, zal een fetch de nieuwe merge commits ophalen, waardoor de commit historie er zo uit ziet: Figuur 68. De historie van Jessica na het mergen van haar beide topic branches. Veel groepen schakelen om naar Git juist vanwege de mogelijkheid om meerdere teams in parallel te kunnen laten werken, waarbij de verschillende lijnen van werk laat in het proces gemerged worden. De mogelijkheid van kleinere subgroepen of een team om samen te werken via remote branches zonder het hele team erin te betrekken of te hinderen is een enorm voordeel van Git. De volgorde van de workflow die je hier zag is ongeveer dit: Figuur 69. Eenvoudige volgorde in de workflow van dit aangestuurde team. Gevorkt openbaar project Het bijdragen aan openbare, of publieke, projecten gaat op een iets andere manier. Omdat je niet de toestemming hebt om de branches van het project rechtstreeks te updaten, moet je het werk op een andere manier naar de beheerders krijgen. Dit eerste voorbeeld beschrijft het bijdragen via afsplitsen (forken) op Git hosts die het eenvoudig aanmaken van forks ondersteunen. Vele hosting sites ondersteunen dit (waaronder GitHub, BitBucket, Google Code, repo.or.cz, en andere), en veel project beheerders verwachten deze manier van bijdragen. De volgende paragraaf behandelt projecten die de voorkeur hebben om bijdragen in de vorm van patches via e-mail te ontvangen. Eerst zal je waarschijnlijk de hoofdrepository klonen, een topic branch maken voor de patch of reeks patches die je van plan bent bij te dragen, en je werk daarop doen. De te volgen stappen zien er in principe zo uit: $ git clone <url> $ cd project $ git checkout -b featureA ... work ... $ git commit ... work ... $ git commit Noot Je kunt eventueel besluiten rebase -i te gebruiken om je werk in één enkele commit samen te persen (squash), of het werk in de commits te herschikken om de patch eenvoudiger te kunnen laten reviewen door de beheerders — zie Geschiedenis herschrijven voor meer informatie over het interactief rebasen. Als je werk op de branch af is, en je klaar bent om het over te dragen aan de beheerders, ga je naar de originele project pagina en klik op de “Fork” knop. Hiermee maak je een eigen overschrijfbare fork van het project. Je moet de URL van deze nieuwe repository URL toevoegen als een tweede remote, en laten we deze myfork noemen: $ git remote add myfork <url> Dan moet je je werk daar naartoe pushen. Het is het makkelijkst om de topic branch waar je op zit te werken te pushen naar je repository, in plaats van het te mergen in je master branch en die te pushen. De reden hiervan is, dat als het werk niet wordt geaccepteerd of alleen ge-cherry picked (deels overgenomen), je jouw master branch niet hoeft terug te draaien (de Git cherry-pick operatie wordt meer gedetailleerd behandeld in Rebasende en cherry pick workflows ). Als de beheerders je werk mergen, rebasen of cherry picken, dan krijg je het uiteindelijk toch binnen door hun repository te pullen. Hoe dan ook, je kunt je werk pushen met: $ git push -u myfork featureA Als jouw werk gepusht is naar jouw fork van de repository, dan moet je de beheerder van het oorspronkelijke project inlichten dat je werk hebt dat je ze wil laten mergen. Dit wordt een pull request (haal-binnen-verzoek) genoemd, en je kunt deze via de website genereren - GitHub heeft een eigen “Pull Request” mechanisme die we verder zullen behandelen in GitHub of je roept het git request-pull commando aan en stuurt een mail met de uitvoer handmatig naar de projectbeheerder. Het request-pull commando neemt de basis branch waarin je de topic branch gepulld wil krijgen, en de URL van de Git repository waar je ze uit wil laten pullen, en maakt een samenvatting van alle wijzigingen die je gepulld wenst te hebben. Bijvoorbeeld, als Jessica John een pull request wil sturen, en ze heeft twee commits gedaan op de topic branch die ze zojuist gepusht heeft, dan kan ze dit uitvoeren: $ git request-pull origin/master myfork The following changes since commit 1edee6b1d61823a2de3b09c160d7080b8d1b3a40: Jessica Smith (1): added a new function are available in the git repository at: git://githost/simplegit.git featureA Jessica Smith (2): add limit to log function change log output to 30 from 25 lib/simplegit.rb | 10 +++++++++- 1 files changed, 9 insertions(+), 1 deletions(-) De uitvoer kan naar de beheerders gestuurd worden — het vertelt ze waar het werk vanaf gebrancht is, vat de commits samen en vertelt waar vandaan ze dit werk kunnen pullen. Bij een project waarvan je niet de beheerder bent, is het over het algemeen eenvoudiger om een branch zoals master altijd de origin/master te laten tracken, en je werk te doen in topic branches die je eenvoudig weg kunt gooien als ze geweigerd worden. Als je je werkthema’s gescheiden houdt in topic branches maakt dat het ook eenvoudiger voor jou om je werk te rebasen als de punt van de hoofd-repository in de tussentijd verschoven is en je commits niet langer netjes toegepast kunnen worden. Bijvoorbeeld, als je een tweede onderwerp wilt bijdragen aan een project, ga dan niet verder werken op de topic branch die je zojuist gepusht hebt — begin opnieuw vanaf de master -branch van de hoofd repository: $ git checkout -b featureB origin/master ... work ... $ git commit $ git push myfork featureB $ git request-pull origin/master myfork ... email generated request pull to maintainer ... $ git fetch origin Nu zijn al je onderwerpen opgeslagen in een silo — vergelijkbaar met een patch reeks (queue) — die je kunt herschrijven, rebasen en wijzigen zonder dat de onderwerpen elkaar beïnvloeden of van elkaar afhankelijk, hier is hoe je het kunt doen: Figuur 70. Initiële commit historie met werk van featureB . Stel dat de project-beheerder een verzameling andere patches binnengehaald heeft en jouw eerste branch geprobeerd heeft, maar dat die niet meer netjes merged. In dat geval kun je proberen die branch te rebasen op origin/master , de conflicten op te lossen voor de beheerder, en dan je wijzigingen opnieuw aanbieden: $ git checkout featureA $ git rebase origin/master $ git push -f myfork featureA Dit herschrijft je historie zodat die eruit ziet als in Commit historie na werk van featureA . . Figuur 71. Commit historie na werk van featureA . Omdat je de branch gerebased hebt, moet je de -f specificeren met je push commando om in staat te zijn de featureA -branch op de server te vervangen met een commit die er geen afstammeling van is. Een alternatief zou zijn dit nieuwe werk naar een andere branch op de server te pushen (misschien featureAv2 genaamd). Laten we eens kijken naar nog een mogelijk scenario: de beheerder heeft je werk bekeken in je tweede branch en vindt het concept goed, maar zou willen dat je een implementatie detail verandert. Je gebruikt deze gelegenheid meteen om het werk te baseren op de huidige master -branch van het project. Je begint een nieuwe branch gebaseerd op de huidige origin/master -branch, squasht de featureB wijzigingen er naartoe, lost eventuele conflicten op, doet de implementatie wijziging en pusht deze terug als een nieuwe branch: $ git checkout -b featureBv2 origin/master $ git merge --squash featureB ... change implementation ... $ git commit $ git push myfork featureBv2 De --squash optie pakt al het werk op de gemergde branch en perst dat samen in één wijzigingsset (changeset) die de staat van de repository geeft alsof er een echte merge zou hebben plaatsgevonden, zonder feitelijk een merge commit te doen. Dit betekent dat je toekomstige commit maar één ouder heeft en geeft je de ruimte om alle wijzigingen te introduceren uit een andere branch en daarna meer wijzigingen te maken voordat de nieuwe commit wordt vastgelegd. Daarnaast kan de --no-commit handig zijn door de merge commit uit te stellen in plaats van het standaard merge proces. Je kunt de beheerder nu een bericht sturen dat je de gevraagde wijzigingen gemaakt hebt en dat ze die wijzigingen kunnen vinden in je featureBv2 -branch. Figuur 72. Commit historie na het featureBv2 werk. Openbaar project per e-mail Veel projecten hebben vastgestelde procedures voor het accepteren van patches — je zult de specifieke regels voor elk project moeten controleren, ze zullen namelijk verschillen. Omdat er verscheidene oudere, grotere projecten zijn die patches accepteren via ontwikkelaar-maillijsten, zullen we nu een voorbeeld hiervan behandelen. De workflow is vergelijkbaar met het vorige geval — je maakt topic branches voor iedere patch waar je aan werkt. Het verschil is hoe je die aanlevert bij het project. In plaats van het project te forken en naar je eigen schrijfbare versie te pushen, genereer je e-mail versies van iedere reeks commits en mailt die naar de ontwikkelaar-maillijst: $ git checkout -b topicA ... work ... $ git commit ... work ... $ git commit Nu heb je twee commits die je naar de maillijst wilt sturen. Je gebruikt git format-patch om de mbox-geformatteerde bestanden te genereren die je kunt mailen naar de lijst. Dit maakt van iedere commit een e-mail bericht met de eerste regel van het commit bericht als het onderwerp, en de rest van het bericht plus de patch die door de commit wordt geïntroduceerd als de inhoud. Het prettige hieraan is dat met het toepassen van een patch uit een mail die gegenereerd is met format-patch alle commit informatie blijft behouden. $ git format-patch -M origin/master 0001-add-limit-to-log-function.patch 0002-changed-log-output-to-30-from-25.patch Het format-patch commando drukt de namen af van de patch bestanden die het maakt. De -M optie vertelt Git te letten op naamswijzigingen. De bestanden komen er uiteindelijk zo uit te zien: $ cat 0001-add-limit-to-log-function.patch From 330090432754092d704da8e76ca5c05c198e71a8 Mon Sep 17 00:00:00 2001 From: Jessica Smith <jessica@example.com> Date: Sun, 6 Apr 2008 10:17:23 -0700 Subject: [PATCH 1/2] add limit to log function Limit log functionality to the first 20 --- lib/simplegit.rb | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/lib/simplegit.rb b/lib/simplegit.rb index 76f47bc..f9815f1 100644 --- a/lib/simplegit.rb +++ b/lib/simplegit.rb @@ -14,7 +14,7 @@ class SimpleGit end def log(treeish = 'master') - command("git log #{treeish}") + command("git log -n 20 #{treeish}") end def ls_tree(treeish = 'master') -- 2.1.0 Je kunt deze patch bestanden ook aanpassen om meer informatie, die je niet in het commit bericht wilt laten verschijnen, voor de maillijst toe te voegen. Als je tekst toevoegt tussen de --- regel en het begin van de patch (de diff --git regel), dan kunnen ontwikkelaars dit lezen, maar tijdens het toepassen van de patch wordt dit genegeerd. Om dit te mailen naar een maillijst, kan je het bestand in je mail-applicatie plakken of het sturen via een commandoregel programma. Het plakken van de tekst veroorzaakt vaak formaterings problemen, in het bijzonder bij “slimmere” clients die de newlines en andere witruimte niet juist behouden. Gelukkig levert Git een gereedschap die je helpt om juist geformatteerde patches via IMAP te versturen, wat het alweer een stuk makkelijker voor je kan maken. We zullen je laten zien hoe je een patch via Gmail stuurt, wat de mail-applicatie is waar we het meest bekend mee zijn. Je kunt gedetailleerde instructies voor een aantal mail programma’s vinden aan het eind van het eerder genoemde Documentation/SubmittingPatches bestand in de Git broncode. Eerst moet je de imap sectie in je ~/.gitconfig bestand instellen. Je kunt iedere waarde apart instellen met een serie git config commando’s, of je kunt ze handmatig toevoegen, maar uiteindelijk moet je config bestand er ongeveer zo uitzien: [imap] folder = "[Gmail]/Drafts" host = imaps://imap.gmail.com user = user@gmail.com pass = YX]8g76G_2^sFbd port = 993 sslverify = false Als je IMAP server geen SSL gebruikt, zijn de laatste twee regels waarschijnlijk niet nodig, en de waarde voor host zal imap:// zijn in plaats van imaps:// . Als dat ingesteld is, kun je git imap-send gebruiken om de patch reeks in de Drafts map van de gespecificeerde IMAP server te zetten: $ cat *.patch |git imap-send Resolving imap.gmail.com... ok Connecting to [74.125.142.109]:993... ok Logging in... sending 2 messages 100% (2/2) done Nu zou je in staat moeten zijn om naar je Drafts folder te gaan, het To veld te veranderen in het adres van de mail lijst waar je de patch naar toe stuurt, en mogelijk de onderhouder of de persoon verantwoordelijk voor dat deel te CC-en, en het te versturen. Je kunt de patches ook via een SMTP server sturen. Net als hierboven, kan je elke waarde apart zetten met een reeks van git config commando’s, of je kunt ze handmatig in de sendemail sectie in je ~/.gitconfig bestand toevoegen: [sendemail] smtpencryption = tls smtpserver = smtp.gmail.com smtpuser = user@gmail.com smtpserverport = 587 Als dit gedaan is, kan je git send-email gebruiken om je patches te sturen: $ git send-email *.patch 0001-added-limit-to-log-function.patch 0002-changed-log-output-to-30-from-25.patch Who should the emails appear to be from? [Jessica Smith <jessica@example.com>] Emails will be sent from: Jessica Smith <jessica@example.com> Who should the emails be sent to? jessica@example.com Message-ID to be used as In-Reply-To for the first email? y Dan zal Git een berg log-informatie oplepelen die er ongeveer zo uitziet voor elke patch die je aan het versturen bent: (mbox) Adding cc: Jessica Smith <jessica@example.com> from \line 'From: Jessica Smith <jessica@example.com>' OK. Log says: Sendmail: /usr/sbin/sendmail -i jessica@example.com From: Jessica Smith <jessica@example.com> To: jessica@example.com Subject: [PATCH 1/2] added limit to log function Date: Sat, 30 May 2009 13:29:15 -0700 Message-Id: <1243715356-61726-1-git-send-email-jessica@example.com> X-Mailer: git-send-email 1.6.2.rc1.20.g8c5b.dirty In-Reply-To: <y> References: <y> Result: OK Samenvatting In dit hoofdstuk is een aantal veel voorkomende workflows behandeld, die je kunt gebruiken om te kunnen werken in een aantal zeer verschillende typen Git projecten die je kunt tegenkomen. En er zijn een aantal nieuwe gereedschappen geïntroduceerd die je helpen om dit proces te beheren. Wat hierna volgt zal je laten zien hoe je aan de andere kant van de tafel werkt: een Git project beheren. Je zult leren hoe een welwillende dictator of integratie manager te zijn. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/freshworks-inc-freshchat/?trk=products_seo_search
Freshchat | LinkedIn Skip to main content LinkedIn Freshworks in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Freshchat Live Chat Software by Freshworks See who's skilled in this Add as skill Learn more Report this product About Freshchat helps B2C businesses have smarter conversations with their customers to help convert, engage, and support them, wherever they are. Our solution helps you connect with customers across channels, provide intuitive deflection with AI-powered chatbots, and enable personalized and proactive self-service at scale. This product is intended for Customer Service Representative Customer Service Manager Customer Support Specialist Marketing Representative Vice President Customer Service Customer Support Representative Product Marketing Manager Senior Product Marketing Manager Media Products media viewer No more previous content Meet Freshchat Freshchat uses conversational engagement, AI-powered chat, and bots to anticipate the needs of your customers. Our single solution makes onboarding a breeze, giving you more time to delight your customers. Grow customer love with smarter conversations. Today’s relationships are digital and always on. Power your business with intuitive and intelligent conversations. Scale support with AI-driven bots on any channel Customers seek quick and effortless self-service on the channels of their choice. Easily create and deploy AI-driven bot experiences across your website, native app, and messaging channels. Simplified and contextual agent experiences. Give agents real-time context to perform faster and with ease. Modern, cost-effective support Adopt a digital-first approach to resolve faster, control costs, and scale customer service. No more next content Featured customers of Freshchat Klarna Software Development 390,433 followers Travix Software Development 16,312 followers Discover Financial Services 234,651 followers Supr Daily by Swiggy Technology, Information and Internet 26,890 followers FinAccel Financial Services 65,708 followers KaafMeem Retail Apparel and Fashion 2,719 followers KKday Travel Arrangements 68,182 followers Musafir.com Travel Arrangements 64,614 followers GetYourGuide Software Development 105,905 followers Zoomcar Travel Arrangements 58,259 followers instantprintuk Printing Services 2,823 followers CISS S.A - Gestão para o Varejo IT Services and IT Consulting 21,693 followers Dunzo Technology, Information and Internet 356,114 followers HP IT Services and IT Consulting 5,229,473 followers Decathlon Sports India Retail 330,390 followers Harvey Norman Retail 44,761 followers Blue Nile Retail 33,285 followers MakeMyTrip Technology, Information and Internet 992,004 followers Show more Show less Similar products ChatR ChatR Live Chat Software ICE Chat ICE Chat Live Chat Software Webchat Webchat Live Chat Software Microsoft Teams Microsoft Teams Live Chat Software LiveChat LiveChat Live Chat Software Chat Chat Live Chat Software Sign in to see more Show more Show less Freshworks products Freshdesk Freshdesk Help Desk Software Freshdesk Omni Freshdesk Omni Help Desk Software Freshmarketer Freshmarketer Marketing Automation Software Freshsales Freshsales Customer Relationship Management (CRM) Software Freshservice Freshservice Service Desk Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Gedistribueerd-Git-Gedistribueerde-workflows
Git - Gedistribueerde workflows About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 5.1 Gedistribueerd Git - Gedistribueerde workflows Nu je een remote Git repository hebt ingesteld als een plaats waar alle ontwikkelaars hun code kunnen delen, en je bekend bent met fundamentele Git commando’s in een lokale workflow, zul je hier zien hoe je enkele van de gedistribueerde workflows kunt gebruiken waar Git je toe in staat stelt. In dit hoofdstuk zul je zien hoe je met Git kunt werken in een gedistribueerde omgeving als een bijdrager (contributor) en als een integrator. Dat wil zeggen, je zult leren hoe je succesvol code kunt bijdragen aan een project en hoe je het zo makkelijk mogelijk maakt voor jou en de beheerder van het project, en ook hoe je een project succesvol kunt onderhouden waar een aantal ontwikkelaars aan bijdragen. Gedistribueerde workflows In tegenstelling tot gecentraliseerde versiebeheersystemen (CVCSen), stelt de gedistribueerde aard van Git je in staat om veel flexibeler te zijn in de manier waarop ontwikkelaars samenwerken in projecten. Bij gecentraliseerde systemen is iedere ontwikkelaar een knooppunt dat min of meer gelijkwaardig werkt op een centraal punt. In Git echter is iedere ontwikkelaar zowel een knooppunt als een spil — dat wil zeggen, iedere ontwikkelaar kan zowel code bijdragen aan andere repositories, als ook een publiek repository beheren waarop andere ontwikkelaars hun werk baseren en waaraan zij kunnen bijdragen. Dit stelt je project en/of je team in staat om een enorm aantal workflows er op na te houden, dus ik zal een aantal veel voorkomende manieren behandelen die gebruik maken van deze flexibiliteit. We zullen de sterke en mogelijke zwakke punten van ieder ontwerp behandelen; je kunt er een kiezen om te gebruiken, of je kunt van iedere wijze een paar eigenschappen overnemen en mengen. Gecentraliseerde workflow In gecentraliseerde systemen is er over het algemeen een enkel samenwerkingsmodel — de gecentraliseerde workflow. Eén centraal punt, of repository , kan code aanvaarden en iedereen synchroniseert zijn werk daarmee. Een aantal ontwikkelaars zijn knopen — gebruikers van dat centrale punt - en synchroniseren met die ene plaats. Figuur 54. Gecentraliseerde workflow. Dit houdt in dat als twee ontwikkelaars klonen van het gecentraliseerde punt en beiden wijzigingen doen, de eerste ontwikkelaar zijn wijzigingen zonder problemen kan pushen. De tweede ontwikkelaar zal het werk van de eerste in het zijne moeten mergen voordat hij het zijne kan pushen, om zo niet het werk van de eerste te overschrijven. Dit concept werkt in Git zoals het ook werkt in Subversion (of ieder ander CVCS), en dit model werkt prima in Git. Als je al vertrouwd bent met een gecentraliseerde workflow in je bedrijf of team, dan kun je eenvoudigweg doorgaan met het gebruiken van die workflow met Git. Stel eenvoudigweg een enkele repository in, en geef iedereen in je team push-toegang; Git zal gebruikers niet toestaan om elkaars wijzigingen te overschrijven. Stel dat John en Jessica beiden tegelijkertijd beginnen te werken. John is klaar met zijn wijziging en pusht deze naar de server. Vervolgens probeert Jessica haar wijzigingen te pushen, maar de server weigert deze. Haar wordt verteld dat ze non-fast-forward wijzigingen probeert te pushen, en dat ze niet kan committen totdat ze gefetcht en gemerged heeft. Deze workflow spreekt veel mensen aan omdat het een werkwijze is waarmee veel mensen bekend zijn en zich hierbij op hun gemak voelen. Deze workflow is echter niet beperkt tot alleen kleine teams. Met het branching model van Git is het mogelijk om honderden ontwikkelaars tegelijktijdig succesvol te laten werken op een enkel project middels tientallen branches. Integratie-manager workflow Omdat Git je toestaat om meerdere remote repositories te hebben, is het mogelijk om een workflow te hebben waarbij iedere ontwikkelaar schrijftoegang heeft tot zijn eigen publieke repository en leestoegang op die van de anderen. In dit scenario is er vaak een gezagdragend (canonical) repository dat het “officiële” project vertegenwoordigt. Om bij te kunnen dragen tot dat project, maak je je eigen publieke kloon van het project en pusht je wijzigingen daarin terug. Daarna stuur je een verzoek naar de eigenaar van het hoofdproject om jouw wijzigingen binnen te halen. Hij kan je repository toevoegen als een remote, je wijzigingen lokaal testen, ze in zijn branch mergen, en dan naar zijn repository pushen. Het proces werkt als volgt (zie Integratie-manager workflow. ): De projecteigenaar pusht naar de publieke repository. Een bijdrager kloont die repository en maakt wijzigingen. De bijdrager pusht naar zijn eigen publieke kopie. De bijdrager stuurt de eigenaar een e-mail met de vraag om de wijzigingen binnen te halen. De eigenaar voegt de repo van de bijdrager toe als een remote en merget lokaal. De eigenaar pusht de gemergde wijzigingen terug in de hoofdrepository. Figuur 55. Integratie-manager workflow. Dit is een veel voorkomende workflow bij websites zoals GitHub of GitLab, waarbij het eenvoudig is om een project af te splitsen (fork) en je wijzigingen te pushen in jouw afgesplitste project waar iedereen ze kan zien. Een van de grote voordelen van deze aanpak is dat je door kunt gaan met werken, en de eigenaar van de hoofdrepository jouw wijzigingen op ieder moment kan binnenhalen. Bijdragers hoeven niet te wachten tot het project hun bijdragen invoegt — iedere partij kan in zijn eigen tempo werken. Dictator en luitenanten workflow Dit is een variant op de multi-repository workflow. Het wordt over het algemeen gebruikt bij enorme grote projecten met honderden bijdragers; een bekend voorbeeld is de Linux-kernel. Een aantal integrators hebben de leiding over bepaalde delen van de repository, zij worden luitenanten genoemd. Alle luitenanten hebben één integrator die bekend staat als de welwillende dictator (benevolent dictator). De repository van de welwillende dictator dient als het referentie-repository vanwaar alle bijdragers dienen te pullen. Het proces werkt als volgt (zie Benevolent dictator workflow. ): Reguliere ontwikkelaars werken op hun eigen onderwerp (topic) branch en rebasen hun werk op de master . De master -branch is die van de referentie-repository waar de dictator naar toe pusht. Luitenanten mergen de topic branches van de ontwikkelaars in hun master -branch. De dictator merged de master -branches van de luitenanten in de master -branch van de dictator. De dictator pusht zijn master -branch ten slotte terug naar het referentie-repository zodat de andere ontwikkelaars op deze kunnen rebasen. Figuur 56. Benevolent dictator workflow. Deze manier van werken is niet gewoon, maar kan handig zijn in hele grote projecten of in zeer hiërarchische omgevingen Het stelt de projectleider (de dictator) in staat het meeste werk te delegeren en grote subsets van code te verzamelen op meerdere punten alvorens ze te integreren. Workflows samenvatting Dit zijn een aantal veel voorkomende workflows die mogelijk zijn met een gedistribueerd systeem als Git, maar je kunt zien dat er veel variaties mogelijk zijn om ze te laten passen bij jouw specifieke workflow. Nu dat je (naar we hopen) in staat bent om te bepalen welke combinatie van workflows voor jou werkt, zullen we wat specifiekere voorbeelden behandelen hoe je de belangrijkste rollen kunt vervullen die in de verschillende workflows voorkomen. In de volgende paragraaf zal je kennis maken met een aantal reguliere patronen voor het bijdragen aan een project. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/posts/bmc-software_control-m-reviews-on-trustradius-activity-7416313849877774337-XRW9
Control-M Reviews on TrustRadius | BMC Software Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Skip to main content LinkedIn Top Content People Learning Jobs Games Sign in Join now BMC Software’s Post BMC Software 329,940 followers 1d Report this post Every review tells a story, and we’re so proud to be part of yours. 🧡 These TrustRadius reviews are a testament to the incredible work our customers are doing with Control-M. Thank you for trusting us to help you simplify, innovate, and succeed. Read more: https://bit.ly/4k6xWXQ 50 1 Comment Like Comment Share Copy LinkedIn Facebook X Jérémy T. 1d Report this comment Windows 11 vous force à vous connecter à Internet lors de l'installation ? Voici l'astuce pour contourner ça en 30 secondes. 🛑💻 On a tous connu ce moment frustrant lors d'une nouvelle installation de Windows (OOBE) : 1. L'écran de connexion réseau apparaît. 2. Le bouton "Je n'ai pas Internet" est introuvable. 3. Impossible d'avancer sans créer de compte Microsoft. Que ce soit pour installer un compte local, pour des questions de confidentialité, ou simplement parce que vos drivers Wi-Fi ne sont pas encore détectés, voici la commande magique à connaître. 👇 La Procédure : 1️⃣ Arrivé à l'écran de sélection du réseau, ne touchez à rien. 2️⃣ Ouvrez l'invite de commande avec le raccourci : 👉 MAJ + F10 (ou Fn + MAJ + F10 sur certains ordinateurs portables). 3️⃣ Tapez exactement cette commande et validez par Entrée : OOBE\BYPASSNRO 4️⃣ Le PC va redémarrer. ✅ Résultat : Au retour sur l'écran réseau, vous verrez apparaître l'option "Je n'ai pas Internet". Cliquez dessus, puis choisissez "Continuer avec l'installation limitée". Like Reply 1 Reaction To view or add a comment, sign in 329,940 followers View Profile Connect More from this author 2025 Highlights: Wins, Proof Points, and Momentum BMC Software 2w Control-M Ranks #1 for the 8th Consecutive Year. What EMA and Gartner® Say About Orchestration BMC Software 1mo BMC Named a Leader in the 2025 Gartner® Magic Quadrant™ for SOAP BMC Software 2mo Explore content categories Career Productivity Finance Soft Skills & Emotional Intelligence Project Management Education Technology Leadership Ecommerce User Experience Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language Sign in to view more content Create your free account or sign in to continue your search Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy .
2026-01-13T09:29:19
https://www.linkedin.com/products/hclsoftware-hcl-sametime/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_image-click
HCL Sametime | LinkedIn Skip to main content LinkedIn HCLSoftware in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in HCL Sametime Employee Communications Software by HCLSoftware See who's skilled in this Add as skill Learn more Report this product About HCL Sametime is a proven and trusted persistent chat and video meetings platform that complex industries and government entities use to ensure their most important communications are secure. It’s been rebuilt for how today’s modern organizations need to be working — nimbly, securely, and remotely — all at a significantly lower cost than the competition. This product is intended for Chief Technology Officer Technical Manager Senior System Analyst Head of Information Technology Department Information Security Engineer Director Network Engineering Chief Security Officer Information Technology Network Administrator Lead Information Technology Specialist DevOps Engineer Similar products Zoho Cliq Zoho Cliq Employee Communications Software Workvivo by Zoom Workvivo by Zoom Employee Communications Software Flip Flip Employee Communications Software Poppulo Poppulo Employee Communications Software Staffbase Staffbase Employee Communications Software Firstup Firstup Employee Communications Software Sign in to see more Show more Show less HCLSoftware products HCL BigFix HCL BigFix Endpoint Detection & Response (EDR) Software HCL Commerce+ HCL Commerce+ E-Commerce Platforms HCL Connections HCL Connections Enterprise Social Networking Software HCL Digital Experience HCL Digital Experience Digital Experience Platforms (DXP) HCL Domino HCL Domino Business Process Management Software HCL Mainframe Solutions HCL Mainframe Solutions Mainframe Management Software HCL Software DevOps HCL Software DevOps DevOps Software HCL Unica HCL Unica Marketing Automation Software HCL Volt MX HCL Volt MX Low-Code Development Platforms Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/pl/v2/Git-na-serwerze-Inne-opcje-hostowania-przez-podmioty-zewn%c4%99trzne
Git - Inne opcje hostowania przez podmioty zewnętrzne About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pierwsze kroki 1.1 Wprowadzenie do kontroli wersji 1.2 Krótka historia Git 1.3 Podstawy Git 1.4 Linia poleceń 1.5 Instalacja Git 1.6 Wstępna konfiguracja Git 1.7 Uzyskiwanie pomocy 1.8 Podsumowanie 2. Podstawy Gita 2.1 Pierwsze repozytorium Gita 2.2 Rejestrowanie zmian w repozytorium 2.3 Podgląd historii rewizji 2.4 Cofanie zmian 2.5 Praca ze zdalnym repozytorium 2.6 Tagowanie 2.7 Aliasy 2.8 Podsumowanie 3. Gałęzie Gita 3.1 Czym jest gałąź 3.2 Podstawy rozgałęziania i scalania 3.3 Zarządzanie gałęziami 3.4 Sposoby pracy z gałęziami 3.5 Gałęzie zdalne 3.6 Zmiana bazy 3.7 Podsumowanie 4. Git na serwerze 4.1 Protokoły 4.2 Uruchomienie Git na serwerze 4.3 Generowanie Twojego publicznego klucza SSH 4.4 Konfigurowanie serwera 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Inne opcje hostowania przez podmioty zewnętrzne 4.10 Podsumowanie 5. Rozproszony Git 5.1 Rozproszone przepływy pracy 5.2 Wgrywanie zmian do projektu 5.3 Utrzymywanie projektu 5.4 Podsumowanie 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Narzędzia Gita 7.1 Wskazywanie rewizji 7.2 Interaktywne używanie przechowali 7.3 Schowek i czyszczenie 7.4 Signing Your Work 7.5 Searching 7.6 Przepisywanie historii 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugowanie z Gitem 7.11 Moduły zależne 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Podsumowanie 8. Dostosowywanie Gita 8.1 Konfiguracja Gita 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git i inne systemy 9.1 Git jako klient 9.2 Migracja do Gita 9.3 Podsumowanie 10. Mechanizmy wewnętrzne w Git 10.1 Komendy typu plumbing i porcelain 10.2 Obiekty Gita 10.3 Referencje w Git 10.4 Spakowane pliki (packfiles) 10.5 Refspec 10.6 Protokoły transferu 10.7 Konserwacja i odzyskiwanie danych 10.8 Environment Variables 10.9 Podsumowanie A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.9 Git na serwerze - Inne opcje hostowania przez podmioty zewnętrzne Inne opcje hostowania przez podmioty zewnętrzne Jeśli nie chcesz przechodzić przez wszystkie prace związane z tworzeniem własnego serwera Gita, masz do wyboru kilka opcji hostingu swojego projektu na zewnętrznej stronie hostingowej. Sposób ten oferuje szereg zalet: strony hostingowe są zazwyczaj szybkie w konfiguracji i łatwe do uruchomienia projektu, nie masz własnego zaangażowania w monitorowanie i obsługę serwerów. Nawet jeśli założysz swój własny wewnętrzny serwer to nadal możesz korzystać w publicznej witryny, gdzie dużo łatwiej znaleźć pomoc. Na dzień dzisiejszy masz do wyboru bardzo dużo stron hostingowych. Każda z nich posiada swoje wady i zalety. Aby zobaczyć aktualną listę takich stron odwiedź adres https://git.wiki.kernel.org/index.php/GitHosting Omówimy szczegółowo korzystanie z GitHuba w rozdziale GitHub , ponieważ jest to największy hostingowy serwis Gita i możesz potrzebować interakcji z innymi projektami hostowanymi na nim. Niemniej jednak, istnieją dziesiątki innych hostingów, z których możesz skorzystać jeśli nie chcesz konfigurować własnego serwera Git. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/posts/bmc-software_from-our-houston-headquarters-to-teams-around-activity-7409578765472694273-Jzmv
From our Houston headquarters to teams around the world — wishing everyone a wonderful holiday season. | BMC Software Agree & Join LinkedIn By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . Skip to main content LinkedIn Top Content People Learning Jobs Games Sign in Join for free BMC Software’s Post BMC Software 329,940 followers 2w Report this post From our Houston headquarters to teams around the world — wishing everyone a wonderful holiday season. +4 418 1 Comment Like Comment Share Copy LinkedIn Facebook X Erez Gordon 2w Report this comment John McKenny love the Jacket💯 Like Reply 2 Reactions 3 Reactions To view or add a comment, sign in 329,940 followers View Profile Follow More from this author 2025 Highlights: Wins, Proof Points, and Momentum BMC Software 2w Control-M Ranks #1 for the 8th Consecutive Year. What EMA and Gartner® Say About Orchestration BMC Software 1mo BMC Named a Leader in the 2025 Gartner® Magic Quadrant™ for SOAP BMC Software 2mo Explore content categories Career Productivity Finance Soft Skills & Emotional Intelligence Project Management Education Technology Leadership Ecommerce User Experience Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional)) Language Sign in to view more content Create your free account or sign in to continue your search Sign in Welcome back Email or phone Password Show Forgot password? Sign in or By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . New to LinkedIn? Join now or New to LinkedIn? Join now By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy .
2026-01-13T09:29:19
https://git-scm.com/book/de/v2/Erste-Schritte-Git-installieren
Git - Git installieren About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Erste Schritte 1.1 Was ist Versionsverwaltung? 1.2 Kurzer Überblick über die Historie von Git 1.3 Was ist Git? 1.4 Die Kommandozeile 1.5 Git installieren 1.6 Git Basis-Konfiguration 1.7 Hilfe finden 1.8 Zusammenfassung 2. Git Grundlagen 2.1 Ein Git-Repository anlegen 2.2 Änderungen nachverfolgen und im Repository speichern 2.3 Anzeigen der Commit-Historie 2.4 Ungewollte Änderungen rückgängig machen 2.5 Mit Remotes arbeiten 2.6 Taggen 2.7 Git Aliases 2.8 Zusammenfassung 3. Git Branching 3.1 Branches auf einen Blick 3.2 Einfaches Branching und Merging 3.3 Branch-Management 3.4 Branching-Workflows 3.5 Remote-Branches 3.6 Rebasing 3.7 Zusammenfassung 4. Git auf dem Server 4.1 Die Protokolle 4.2 Git auf einem Server einrichten 4.3 Erstellung eines SSH-Public-Keys 4.4 Einrichten des Servers 4.5 Git-Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Von Drittanbietern gehostete Optionen 4.10 Zusammenfassung 5. Verteiltes Git 5.1 Verteilter Arbeitsablauf 5.2 An einem Projekt mitwirken 5.3 Ein Projekt verwalten 5.4 Zusammenfassung 6. GitHub 6.1 Einrichten und Konfigurieren eines Kontos 6.2 Mitwirken an einem Projekt 6.3 Ein Projekt betreuen 6.4 Verwalten einer Organisation 6.5 Skripte mit GitHub 6.6 Zusammenfassung 7. Git Tools 7.1 Revisions-Auswahl 7.2 Interaktives Stagen 7.3 Stashen und Bereinigen 7.4 Deine Arbeit signieren 7.5 Suchen 7.6 Den Verlauf umschreiben 7.7 Reset entzaubert 7.8 Fortgeschrittenes Merging 7.9 Rerere 7.10 Debuggen mit Git 7.11 Submodule 7.12 Bundling 7.13 Replace (Ersetzen) 7.14 Anmeldeinformationen speichern 7.15 Zusammenfassung 8. Git einrichten 8.1 Git Konfiguration 8.2 Git-Attribute 8.3 Git Hooks 8.4 Beispiel für Git-forcierte Regeln 8.5 Zusammenfassung 9. Git und andere VCS-Systeme 9.1 Git als Client 9.2 Migration zu Git 9.3 Zusammenfassung 10. Git Interna 10.1 Basisbefehle und Standardbefehle (Plumbing and Porcelain) 10.2 Git Objekte 10.3 Git Referenzen 10.4 Packdateien (engl. Packfiles) 10.5 Die Referenzspezifikation (engl. Refspec) 10.6 Transfer Protokolle 10.7 Wartung und Datenwiederherstellung 10.8 Umgebungsvariablen 10.9 Zusammenfassung A1. Anhang A: Git in anderen Umgebungen A1.1 Grafische Schnittstellen A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Zusammenfassung A2. Anhang B: Git in deine Anwendungen einbetten A2.1 Die Git-Kommandozeile A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Anhang C: Git Kommandos A3.1 Setup und Konfiguration A3.2 Projekte importieren und erstellen A3.3 Einfache Snapshot-Funktionen A3.4 Branching und Merging A3.5 Projekte gemeinsam nutzen und aktualisieren A3.6 Kontrollieren und Vergleichen A3.7 Debugging A3.8 Patchen bzw. Fehlerkorrektur A3.9 E-mails A3.10 Externe Systeme A3.11 Administration A3.12 Basisbefehle 2nd Edition 1.5 Erste Schritte - Git installieren Git installieren Bevor du mit Git loslegen kannst, muss es natürlich zuerst installiert werden. Auch wenn es bereits vorhanden ist, ist es vermutlich eine gute Idee, auf die neueste Version zu aktualisieren. Du kannst es entweder als Paket oder über ein anderes Installationsprogramm installieren oder den Quellcode herunterladen und selbst kompilieren. Anmerkung Dieses Buch wurde auf Basis der Git-Version 2 geschrieben. Da Git hervorragend darin ist, die Abwärtskompatibilität aufrechtzuerhalten, sollte jede neuere Version problemlos funktionieren. Auch wenn die meisten Befehle, die wir anwenden werden, auch in älteren Versionen funktionieren, kann es doch sein, dass die Befehlsausgabe oder das Verhalten leicht anders ist. Installation unter Linux Wenn du die grundlegenden Git-Tools unter Linux über ein Installationsprogramm installieren möchten, kannst du das in der Regel über das Paketverwaltungstool der Distribution tun. Wenn du mit Fedora (oder einer anderen eng damit verbundenen RPM-basierten Distribution, wie RHEL oder CentOS) arbeitest, kannst du dnf verwenden: $ sudo dnf install git-all Auf einem Debian-basierten System, wie Ubuntu, steht apt zur Verfügung: $ sudo apt install git-all Auf der Git-Homepage https://git-scm.com/download/linux findet man weitere Möglichkeiten und Optionen, wie man Git unter einem Unix-basierten Betriebssystem installieren kann. Installation unter macOS Es gibt mehrere Möglichkeiten, Git auf einem Mac zu installieren. Am einfachsten ist es wahrscheinlich, die Xcode Command Line Tools zu installieren. Bei Mavericks (10.9) oder neueren Versionen kann man dazu einfach git im Terminal eingeben. $ git --version Wenn Git noch nicht installiert ist, erscheint eine Abfrage, ob man es installieren möchte. Wenn man eine sehr aktuelle Version einsetzen möchte, kann man Git auch über ein Installationsprogramm installieren. Auf der Git-Website https://git-scm.com/download/mac findet man die jeweils aktuellste Version und kann sie von dort herunterladen. Abbildung 7. Git macOS Installationsprogramm Installation unter Windows Auch für Windows gibt es einige Möglichkeiten zur Installation von Git. Eine offizielle Windows-Version findet man direkt auf der Git-Homepage. Gehe dazu auf https://git-scm.com/download/win und der Download sollte dann automatisch starten. Man sollte dabei beachten, dass es sich hierbei um das Projekt „Git for Windows“ handelt, welches unabhängig von Git selbst ist. Weitere Informationen hierzu findest du unter https://msysgit.github.io/ . Um eine automatisierte Installation zu erhalten, kannst du das Git Chocolatey Paket verwenden. Beachte, dass das Chocolatey-Paket von der Community gepflegt wird. Aus dem Quellcode installieren Viele Leute kompilieren Git auch auf ihrem eigenen Rechner, weil sie damit die jeweils aktuellste Version erhalten. Die vorbereiteten Pakete hinken meist ein wenig hinterher, obwohl Git in den letzten Jahren ausgereifter geworden ist und dies somit wesentlich besser geworden ist. Wenn du Git aus dem Quellcode installieren möchtest, benötigst du die folgenden Bibliotheken, von denen Git abhängt: autotools, curl, zlib, openssl, expat und libiconv. Wenn du dich beispielsweise auf einem System befinden, das Paketverwaltungen, wie dnf (Fedora) oder apt-get (ein Debian-basiertes System) hat, kannst du mit einem dieser Befehle die minimalen Abhängigkeiten für die Kompilierung und Installation der Git-Binärdateien installieren: $ sudo dnf install dh-autoreconf curl-devel expat-devel gettext-devel \ openssl-devel perl-devel zlib-devel $ sudo apt-get install dh-autoreconf libcurl4-gnutls-dev libexpat1-dev \ gettext libz-dev libssl-dev Um die Dokumentation in verschiedenen Formaten (doc, html, info) zu erstellen, sind weitere Abhängigkeiten notwendig: $ sudo dnf install asciidoc xmlto docbook2X $ sudo apt-get install asciidoc xmlto docbook2x Anmerkung Benutzer von RHEL und RHEL-Derivaten wie CentOS und Scientific Linux müssen das EPEL-Repository aktivieren , um das Paket docbook2X herunterzuladen. Wenn du eine Debian-basierte Distribution (Debian, Ubuntu oder deren Derivate) verwendest, benötigst du auch das Paket install-info : $ sudo apt-get install install-info Wenn du eine RPM-basierte Distribution (Fedora, RHEL oder deren Derivate) verwendest, benötigst du auch das Paket getopt (welches auf einer Debian-basierten Distribution bereits installiert ist): $ sudo dnf install getopt Wenn du Fedora- oder RHEL-Derivate verwendest, musst du wegen der unterschiedlichen Paketnamen zusätzlich einen Symlink erstellen, indem du folgenden Befehl: $ sudo ln -s /usr/bin/db2x_docbook2texi /usr/bin/docbook2x-texi aufgrund von binären Namensunterschieden ausführst. Wenn du alle notwendigen Abhängigkeiten installiert hast, kannst du dir als nächstes die jeweils aktuellste Version als Tarball von verschiedenen Stellen herunterladen. Man findet die Quellen auf der Kernel.org-Website unter https://www.kernel.org/pub/software/scm/git , oder einen Mirror auf der GitHub-Website unter https://github.com/git/git/releases . Auf der GitHub-Seite ist es einfacher herauszufinden, welches die jeweils aktuellste Version ist. Auf kernel.org dagegen werden auch Signaturen zur Verifikation des Downloads der jeweiligen Pakete zur Verfügung gestellt. Nachdem man sich so die Quellen beschafft hat, kann man Git kompilieren und installieren: $ tar -zxf git-2.8.0.tar.gz $ cd git-2.8.0 $ make configure $ ./configure --prefix=/usr $ make all doc info $ sudo make install install-doc install-html install-info Nachdem dies erledigt ist, kannst du Git für Updates auch über Git selbst beziehen: $ git clone https://git.kernel.org/pub/scm/git/git.git prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/categories/managed-file-transfer-software?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_subtitle-click
Best Managed File Transfer (MFT) Software | Products | LinkedIn Skip to main content LinkedIn Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Clear text Used by Used by Information Technology Manager (8) System Engineer (3) Technical Architect (3) Compliance Officer (3) Infrastructure Manager (2) See all products Find top products in Managed File Transfer (MFT) Software category Software used to move large volumes of files over established networks with end-to-end security and activity monitoring. - Use built-in security, encryption, and compliance features not included in FTP solutions - Scale delivery between systems, automate file transfer procedures, and support multiple protocols - Control auditing and authentication and receive confirmation of successful transfers 84 results Axway Managed File Transfer Managed File Transfer (MFT) Software by Axway Secure, reliable, and easy-to-manage solution for transferring data between people, partners, businesses, and applications. Manage and control how your organization sends and receives data so you can ensure it stays protected and meets compliance regulations such as HIPAA, PCI, DSS, and GDPR. View product Progress MOVEit Managed File Transfer (MFT) Software by Progress Software Enterprise-grade secure file transfer and automation software for reliable, secure data transfer between partners, customers and systems. View product Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software by SolarWinds Serv-U® Managed File Transfer Server offers comprehensive security, automation, and control over file transfers in and outside your organization. View product JSCAPE by Redwood Managed File Transfer (MFT) Software by Redwood Software Automate your file transfers while staying secure and compliant with JSCAPE by Redwood, able to support any platform, protocol or environment. Easy to use and set up, IT teams are empowered with a single view into all file transfer operations, with real-time notifications and alerts. With a built-in library of 100+ no-code/low-code workflow automation templates, REST APIs and off-the-shelf integrations, your experts have the tools they need to deliver reliable file transfer operations. Experience the power of scalable file transfer with JSCAPE. Experience the power of secure, seamless and scalable file transfer with JSCAPE. Elevate your enterprise's data management capabilities and unlock new levels of productivity and agility today. View product Cerberus by Redwood Managed File Transfer (MFT) Software by Redwood Software Secure file transfer just got easier with Cerberus FTP by Redwood. Since 2001, Cerberus has been simplifying SFTP for small, midsize and enterprise organizations. Known for its ease of use, robust security features and award-winning support, Cerberus is the preferred Windows SFTP solution for IT professionals. Security and compliance features help you meet requirements like HIPAA, GPDR and more, plus regular penetration testing and third-party auditing identifies any potential product weaknesses for remediation. Start sharing your files securely and reliably with Cerberus FTP today. View product Find products trusted by professionals in your network See which products are used by connections in your network and those that share similar job titles Sign in to view full insights MLADU Managed File Transfer (MFT) Software by Dantech Corporation Inc. MLADU redefines data transfer with industry-leading scalability, security, availability, and performance. Designed for businesses of all sizes and across all industries, MLADU effortlessly handles data transfers of any volume, supporting use cases such as acquisitions, clinical trials, omics pipelines, scientific research, enterprise applications, and big data analytics. Its robust features empower you to easily initiate, track, organize, and configure transfers to align perfectly with your business, organizational, and compliance needs. Engineered to overcome the toughest data transfer challenges, MLADU addresses complexities like data corruption, system incompatibility, duplication, and massive data volumes. It also ensures the highest standards of security, regulatory compliance, and data integrity. Whether you're navigating diverse platforms, managing varying formats, or mitigating potential human errors, MLADU is your trusted partner in seamless, reliable data transfer solutions View product dDataBox Managed File Transfer (MFT) Software by Dataport AöR Bei der Kommunikation mit unseren Kolleg*innen und Externen greifen wir täglich auf E-Mails zurück. Für den Versand sehr großer Dateien eignen sie sich allerdings nicht. Zudem bieten E-Mails nicht immer die erforderliche Vertraulichkeit. Mit der dDataBox können Sie Ihren Adressat*innen nun große Dateien sicher bereitstellen. Auf einen Blick: ✅ Ende-zu-Ende Verschlüsselung ✅ Smartphone-App (Android, iOS) ✅ Temporärer Down- und Uploadlink für Externe ✅ Benutzer*innen und Speichervolumen eigenständig verwalten ✅ Betrieb im hochsicheren Twin Data Center ✅ Günstiger Speicher View product LinShare Managed File Transfer (MFT) Software by LINAGORA LinShare is a Private and secure file sharing and cloud storage solution. Fully compliant with GDPR. Made and hosted in the EU. LinShare has been free to use and powered by open-source since 2009. Check us out on GitHub at https://github.com/linagora/linshare LinShare gives you access to features such as notifications whenever your file is opened or downloaded, activity logs and secure uploads. With LinShare you can organize your files conveniently and use them for efficient collaboration with your team. Your privacy is our priority so LinShare uses data encryption, password-safe links and multiple layers of protection. View product AnyShare – Rapid File Transfer Managed File Transfer (MFT) Software by Technozer Solution No. 1 App to transfer APKs, videos and photos to any devices A simple and easy use app to manage files on your android, iOS or PC devices. Avoid using USB cables to transfer files, movies, photos and songs from your device to laptops/PCs/Macs and other android devices without internet. AnyShare sets itself apart from other share apps in the market by offering perks over perks. So, what are you waiting for? Download AnyShare – Rapid File Transfer, Share Files Now. View product Secure File Sharing Managed File Transfer (MFT) Software by Tresorit | Swiss Post Digital Share files securely. Stay in control of your data. Protect your company’s data with zero-knowledge end-to-end encryption when sharing files inside or outside your organization. Keep your files in sight long after they’ve been shared. Data transfers are crucial to running a business in the digital age but this has led to new challenges. The exponential increase in data shared online, an emerging fast-paced working culture, and limitless remote collaboration pose significant risks to companies. Your most valuable data is constantly being exposed to a myriad of security and privacy threats. - Exchange files securely with clients, partners, and other teams using encrypted links. - Ultra-secure share links enable encrypted document and folder sharing with anyone, even if they don’t have a Tresorit account. - Collect files from anyone, anywhere with secure, encrypted file requests. Fully compliant with: - GDPR, CCPA - HIPAA - TISAX(Chapter 5) - ITAR - FINRA View product See more How it works Explore Discover the best product for your need from a growing catalog of 25,000 products and categories trusted by LinkedIn professionals Learn Evaluate new tools, explore trending products in your industry and see who in your network is skilled in the product Grow Join communities of product users to learn best practices, celebrate your progress and accelerate your career LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English Language
2026-01-13T09:29:19
https://www.linkedin.com/products/ovhgroup-ovh-antiddos-protection/?trk=products_seo_search
OVHcloud Anti-DDoS Protection | LinkedIn Skip to main content LinkedIn OVHcloud in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in OVHcloud Anti-DDoS Protection DDoS Protection Software by OVHcloud See who's skilled in this Add as skill Learn more Report this product About Mitigate your risk of DDoS attacks and keep your services available at all times. Free with every OVH service. Learn more. Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less OVHcloud products OVHcloud Block Storage OVHcloud Block Storage Block Storage Software OVHcloud Marketplace OVHcloud Marketplace Marketplace Platforms OVHcloud Object Storage OVHcloud Object Storage Object Storage Software OVHcloud Zerto DRP OVHcloud Zerto DRP Disaster Recovery as a Service (DRaaS) Software LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://hackercombat.com
Hacker Combat - Cyber Security and Hacking News | HackerCombat Tuesday 13 January 2026 / 9:29 am Submit News Tips Forum FREE EDR Toggle navigation News Technology Malware Hacks Ransomware Partner Write For Us FREE EDR Threats What Is Advanced Threat Protection? September 19, 2019 by Julia Sowells 0 Cyber Attacks Advanced Persistent Threat: What You Need to Know August 26, 2019 by Kevin Jones 0 Today's Featured Security News Tech Google Patches Critical Android Flaw August 2, 2022 by Paul Hamilton‏ 0 Snowflake Data Breach: What Happened and How to Prevent It August 5, 2025 Ways to Mitigate Risk in Cybersecurity: Cybersecurity Risk Management December 13, 2024 Zero Trust Architecture  December 2, 2024 Malware Leaks North Korean hackers target crypto experts with fake Coinbase job offers August 8, 2022 by Paul Hamilton‏ 0 Security News cybersecurity Vulnerability Microsoft Researches Ransomware Attack Targeting App Developers September 20, 2021 by Paul Hamilton‏ 0 Latest Security News Snowflake Data Breach: What Happened and How to Prevent It August 5, 2025 by Hacker Combat 0 0 In 2024, the cybersecurity landscape was shaken by an unexpected and widespread incident—the Snowflake data breach. Despite being a leading provider of cloud-based data warehousing solutions, Snowflake found itself at… Read More Ways to Mitigate Risk in Cybersecurity: Cybersecurity Risk Management December 13, 2024 by Hacker Combat 0 0 Cyber threats can wreak havoc on businesses, from data breaches to loss of reputation. Luckily, there are effective strategies available that can reduce cybersecurity risk. Avoidance is one of the… Read More Zero Trust Architecture  December 2, 2024 by Hacker Combat 0 0 Zero trust security takes an “never trust, always verify” approach to access control. Access is only granted once an individual’s identity and context have been confirmed through multifactor authentication and… Read More What Is a Security Operations Center (SOC)? December 2, 2024 by Hacker Combat 0 0 A Security Operations Center (SOC) specializes in monitoring and analyzing data to detect cyber threats and prevent attacks from them. They work to sort actual threats from false positives before… Read More XDR vs SIEM Security Information and Event Management November 29, 2024 by Hacker Combat 0 0 The Extended Detection and Response Platform (XDR) ingestion and correlation technology captures and correlates high-fidelity data across your security layers, such as endpoint, network, logs, cloud services and identities to… Read More Best Free EDR for Windows PC November 29, 2024 by Hacker Combat 0 0 Endpoint detection and response (EDR) tools offer businesses that employ hybrid work models or remote employees an extra layer of cybersecurity protection. Utilizing artificial intelligence (AI) and machine learning (ML),… Read More Free EDR Solutions for Home Users in 2025 November 26, 2024 by Hacker Combat 0 0 EDR can detect and respond to emerging and advanced cyber threats quickly and efficiently, making it an essential component of modern business ecosystems. Beyond signature-based detection capabilities, its features go… Read More Cloud Security Essentials October 28, 2024 by Hacker Combat 0 0 Cloud security involves employing perimeter defenses like firewalls, IDPSs and VPNs as well as guaranteeing isolation through network segmentation and virtual LANs while monitoring traffic for anomalies and threats –… Read More Antivirus Software October 28, 2024 by Hacker Combat 0 0 Antivirus software protects devices against viruses, malware, and other cyberthreats by detecting, quarantining, and deleting malicious code. Modern antivirus products also offer additional security features such as password protection, identity… Read More How to Protect Against Ransomware Attacks? October 25, 2024 by Hacker Combat 0 0 Criminal hackers employ ransomware attacks against their targets by encrypting their data and demanding that a ransom be paid within an allotted timeframe or risk losing it forever. When an… Read More Show me More Stay Connected 100K+ Followers 15k+ Fans 6899 Followers 2067 Followers Recent Posts Snowflake Data Breach: What Happened and How to Prevent It Ways to Mitigate Risk in Cybersecurity: Cybersecurity Risk Management Zero Trust Architecture  What Is a Security Operations Center (SOC)? XDR vs SIEM Security Information and Event Management Best Free EDR for Windows PC Free EDR Solutions for Home Users in 2025 About Hackercombat is a news site, which acts as a source of information for IT security professionals across the world. We have lived it for 2 years, sharing IT expert guidance and insight, in-depth analysis, and news. We also educate people with product reviews in various content forms. As a dedicated cybersecurity news platform, HC has been catering unbiased information to security professionals, on the countless security challenges that they come across every day. We publish data on comprehensive analysis, updates on cutting-edge technologies and features with contributions from thought leaders. Hackercombat also has a section extensively for product reviews and forums. Categories Android Security Antivirus Cloud Security Cryptocurrency Cyber Attacks Cyber Crime cybersecurity Data breach Data Protection Data security DDoS Attacks Endpoint Security Firewall GDPR Hacks Infographics Internet & Network Internet & Network IoT Leaks Malware Mobile Security Network Security Phishing Attacks Ransomware Security & Privacy Security News Tech Threats VPN Vulnerability Web Security Website Security Quick Links News Cyber Attacks Hacking Data Breach Security Data Security Network Security Web Security Submit News Tips Technology Vulnerability Website Malware Scanner Fix Hacked Website Submit News Tips Stay Connected Hacker Combat LLC 2025. Under the Section 18-201 of the Delaware Limited Liability Company Act. Login Welcome! Login in to your account Username or email address * Password * Login Remember me Lost your password? Don't have account. Register Lost Password Type your email address * Submit Back to Login Register Your Username * Your Email Adress * Your Password * Confirm Password * Back to Login
2026-01-13T09:29:19
https://visualstudio.microsoft.com/cs/vs/
Visual Studio IDE – AI pro psaní, ladění a testování kódu Přeskočit na hlavní obsah Microsoft Visual Studio Visual Studio Visual Studio Domů Vývojářské nástroje Visual Studio Visual Studio Přehled Funkce Edice Rozšíření Začínáme Stav služby Podpora nejčastějších dotazů Dokumentace Blog Visual Studio Code Visual Studio Code Přehled Rozšíření Podpora Dokumentace Blog VS Code for Education Soubory ke stažení Koupit Předplatné Přehled Podpora​ Portál správce​ Portál pro předplatitele​ Dev Essentials Centrum sady Visual Studio Více Visual Studio zdarma Vše od Microsoftu Global Microsoft Security Azure Dynamics 365 Microsoft 365 Microsoft Teams Windows 365 Technologie a inovace Technologie a inovace Umělá inteligence Azure Space Hybridní realita Microsoft HoloLens Microsoft Viva Kvantové výpočetní prostředky Udržitelnost Odvětví Odvětví Vzdělávání Automobilový průmysl Finanční služby Státní správa Zdravotní péče Výroba Maloobchod Všechna průmyslová odvětví Partneři Partneři Najít partnera Staňte se partnerem Partnerská síť Microsoft Marketplace Marketplace Rewards Software development companies Zdroje Zdroje Blog Microsoft Advertising Středisko pro vývojáře Dokumentace Události Licencování Microsoft Learn Microsoft Research Zobrazit mapu stránek Přihlásit se This device is not currently supported for these products. Visual Studio IDE Visual Studio for Mac Visual Studio Code To continue downloading, click here Visual Studio IDE – AI pro psaní, ladění a testování kódu 2025-12-08T11:35:54-08:00 Výkonné integrované vývojové prostředí (IDE) – inteligentnější s využitím AI Režim agenta – nyní k dispozici v GitHub Copilot Free Vyzkoušejte si pomoc při psaní kódů od umělé inteligence, která dokáže zdůvodnit problémy, koordinovat další kroky, aplikovat změny a iterovat při chybách. Stáhnout Visual Studio Community Professional Enterprise Insiders Začněte zadarmo Zvýšení produktivity s asistencí AI Skutečná řešení AI s režimem agenta Plánujte, sestavujte, testujte a opravujte – vše z jediné výzvy. Popište svou potřebu v přirozeném jazyce a nechte Copilota, ať udělá zbytek. Spouštějte lintry, testy a příkazy, aniž byste opustili Visual Studio. Aktivujte režim agenta hned Vždy aktuální modely AI Získejte přístup k nejnovějším modelům AI a získejte nejpřesnější kontextové návrhy v chatu nebo inline. Vyberte model, který odpovídá potřebě každého úkolu. Získejte nejnovější modely Bezplatné a plně integrované v sadě Visual Studio Pomoc s psaním kódu s využitím AI integrovaná do Visual Studio. Získejte 2 000 dokončení kódu. Získejte 50 chatových zpráv zdarma za měsíc. Povolit Copilot Free Sledujte GitHub Copilot v akci Už nikdy nebudete muset psát zprávu o potvrzení: Díky GitHub Copilote Úpravy více souborů pro GitHub Copilot ve Visual Studiu Začínáme s GitHub Copilot Free ve Visual Studiu Zachovejte si přehled o dění kolem Visual Studia Navštivte Visual Studio Hub Vývoj Ladění Testování Správa verzí Spolupráce Cloud Návrh Profilování Data Copilot: váš kolega programátor Urychlete vývoj pomocí kontextových návrhů pro dokončování kódu pro jeden řádek nebo celý fragment kódu. Pomocí Copilot Edits můžete upravovat více souborů s prostředím pro revize kódu, náhledy v souboru a vracení změn zpět. Jak používat Copilot pro návrhy kódu Chat pro pomoc s kódem a dotazy Získejte nápovědu k psaní dotazů týkajících se psaní kódu v reálném čase pomocí chatu GitHub Copilot. Konverzační rozhraní chatu poskytuje okamžité návrhy s podporou kontextu, pomáhá s laděním a generuje testy jednotek, a to vše ve vašem vývojovém prostředí. Tipy pro vylepšení výsledků chatu Copilot Snadné ladění s využitím služby Copilot Zrychlete ladění s podrobnými přehledy o vašem kódu s využitím AI. Copilot nabízí hloubkovou analýzu a vysvětlení a poskytuje cílená řešení pro výjimky, zablokování a další scénáře ladění. Zjednodušení procesu ladění Zvýšení efektivity testování prostřednictvím Copilotu Zkraťte dobu strávenou ručním vytvářením testů a udržováním kvality kódu. Automatizace s využitím služby Copilot pomáhá zvýšit spolehlivost včasnou identifikací chyb pomocí testů jednotek a integrace na základě vašeho stávajícího kódu. Oprava chyb pomocí služby Copilot Zjednodušení správy verzí s využitím Copilotu Udržujte čistý a uspořádaný základ kódu pomocí zpráv o potvrzení generovaných AI na základě změn kódu. GitHub Copilot zjednodušuje správu verzí tím, že zajišťuje srozumitelnost pracovního postupu, správy větví a konfliktů při slučování. Generování zpráv o potvrzení pomocí chatu GitHub Copilot Sdílejte více než obrazovky Relace spolupráce v reálném čase Live Share urychlují cykly úprav a ladění vašeho týmu bez ohledu na jazyk nebo platformu. Přizpůsobené relace s řízením přístupu a vlastním nastavením editoru zajistí, aby kód všech zůstal konzistentní. Spolupracujte s Live Share Azure v sadě Visual Studio Vytvářejte, nasazujte a škálujte s využitím Azure ve Visual Studiu: Efektivní a bezpečné vytváření cloudových aplikací s využitím integrovaných nástrojů a možností AI – to vše v rámci integrovaného vývojového prostředí (IDE). Bez ohledu na to, jestli vyvíjíte webové aplikace, rozhraní API nebo řešení založená na AI, Visual Studio nabízí bezproblémovou integraci Azure pro jednodušší kódování. Prozkoumejte kurzy, zvyšte svou úroveň psaní kódu a automatizujte nasazení pomocí Azure DevOps a GitHub Actions. Připojte se ke službám Azure, nasazujte aplikace a laďte v reálném čase pomocí výkonných nástrojů ve Visual Studiu. Další informace o Azure ve Visual Studiu Navrhujte krásné multiplatformní aplikace Visual Studio je nejlepší integrované vývojové prostředí (IDE) pro vytváření moderních multiplatformových aplikací pro Windows, Mac, Linux, iOS a Android. Vytvářejte bohaté klientské aplikace s WinForms, WPF, WinUI a MAUI pomocí výkonných vizuálních návrhářů, které zjednodušují komplexní rozvržení a oživují vaše uživatelské rozhraní. Vizualizace a modelování Sestavujte rychlé a responzivní aplikace pomocí integrovaných nástrojů pro profilaci Pomocí nástrojů pro profilaci můžete ve Visual Studiu vytvářet rychlé a responzivní aplikace .NET a C++ pro web a klienta. Díky profilerům pro procesor, paměť, XAML, databáze, události a další funkce máte celou sadu nástrojů, které vám pomůžou identifikovat a řešit problémy s výkonem. Měřte výkon aplikací Nástroje pro usnadnění práce s SQL Pracujte s projekty SQL bez opuštění integrovaného vývojového prostředí (IDE). Vytvářejte dotazy s bohatými funkcemi pro úpravy, zobrazujte a zkoumejte servery a databáze. Vizualizujte schémata databáze a dotazy z programu Visual Studio. Přístup k datům ve Visual Studiu Nastavte si to po svém Nakonfigurujte a využívejte motivy, jak byste očekávali od jakéhokoli moderního vývojového nástroje. Ať už zvolíte jeden z výchozích motivů, který je už nainstalovaný, nebo prozkoumáte nové motivy z Visual Studio Marketplace, budete moct svůj editor nebo integrované vývojové prostředí (IDE) nastavit jako skutečně své. Prozkoumat motivy S rozšířeními to můžete dotáhnout dál. Rozšíření jsou doplňky, které přizpůsobují a vylepšují Visual Studio, včetně dalších nastavení, funkcí nebo použití pro existující nástroje. S tisíci rozšířeními na Marketplace máte spoustu možností, jak zvýšit produktivitu a vyhovět vašim pracovním postupům. Prozkoumat rozšíření GitHub Actions Instalace Okno, které poskytuje zobrazení GitHub Actions pro aktuální úložiště otevřeného řešení ve Visual Studiu. Poskytuje @timheuer Získat rozšíření Markdown Editor v2 Instalace Plnohodnotný editor Markdownu s živým náhledem a zvýrazněním syntaxe. Podporuje Markdown ve formátu používaném v GitHubu Získat rozšíření Balíček motivů sady Visual Studio Instalace Kolekce oblíbených motivů, která je teď dostupná pro Visual Studio. Další informace Sada nástrojů SQLite a SQL Server Compact Instalace Toto rozšíření přidává několik funkcí, které vám pomůžou s vývojem vložených databází. Získat rozšíření Visual Studio Community Bezplatné IDE se všemi funkcemi pro studenty, open source a samostatné vývojáře Další informace Stažení zdarma Visual Studio Professional Výkonné vývojářské nástroje, služby a výhody předplatného pro efektivní spolupráci malých týmů Další informace Bezplatná zkušební verze Visual Studio Enterprise Pokročilé ladění, podnikové zabezpečení, vývoj s AI asistencí a navíc kredity Azure a ceny pro vývoj/testování pro cloud Další informace Bezplatná zkušební verze Sledujte nás Dev Essentials Developer Community Blogy: Visual Studio .NET C++ Co je nového Surface Pro Surface Laptop Surface Laptop Studio 2 Copilot pro organizace Copilot pro osobní používání Microsoft 365 Prohlédnout produkty Microsoft Microsoft Store Profil účtu Centrum stahování Podpora pro Microsoft Store Vrácení Sledování objednávky Recyklace Commercial Warranties Vzdělávání Microsoft Education Zařízení pro vzdělávání Microsoft Teams pro vzdělávání Microsoft 365 Education Office Education Vzdělávání a rozvoj pedagogů Akce pro studenty a rodiče Azure pro studenty Firmy Microsoft Security Azure Dynamics 365 Microsoft 365 Microsoft Advertising Microsoft 365 Copilot Microsoft Teams Vývojáři a IT Microsoft pro vývojáře Microsoft Learn Podpora aplikací s umělou inteligenci v obchodě Marketplace Odborná komunita Microsoft Microsoft Marketplace Microsoft Power Platform Marketplace Rewards Visual Studio Společnost Kariéra Novinky u společnosti Microsoft a ochrana osobních údajů Investoři Udržitelnost Čeština (Česko) Ikona nesouhlasu s volbami ochrany osobních údajů Vaše volby ochrany osobních údajů Ikona nesouhlasu s volbami ochrany osobních údajů Vaše volby ochrany osobních údajů Ochrana osobních údajů spotřebitele ve zdravotnictví Kontaktujte nás Ochrana osobních údajů a soubory cookie Spravovat soubory cookie Podmínky použití Ochranné známky O našich reklamách © Microsoft 2026
2026-01-13T09:29:19
https://git-scm.com/book/gr/v2/%ce%a4%ce%b1-%ce%b8%ce%b5%ce%bc%ce%b5%ce%bb%ce%b9%cf%8e%ce%b4%ce%b7-%cf%83%cf%84%ce%bf%ce%b9%cf%87%ce%b5%ce%af%ce%b1-%cf%84%ce%bf%cf%85-Git-%ce%91%cf%80%cf%8c%ce%ba%cf%84%ce%b7%cf%83%ce%b7-%ce%b1%cf%80%ce%bf%ce%b8%ce%b5%cf%84%ce%b7%cf%81%ce%af%ce%bf%cf%85-Git
Git - Απόκτηση αποθετηρίου Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Ξεκινώντας με το Git 1.1 Σχετικά με τον έλεγχο εκδόσεων 1.2 Σύντομο ιστορικό του Git 1.3 Τι είναι το Git; 1.4 Η γραμμή εντολών 1.5 Εγκατάσταση του Git 1.6 Ρύθμιση του Git για πρώτη φορά 1.7 Χρησιμοποιώντας τη βοήθεια 1.8 Ανακεφαλαίωση 2. Τα θεμελιώδη στοιχεία του Git 2.1 Απόκτηση αποθετηρίου Git 2.2 Καταγραφή αλλαγών στο αποθετήριο 2.3 Χρησιμοποιώντας το ιστορικό υποβολών 2.4 Αναιρέσεις (undoing) 2.5 Δουλεύοντας με απομακρυσμένα αποθετήρια 2.6 Ετικέτες 2.7 Συντομεύεσεις στο Git 2.8 Ανακεφαλαίωση 3. Διακλαδώσεις στο Git 3.1 Οι κλάδοι με λίγα λόγια 3.2 Βασικές έννοιες διακλαδώσεων και συγχωνεύσεων 3.3 Διαχείριση κλάδων 3.4 Ροές εργασίας με διακλαδώσεις 3.5 Απομακρυσμένοι κλάδοι 3.6 Αλλαγή βάσης (rebasing) 3.7 Ανακεφαλαίωση 4. Το Git στον διακομιστή 4.1 Τα πρωτόκολλα 4.2 Εγκατάσταση του Git σε διακομιστή 4.3 Δημιουργία δημόσιου κλειδιού SSH 4.4 Στήσιμο του διακομιστή 4.5 Δαίμονες του Git 4.6 Έξυπνο HTTP 4.7 GitWeb 4.8 GitLab 4.9 Επιλογές φιλοξενίας από τρίτους 4.10 Ανακεφαλαίωση 5. Κατανεμημένο Git 5.1 Κατανεμημένες ροές εργασίας 5.2 Συνεισφέροντας σε ένα έργο 5.3 Συντήρηση ενός έργου 5.4 Ανακεφαλαίωση 6. GitHub 6.1 Δημιουργία λογαριασμού και ρύθμισή του 6.2 Συνεισφορά σε έργο 6.3 Συντήρηση ενός έργου 6.4 Διαχείριση οργανισμώνν 6.5 Συγγραφή script στο GitHub 6.6 Ανακεφαλαίωση 7. Εργαλεία του Git 7.1 Επιλογή αναθεώρησης 7.2 Διαδραστική εργασία με το στάδιο καταχώρισης 7.3 Αποθέματα και Καθαρισμός 7.4 Υπογραφή της δουλειάς μας 7.5 Αναζήτηση 7.6 Η ιστορία ξαναγράφεται 7.7 Απομυθοποίηση της reset 7.8 Προχωρημένη Συγχώνευση 7.9 Rerere 7.10 Αποσφαλμάτωση με το Git 7.11 Υπομονάδες 7.12 Δεμάτιασμα δεδομένων 7.13 Replace 7.14 Αποθήκευση διαπιστευτηρίων 7.15 Ανακεφαλαίωση 8. Εξατομίκευση του Git 8.1 Διαμόρφωση Git 8.2 Γνωρίσματα του Git 8.3 Τα άγκιστρα του Git 8.4 Ένα παράδειγμα επιβολής πολιτικής από το Git 8.5 Ανακεφαλαίωση 9. Το Git και άλλα συστήματα 9.1 Το Git ως πελάτης 9.2 Μετανάστευση στο Git 9.3 Ανακεφαλαίωση 10. Εσωτερική λειτουργία του Git 10.1 Διοχετεύσεις και πορσελάνες 10.2 Αντικείμενα του Git 10.3 Αναφορές του Git 10.4 Πακετάρισμα αρχείων 10.5 Τα refspec 10.6 Πρωτόκολλα μεταφοράς 10.7 Διατήρηση και ανάκτηση δεδομένων 10.8 Μεταβλητές περιβάλλοντος 10.9 Ανακεφαλαίωση A1. Appendix A: Το Git σε άλλα περιβάλλοντα A1.1 Γραφικές διεπαφές A1.2 Το Git στο Visual Studio A1.3 Git στο Visual Studio Code A1.4 Git στο IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git στο Sublime Text A1.6 Το Git στο Bash A1.7 Το Git στο Zsh A1.8 Το Git στο Powershell A1.9 Ανακεφαλαίωση A2. Appendix B: Ενσωμάτωση του Git στις εφαρμογές μας A2.1 Γραμμή εντολών Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Appendix C: Εντολές Git A3.1 Ρύθμιση και διαμόρφωση A3.2 Λήψη και δημιουργία έργων A3.3 Βασική λήψη στιγμιοτύπων A3.4 Διακλάδωση και συγχώνευση A3.5 Κοινή χρήση και ενημέρωση έργων A3.6 Επιθεώρηση και σύγκριση A3.7 Αποσφαλμάτωση A3.8 Επιθέματα A3.9 Ηλεκτρονικό ταχυδρομείο A3.10 Εξωτερικά Συστήματα A3.11 Διοίκηση A3.12 Εντολές διοχέτευσης 2nd Edition 2.1 Τα θεμελιώδη στοιχεία του Git - Απόκτηση αποθετηρίου Git Αν μπορούσατε να διαβάσετε μόνο ένα κεφάλαιο για το Git, τότε θα έπρεπε να διαβάσετε αυτό. Το κεφάλαιο αυτό καλύπτει κάθε βασική εντολή που χρειάζεστε για να επιτελέσετε τη συντριπτική πλειονότητα των ενεργειών που μπορείτε να κάνετε με το Git. Στο τέλος αυτού το κεφαλαίου, θα μπορείτε να παραμετροποιείτε (configure) και να αρχικοποιείτε ένα αποθετήριο, να αρχίζετε και να σταματάτε την παρακολούθηση (tracking) αρχείων, να τοποθετείτε αλλαγές στον προθάλαμο (stage) ή να τις υποβάλλετε (commit). Επίσης θα σας δείξουμε πώς να ρυθμίσετε το Git ώστε να αγνοεί ορισμένα αρχεία και μοτίβα αρχείων, πώς να αναιρείτε λάθη εύκολα και γρήγορα, πώς να περιηγείστε στο ιστορικό του έργου σας ώστε να βλέπετε τις αλλαγές μεταξύ των υποβολών και πώς να ωθείτε (push) σε και να τραβάτε (pull) από απομακρυσμένα αποθετήρια. Απόκτηση αποθετηρίου Git Μπορούμε να δημιουργήσουμε ένα έργο στο Git χρησιμοποιώντας δύο βασικές προσεγγίσεις. Να πάρουμε ένα έργο που έχουμε σε κάποιο υπάρχοντα κατάλογο που δεν βρίσκεται υπό έλεγχο εκδόσεων και να τον μετατρέψουμε σε ένα αποθετήριο Git Να κλωνοποιήσουμε (clone) ένα υπάρχον αποθετήριο Git από κάπου αλλού. Σε κάθε περίπτωση, θα αποκτήσουμε ένα αποθετήριο Git τοπικά στον υπολογιστή μας. Αρχικοποίηση αποθετηρίου σε έναν υπάρχοντα κατάλογο Αν έχουμε ένα έργο σε έναν κατάλογο που δεν βρίσκεται υπό έλεγχο εκδόσεων και θέλουμε να ξεκινήσουμε να τον ελέγχουμε με το Git, πρώτα πρέπει να πάμε σε αυτό τον κατάλογο. Αυτό γίνεται με διαφορετικό τρόπο ανάλογα με το σύστημά μας. Σε Linux $ cd /home/user/my_project Σε macOS: $ cd /Users/user/my_project Σε Windows: $ cd C:/Users/user/my_project και πληκτρολογούμε: $ git init Η εντολή αυτή δημιουργεί έναν νέο υποκατάλογο με το όνομα .git ο οποίος περιέχει όλα τα απαραίτητα αρχεία για το αποθετήριο — ένα σκελετό για το αποθετήριό μας. Στο σημείο αυτό, τίποτα δεν παρακολουθείται ακόμα από το έργο μας. Βλ. Εσωτερική λειτουργία του Git για περισσότερες πληροφορίες σχετικά με το τι ακριβώς αρχεία περιέχονται στον κατάλογο .git , που μόλις δημιουργήσαμε. Αν θέλουμε να ξεκινήσουμε τον έλεγχο έκδοσης στα υπάρχοντα αρχεία (εν εντιθέσει με ένα κενό κατάλογο), θα πρέπει να ξεκινήσουμε την παρακολούθηση αυτών των αρχείων και να κάνουμε την πρώτη υποβολή (commit). Για να το πετύχουμε αυτό θα χρειαστούμε μερικές εντολές git add οι οποίες προσδιορίζουν τα αρχεία που θέλουμε να παρακολουθούμε και μια εντολή git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' Θα εξετάσουμε σε λίγο τι κάνουν οι παραπάνω εντολές. Στο σημείο αυτό, έχουμε ένα αποθετήριο Git με κάποια παρακολουθούμενα αρχεία και μια αρχική υποβολή. Κλωνοποίηση υπάρχοντος αποθετηρίου Αν θέλουμε να αποκτήσουμε ένα αντίγραφο ενός υπάρχοντος αποθετηρίου Git — για παράδειγμα, ένα έργο στο οποίο θα θέλουμε να συνεισφέρουμε — η εντολή που χρειαζόμαστε είναι git clone . Αν είμαστε εξεικοιωμένοι με άλλα συστήματα ελέγχου έκδοσης όπως το Subversion, θα παρατηρήσουμε ότι η εντολή είναι "clone" και όχι "checkout". Αυτή είναι μια σημαντική διάκριση — το Git δεν παίρνει απλά ένα αντίγραφο της τρέχουσας κατάστασης του αποθετηρίου, αλλά ένα πλήρες αντίγραφο σχεδόν όλων των δεδομένων που βρίσκονται στον διακομιστή. Με την εντολή git clone όλες οι εκδόσεις κάθε αρχείου του έργου αποθηκεύονται τοπικά. Μάλιστα, αν ο δίσκος του διακομιστή μας αλλοιωθεί, μπορούμε να χρησιμοποιήσουμε οποιονδήποτε από τους κλώνους του ώστε να θέσουμε τον διακομιστή στην κατάσταση που ήταν όταν κλωνοποιήθηκε (μπορεί να χαθεί κάποιο άγκιστρο (hook) από την μεριά του διακομιστή, αλλά τα δεδομένα με έκδοση θα είναι εκεί — βλ. Εγκατάσταση του Git σε διακομιστή για περισσότερες πληροφορίες). Για να κλωνοποιήσουμε ένα αποθετήριο, εκτελούμε την εντολή git clone <url> . Για παράδειγμα, αν θέλουμε να κλωνοποιήσουμε τη βιβλιοθήκη libgit2 του Git, θα πρέπει να εκτελέσουμε: $ git clone https://github.com/libgit2/libgit2 Η εντολή αυτή δημιουργεί έναν κατάλογο με το όνομα “libgit2”, αρχικοποιεί έναν κατάλογο .git μέσα σε αυτό, κατεβάζει όλα τα δεδομένα αυτού του αποθετηρίου καθώς και ένα αντίγραφο από την τελευταία έκδοση. Αν περιηγηθούμε στον καινούριο κατάλογο libgit2 , θα δούμε τα αρχεία του έργου μέσα σε αυτό, έτοιμα προς χρήση ή επεξεργασία. Αν θέλουμε να κλωνοποιήσουμε το αποθετήριο σε έναν κατάλογο με διαφορετικό όνομα, μπορούμε να το ορίσουμε με την παρακάτω εναλλακτική της εντολής: $ git clone https://github.com/libgit2/libgit2 mylibgit Η εντολή αυτή έχει το ίδιο αποτέλεσμα με την προηγούμενη, με τη διαφορά ότι ο κατάλογος που θα δημιουργηθεί θα ονομάζεται mylibgit . Το Git συνεργάζεται με διάφορα πρωτόκολλα μεταφοράς που μπορούμε να χρησιμοποιήσουμε. Το προηγούμενο παράδειγμα χρησιμοποεί το πρωτόκολλο https:// , ενδέχεται επίσης να δούμε το git:// ή το user@server:path/to/repo.git το οποίο χρησιμοποιεί το πρωτόκολλο μεταφοράς SSH. Το κεφάλαιο Εγκατάσταση του Git σε διακομιστή παρουσιάζει όλες τις διαθέσιμες εναλλακτικές που με τις οποίες μπορεί ένας διακομιστής να μας δώσει πρόσβαση σε ένα αποθετήριο Git, καθώς και τα πλεονεκτήματα και μειονεκτήματα της κάθε εναλλακτικής. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/dantech-corporation-mladu/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_full-click
MLADU | LinkedIn Skip to main content LinkedIn Dantech Corporation Inc. in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in MLADU Managed File Transfer (MFT) Software by Dantech Corporation Inc. See who's skilled in this Add as skill Request demo Report this product About MLADU redefines data transfer with industry-leading scalability, security, availability, and performance. Designed for businesses of all sizes and across all industries, MLADU effortlessly handles data transfers of any volume, supporting use cases such as acquisitions, clinical trials, omics pipelines, scientific research, enterprise applications, and big data analytics. Its robust features empower you to easily initiate, track, organize, and configure transfers to align perfectly with your business, organizational, and compliance needs. Engineered to overcome the toughest data transfer challenges, MLADU addresses complexities like data corruption, system incompatibility, duplication, and massive data volumes. It also ensures the highest standards of security, regulatory compliance, and data integrity. Whether you're navigating diverse platforms, managing varying formats, or mitigating potential human errors, MLADU is your trusted partner in seamless, reliable data transfer solutions Media Products media viewer No more previous content What is MLADU MLADU redefines data transfer with industry-leading scalability, security, availability, and performance. Designed for businesses of all sizes and across all industries, MLADU effortlessly handles data transfers of any volume, supporting use cases such as acquisitions, clinical trials, omics pipelines, scientific research, enterprise applications, and big data analytics. Data Transfers Awaiting Approval MLADU has a built-in workflow that lets you auto approve or request approval for your data transfers Data Transfers Completed MLADU allows you to quickly visually track the progression of your data transfer Complete Audit and Tracking of All Data Transfers MLADU automatically provides a detailed history of all data transfers for your data audit and chain of custody tracking. MLADU Analytics MLADU provides built-in analytics across all data transfers No more next content Similar products Axway Managed File Transfer Axway Managed File Transfer Managed File Transfer (MFT) Software Progress MOVEit Progress MOVEit Managed File Transfer (MFT) Software Serv-U Managed File Transfer Server Serv-U Managed File Transfer Server Managed File Transfer (MFT) Software JSCAPE by Redwood JSCAPE by Redwood Managed File Transfer (MFT) Software Cerberus by Redwood Cerberus by Redwood Managed File Transfer (MFT) Software dDataBox dDataBox Managed File Transfer (MFT) Software Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://www.linkedin.com/products/kaspersky-ddos-protection/?trk=products_seo_search
Kaspersky DDoS Protection | LinkedIn Skip to main content LinkedIn Kaspersky in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Kaspersky DDoS Protection DDoS Protection Software by Kaspersky See who's skilled in this Add as skill Learn more Report this product About Distributed Denial of Service (DDoS) can cause financial costs for your business. Keep your service protected with Kaspersky DDoS-Protection! Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Sign in to see more Show more Show less Kaspersky products Kaspersky Container Security Kaspersky Container Security Container Security Software Kaspersky Digital Footprint Intelligence Kaspersky Digital Footprint Intelligence Digital Forensics Software Kaspersky Industrial Cybersecurity Kaspersky Industrial Cybersecurity Enterprise Architecture Software Kaspersky Next Kaspersky Next Endpoint Detection & Response (EDR) Software Kaspersky Password Manager Kaspersky Password Manager Password Management Software Kaspersky Premium Kaspersky Premium Antivirus Software Kaspersky Safe Kids Kaspersky Safe Kids Antivirus Software Kaspersky Threat Intelligence Kaspersky Threat Intelligence Threat Intelligence Platforms Kaspersky VPN Kaspersky VPN Virtual Private Network (VPN) Software KasperskyOS KasperskyOS Operating Systems Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/pl/v2/Git-na-serwerze-Smart-HTTP
Git - Smart HTTP About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Pierwsze kroki 1.1 Wprowadzenie do kontroli wersji 1.2 Krótka historia Git 1.3 Podstawy Git 1.4 Linia poleceń 1.5 Instalacja Git 1.6 Wstępna konfiguracja Git 1.7 Uzyskiwanie pomocy 1.8 Podsumowanie 2. Podstawy Gita 2.1 Pierwsze repozytorium Gita 2.2 Rejestrowanie zmian w repozytorium 2.3 Podgląd historii rewizji 2.4 Cofanie zmian 2.5 Praca ze zdalnym repozytorium 2.6 Tagowanie 2.7 Aliasy 2.8 Podsumowanie 3. Gałęzie Gita 3.1 Czym jest gałąź 3.2 Podstawy rozgałęziania i scalania 3.3 Zarządzanie gałęziami 3.4 Sposoby pracy z gałęziami 3.5 Gałęzie zdalne 3.6 Zmiana bazy 3.7 Podsumowanie 4. Git na serwerze 4.1 Protokoły 4.2 Uruchomienie Git na serwerze 4.3 Generowanie Twojego publicznego klucza SSH 4.4 Konfigurowanie serwera 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Inne opcje hostowania przez podmioty zewnętrzne 4.10 Podsumowanie 5. Rozproszony Git 5.1 Rozproszone przepływy pracy 5.2 Wgrywanie zmian do projektu 5.3 Utrzymywanie projektu 5.4 Podsumowanie 6. GitHub 6.1 Account Setup and Configuration 6.2 Contributing to a Project 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Narzędzia Gita 7.1 Wskazywanie rewizji 7.2 Interaktywne używanie przechowali 7.3 Schowek i czyszczenie 7.4 Signing Your Work 7.5 Searching 7.6 Przepisywanie historii 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugowanie z Gitem 7.11 Moduły zależne 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Podsumowanie 8. Dostosowywanie Gita 8.1 Konfiguracja Gita 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git i inne systemy 9.1 Git jako klient 9.2 Migracja do Gita 9.3 Podsumowanie 10. Mechanizmy wewnętrzne w Git 10.1 Komendy typu plumbing i porcelain 10.2 Obiekty Gita 10.3 Referencje w Git 10.4 Spakowane pliki (packfiles) 10.5 Refspec 10.6 Protokoły transferu 10.7 Konserwacja i odzyskiwanie danych 10.8 Environment Variables 10.9 Podsumowanie A1. Appendix A: Git in Other Environments A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Summary A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 4.6 Git na serwerze - Smart HTTP Smart HTTP We now have authenticated access though SSH and unauthenticated access through git:// , but there is also a protocol that can do both at the same time. Setting up Smart HTTP is basically just enabling a CGI script that is provided with Git called git-http-backend on the server. This CGI will read the path and headers sent by a git fetch or git push to an HTTP URL and determine if the client can communicate over HTTP (which is true for any client since version 1.6.6). If the CGI sees that the client is smart, it will communicate smartly with it, otherwise it will fall back to the dumb behavior (so it is backward compatible for reads with older clients). Let’s walk through a very basic setup. We’ll set this up with Apache as the CGI server. If you don’t have Apache setup, you can do so on a Linux box with something like this: $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env This also enables the mod_cgi , mod_alias , and mod_env modules, which are all needed for this to work properly. Next we need to add some things to the Apache configuration to run the git-http-backend as the handler for anything coming into the /git path of your web server. SetEnv GIT_PROJECT_ROOT /opt/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/libexec/git-core/git-http-backend/ If you leave out GIT_HTTP_EXPORT_ALL environment variable, then Git will only serve to unauthenticated clients the repositories with the git-daemon-export-ok file in them, just like the Git daemon did. Then you’ll have to tell Apache to allow requests to that path with something like this: <Directory "/usr/lib/git-core*"> Options ExecCGI Indexes Order allow,deny Allow from all Require all granted </Directory> Finally you’ll want to make writes be authenticated somehow, possibly with an Auth block like this: <LocationMatch "^/git/.*/git-receive-pack$"> AuthType Basic AuthName "Git Access" AuthUserFile /opt/git/.htpasswd Require valid-user </LocationMatch> That will require you to create a .htaccess file containing the passwords of all the valid users. Here is an example of adding a “schacon” user to the file: $ htdigest -c /opt/git/.htpasswd "Git Access" schacon There are tons of ways to have Apache authenticate users, you’ll have to choose and implement one of them. This is just the simplest example we could come up with. You’ll also almost certainly want to set this up over SSL so all this data is encrypted. We don’t want to go too far down the rabbit hole of Apache configuration specifics, since you could well be using a different server or have different authentication needs. The idea is that Git comes with a CGI called git-http-backend that when invoked will do all the negotiation to send and receive data over HTTP. It does not implement any authentication itself, but that can easily be controlled at the layer of the web server that invokes it. You can do this with nearly any CGI-capable web server, so go with the one that you know best. Note For more information on configuring authentication in Apache, check out the Apache docs here: http://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/pt-br/v2/Fundamentos-de-Git-Obtendo-um-Reposit%c3%b3rio-Git
Git - Obtendo um Repositório Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Começando 1.1 Sobre Controle de Versão 1.2 Uma Breve História do Git 1.3 O Básico do Git 1.4 A Linha de Comando 1.5 Instalando o Git 1.6 Configuração Inicial do Git 1.7 Pedindo Ajuda 1.8 Sumário 2. Fundamentos de Git 2.1 Obtendo um Repositório Git 2.2 Gravando Alterações em Seu Repositório 2.3 Vendo o histórico de Commits 2.4 Desfazendo coisas 2.5 Trabalhando de Forma Remota 2.6 Criando Tags 2.7 Apelidos Git 2.8 Sumário 3. Branches no Git 3.1 Branches em poucas palavras 3.2 O básico de Ramificação (Branch) e Mesclagem (Merge) 3.3 Gestão de Branches 3.4 Fluxo de Branches 3.5 Branches remotos 3.6 Rebase 3.7 Sumário 4. Git no servidor 4.1 Os Protocolos 4.2 Getting Git on a Server 4.3 Gerando Sua Chave Pública SSH 4.4 Setting Up the Server 4.5 Git Daemon 4.6 Smart HTTP 4.7 GitWeb 4.8 GitLab 4.9 Third Party Hosted Options 4.10 Sumário 5. Distributed Git 5.1 Fluxos de Trabalho Distribuídos 5.2 Contribuindo com um Projeto 5.3 Maintaining a Project 5.4 Summary 6. GitHub 6.1 Configurando uma conta 6.2 Contribuindo em um projeto 6.3 Maintaining a Project 6.4 Managing an organization 6.5 Scripting GitHub 6.6 Summary 7. Git Tools 7.1 Revision Selection 7.2 Interactive Staging 7.3 Stashing and Cleaning 7.4 Signing Your Work 7.5 Searching 7.6 Rewriting History 7.7 Reset Demystified 7.8 Advanced Merging 7.9 Rerere 7.10 Debugging with Git 7.11 Submodules 7.12 Bundling 7.13 Replace 7.14 Credential Storage 7.15 Summary 8. Customizing Git 8.1 Git Configuration 8.2 Git Attributes 8.3 Git Hooks 8.4 An Example Git-Enforced Policy 8.5 Summary 9. Git and Other Systems 9.1 Git as a Client 9.2 Migrating to Git 9.3 Summary 10. Funcionamento Interno do Git 10.1 Encanamento e Porcelana 10.2 Objetos do Git 10.3 Referências do Git 10.4 Packfiles 10.5 The Refspec 10.6 Transfer Protocols 10.7 Maintenance and Data Recovery 10.8 Variáveis de ambiente 10.9 Sumário A1. Appendix A: Git em Outros Ambientes A1.1 Graphical Interfaces A1.2 Git in Visual Studio A1.3 Git in Eclipse A1.4 Git in Bash A1.5 Git in Zsh A1.6 Git in Powershell A1.7 Resumo A2. Appendix B: Embedding Git in your Applications A2.1 Command-line Git A2.2 Libgit2 A2.3 JGit A3. Appendix C: Git Commands A3.1 Setup and Config A3.2 Getting and Creating Projects A3.3 Basic Snapshotting A3.4 Branching and Merging A3.5 Sharing and Updating Projects A3.6 Inspection and Comparison A3.7 Debugging A3.8 Patching A3.9 Email A3.10 External Systems A3.11 Administration A3.12 Plumbing Commands 2nd Edition 2.1 Fundamentos de Git - Obtendo um Repositório Git Se você pode ler apenas um capítulo antes de começar a usar o Git, este é ele. Este capítulo cobre cada comando básico que você precisa para fazer a maior parte das coisas com as quais eventualmente você vai se deparar durante seu uso do Git. No final deste capítulo, você será capaz de configurar e inicializar um repositório, iniciar e interromper o rastreamento de arquivos, usar a área de stage e realizar commits das alterações. Também mostraremos como configurar o Git para ignorar certos arquivos e padrões de arquivo, como desfazer erros de maneira rápida e fácil, como navegar no histórico do seu projeto e visualizar alterações entre commits e como fazer push e pull em repositórios remotos. Obtendo um Repositório Git Você pode obter um projeto Git utilizando duas formas principais. 1. Você pode pegar um diretório local que atualmente não está sob controle de versão e transformá-lo em um repositório Git, ou 2. Você pode fazer um clone de um repositório Git existente em outro lugar. Inicializando um Repositório em um Diretório Existente Para você começar a monitorar um projeto existente com Git, você deve ir para o diretório desse projeto. Se você nunca fez isso, use o comando a seguir, que terá uma pequena diferença dependendo do sistema em que está executando: para Linux: $ cd /home/user/your_repository para Mac: $ cd /Users/user/your_repository para Windows: $ cd /c/user/your_repository depois digite: $ git init Isso cria um novo subdiretório chamado .git que contém todos os arquivos necessários de seu repositório – um esqueleto de repositório Git. Neste ponto, nada em seu projeto é monitorado ainda. (Veja [ch10-git-internals] para mais informações sobre quais arquivos estão contidos no diretório .git que foi criado.) Se você quer começar a controlar o versionamento dos arquivos existentes (ao contrário de um diretório vazio), você provavelmente deve começar a monitorar esses arquivos e fazer um commit inicial. Você pode fazer isso com alguns comandos git add que especificam os arquivos que você quer monitorar, seguido de um git commit : $ git add *.c $ git add LICENSE $ git commit -m 'initial project version' Nós já veremos o que esses comandos fazem. Mas neste ponto você já tem um repositório Git com arquivos monitorados e um commit inicial. Clonando um Repositório Existente Caso você queira obter a cópia de um repositório Git existente – por exemplo, um projeto que você queira contribuir – o comando para isso é git clone . Se você estiver familiarizado com outros sistemas VCS, tal como Subversion, você vai notar que o comando é clone e não checkout . Essa é uma diferença importante – em vez de receber apenas uma cópia para trabalho, o Git recebe uma cópia completa de praticamente todos os dados que o servidor possui. Cada versão de cada arquivo no histórico do projeto é obtida por padrão quando você executa git clone . De fato, se o disco do servidor ficar corrompido, em geral você pode usar qualquer uma das cópias de qualquer um dos clientes para reverter o servidor ao estado em que estava quando foi clonado (talvez você perca algumas configurações do servidor, mas todos os dados versionados estarão lá — veja Getting Git on a Server para mais detalhes). Você clona um repositório com git clone [url] . Por exemplo, caso você queria clonar a biblioteca Git Linkable chamada libgit2, você pode fazer da seguinte forma: $ git clone https://github.com/libgit2/libgit2 Isso cria um diretório chamado libgit2 , inicializa um diretório .git dentro dele, recebe todos os dados deste repositório e deixa disponível para trabalho a cópia da última versão. Se você entrar no novo diretório libgit2 , você verá os arquivos do projeto nele, pronto para serem editados ou utilizados. Caso você queira clonar o repositório em um diretório diferente de libgit2 , é possível especificar esse diretório utilizando a opção abaixo: $ git clone https://github.com/libgit2/libgit2 mylibgit Este comando faz exatamente a mesma coisa que o anterior, mas o diretório de destino será chamado mylibgit . O Git possui diversos protocolos de transferência que você pode utilizar. O exemplo anterior usa o protocolo https:// , mas você também pode ver git:// ou user@server:path/to/repo.git , que usam o protocolo de transferência SSH. Em Getting Git on a Server é apresentado todas as opções disponíveis com as quais o servidor pode ser configurado para acessar o seu repositório Git, e os prós e contras de cada uma. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Branchen-in-Git-Samenvatting
Git - Samenvatting About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 3.7 Branchen in Git - Samenvatting Samenvatting We hebben de basis van branchen en mergen in Git behandeld. Je zou je op je gemak moeten voelen met het maken en omschakelen naar nieuwe branches, omschakelen tussen branches, en lokale branches te mergen. Je zou ook in staat moeten zijn om je branches te delen door ze naar een gedeelde server te pushen, met anderen op gedeelde branches samen te werken en je branches te rebasen voordat ze gedeeld worden. In het volgende deel gaan we de materie behandelen wat je nodig gaat hebben om jouw eigen Git repository-hosting server op te zetten. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/intercom/?trk=products_details_guest_similar_products_section_similar_products_section_product_link_result-card_full-click
Intercom | LinkedIn Skip to main content LinkedIn Intercom in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Intercom Service Desk Software by Intercom See who's skilled in this Add as skill Learn more Report this product About Intercom is the only complete AI-first customer service platform, enhancing the customer experience, improving operational efficiency, and scaling with your business every step of the way. Our AI-first platform is built on a single AI system, with two major components that will allow you to deliver the remarkable customer service you’ve spent decades striving for: Fin AI Agent: The human-quality AI agent that works with any Helpdesk Intercom Customer Service Platform: The AI-first platform trusted by thousands of support leaders. Founded in 2011 and backed by leading venture capitalists, including Kleiner Perkins, Bessemer Venture Partners and Social Capital, Intercom believes there's a new way to do customer service. Featured customers of Intercom Coda Software Development 43,019 followers Chess.com Technology, Information and Internet 53,922 followers Cloudinary Technology, Information and Internet 61,404 followers Outschool Technology, Information and Internet 34,687 followers Flipdish Software Development 27,258 followers Tines Software Development 55,851 followers Shutterstock Software Development 128,637 followers Culture Amp Software Development 117,670 followers Nuuly Retail 45,152 followers Lightspeed Commerce Software Development 97,468 followers Atlassian Software Development 2,340,590 followers monday.com Software Development 328,493 followers Anthropic Research Services 1,881,196 followers Betr Spectator Sports 11,640 followers MONY Group plc Software Development 20,081 followers Amplitude Software Development 103,066 followers Carvana Technology, Information and Internet 98,101 followers Show more Show less Similar products Jira Service Management Jira Service Management Service Desk Software Freshservice Freshservice Service Desk Software Atomicwork Atomicwork Service Desk Software SolarWinds Service Desk SolarWinds Service Desk Service Desk Software Service Desk Service Desk Service Desk Software TOPdesk TOPdesk Service Desk Software Sign in to see more Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/GitHub-GitHub-Scripten
Git - GitHub Scripten About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 6.5 GitHub - GitHub Scripten GitHub Scripten We hebben nu alle hoofdfuncties en workflows van GitHub hebben behandeld, maar elke grote groep of project zal aanpassingen hebben die ze willen maken of services van buitenaf die ze willen integreren. Gelukkig voor ons is GitHub echt op vele manieren redelijk te hacken. In deze paragraaf zullen we behandelen hoe het zgn. GitHub haken (hooks) en API systeem te gebruiken om GitHub zich te laten gedragen zoals we willen. Services en Hooks Het Hooks en Services deel van het GitHub repository beheer is de eenvoudigste manier om GitHub te laten samenwerken met externe systemen. Services Eerst zullen we naar Services kijken. Zowel Hooks als Services integratie kunnen in het Settings gedeelte van je repository gevonden worden, waar we eerder naar gekeken hebben bij het toevoegen van medewerkers en het wijzigen van de standaard-branch van je project. Onder de “Webhooks and Services” tab zul je iets als Services and Hooks configuratie deel. zien. Figuur 130. Services and Hooks configuratie deel. Er zijn tientallen services waar je uit kunt kiezen, de meeste zijn integraties naar andere commerciele en open source systemen. De meeste daarvan zijn Continuous Integration services, bug en issue trackers, chat room systemen en documentatie systemen. We zullen je door het opzetten van een eenvoudige leiden: de Email-hook. Als je “email” kiest uit de “Add Service” dropdown, krijg je een configuratie scherm zoals E-mail service configuratie. . Figuur 131. E-mail service configuratie. In dit geval, als we de “Add service” knop klikken, zal het e-mail adres die we intypen elke keer een bericht ontvangen als iemand naar de repository pusht. Services kunnen luisteren naar verschillende type gebeurtenissen, maar de meeste luisteren alleen naar push-events en doen dan iets met die gegevens. Als er een systeem is dat je gebruikt en die je wilt integreren met GitHub, zou je hier moeten kijken om te zien of er een bestaande service integratie beschikbaar is. Je zou, als je bijvoorbeeld Jenkins gebruikt om tests te draaien op je codebase, de ingebouwde Jenkins service integratie kunnen aanzetten om een testrun af te trappen elke keer als iemand naar jouw repository pusht. Hooks Als je iets meer specifieker nodig hebt, of je wilt een service of site integreren die niet in de lijst staat, kan je in plaats daarvan het meer generieke hooks systeem gebruiken. GitHub repository hooks zijn redelijk eenvoudig. Je geeft een URL op en GitHub zal een HTTP payload posten op die URL bij elke gebeurtenis dat je maar wilt. Hoe dit globaal werkt is dat je een kleine web service kunt opzetten die naar een GitHub hook payload luistert en dan iets met de gegevens doet als het is ontvangen. Om een hook aan te zetten, klik je de “Add webhook” knop in Services and Hooks configuratie deel. . Dit leidt je naar een pagina die eruit ziet als Web hook configuratie. . Figuur 132. Web hook configuratie. De configuratie van een web hook is redelijk eenvoudig. In de meeste gevallen voer je een URL in en een geheime sleutel en klikt “Add webhook”. Er zijn een paar opties voor welke gebeurtenissen je wilt waarvor GitHub je een payload stuurt — standaard is om alleen een payload te ontvangen voor de push gebeurtenis, als iemand nieuwe code naar een van de branches uit je repository pusht. Laten we een kort voorbeeld van een web service bekijken die je zou kunnen opzetten om een web hook te verwerken. We zullen het Ruby web framework Sinatra gebruiken omdat dit redelijk bondig is en je in staat zou moeten zijn om snel te zien wat we aan het doen zijn. Laten we stellen dat we een e-mail willen ontvangen als een bepaald persoon naar een specifieke branch van ons project pusht waarin een zeker bestand wordt gewijzigd. We kunnen dat relatief eenvoudig doen met code zoals deze: require 'sinatra' require 'json' require 'mail' post '/payload' do push = JSON.parse(request.body.read) # parse the JSON # gather the data we're looking for pusher = push["pusher"]["name"] branch = push["ref"] # get a list of all the files touched files = push["commits"].map do |commit| commit['added'] + commit['modified'] + commit['removed'] end files = files.flatten.uniq # check for our criteria if pusher == 'schacon' && branch == 'ref/heads/special-branch' && files.include?('special-file.txt') Mail.deliver do from 'tchacon@example.com' to 'tchacon@example.com' subject 'Scott Changed the File' body "ALARM" end end end Hier nemen we de JSON payload die GitHub ons levert en kijken na wie gepusht heeft, naar welke branch hij gepusht heeft en welke bestanden geraakt zijn in alle commits die zijn gepusht. Dan houden we die gegevens tegen onze criteria en sturen een e-mail als ze passen. Om zoiets te kunnen ontwikkelen en testen, heb je een nette ontwikkelaarsscherm in hetzelfde scherm waar je ook de hook ingesteld hebt. Je kunt de laatste paar leveranties die GitHub heeft proberen te maken voor die webhook zien. Voor elke hook kan je uitvinden wanneer het was afgeleverd, of dit succesvol was en de body en headers voor zowel de vraag en het antwoord. Dit maakt het ongelofelijk eenvoudig om je hooks te testen en te debuggen. Figuur 133. Web hook debug informatie. De andere geweldige mogelijkheid van dit is dat je elk van de payloads opnieuw kan laten afleveren om je service makkelijk te testen. Voor meer informatie over hoe webhooks te schrijven en alle verschillende type gebeurtenissen waar je naar kunt verwijzen we je naar de GitHub Developer documentatie op: https://developer.github.com/webhooks/ De GitHub API Services en hooks bieden je een manier om push berichten te ontvangen van gebeurtenissen die plaatsvinden op je repositories, maar wat nu als je meer informatie hierover nodig hebt? Wat als je iets wilt automatiseren zoals medewerkers toevoegen of issues markeren? Dit is waar de GitHub API handig bij gebruikt kan worden. GitHub heeft ongelofelijk veel API endpoints om bijna alles wat je op de website kan doen op een geautomatiseerde manier kan doen. In deze paragraaf zullen we leren hoe te authenticeren en te verbinden met de API, hoe te reageren op een issue en hoe de status van een Pull Request te wijzigen middels de API. Eenvoudig gebruik Het meest basale wat je kunt doen is een simpele GET request op een endpoint die geen authenticatie behoeft. Dit zou een gebruiker of read-only informatie op een open source project kunnen zijn. Bijvoorbeeld, als we meer willen weten van een gebruiker genaamd “schacon”, kunnen we zoiets uitvoeren: $ curl https://api.github.com/users/schacon { "login": "schacon", "id": 70, "avatar_url": "https://avatars.githubusercontent.com/u/70", # … "name": "Scott Chacon", "company": "GitHub", "following": 19, "created_at": "2008-01-27T17:19:28Z", "updated_at": "2014-06-10T02:37:23Z" } Er zijn honderden van soortgelijke endpoints als deze om informatie over organisaties, projecten, issues, commits te verkrijgen — zo ongeveer alles wat je publiekelijk kan zien op GitHub. Je kunt de API zelfs gebruiken om willekeurige MarkDown te tonen of een .gitignore template vinden. $ curl https://api.github.com/gitignore/templates/Java { "name": "Java", "source": "*.class # Mobile Tools for Java (J2ME) .mtj.tmp/ # Package Files # *.jar *.war *.ear # virtual machine crash logs, see https://www.java.com/en/download/help/error_hotspot.xml hs_err_pid* " } Reageren op een issue Echter, als je een actie wilt uitvoeren op de website zoals reageren op een Issue of Pull Request of als je gesloten informatie wilt zien of ermee interacteren, zal je je moeten authenticeren. Er zijn verschillende manieren om je te authenticeren. Je kunt eenvoudige authenticatie gebruiken met gewoon je gebruikersnaam en wachtwoord, maar over het algemeen is het een beter idee om een persoonlijke toegangs bewijs (access token) te gebruiken. Deze kan je genereren vanaf de “Applications” tab van je instellingen pagina. Figuur 134. Genereer je access token vanaf de “Applications” tab op je instellingen pagina. Het zal je vragen welke contexten je wilt voor dit token en een omschrijving. Zorg ervoor dat je een goede omschrijving gebruikt zo dat je met vertrouwen het token kan weggooien als je script of applicatie niet langer meer in gebruik is. GitHub laat je het token maar één keer zien, dus zorg ervoor dat je het kopieert. Je kunt deze nu gebruiken om te authenticeren in je script in plaats van een gebruikersnaam en wachtwoord. Dit is prettig omdat je de context waarin je iets wilt doen kan beperken en het token is weer intrekbaar. Het heeft ook het bijkomende voordeel dat het je aanvraag limiet verhoogt. Zonder authenticatie ben je gelimiteerd tot 60 aanvragen per uur. Als je authenticeert kan je tot 5.000 aanvragen per uur doen. Dus laten we het gebruiken om een reactie te geven op een van onze issues. Stel dat we een reactie willen geven op een specifieke issue, Issue #6. Om dit te doen moeten we een HTTP POST request op repos/<user>/<repo>/issues/<num>/comments uitvoeren met het token wat we zojuist gegeneerd hebben als een Authorization header. $ curl -H "Content-Type: application/json" \ -H "Authorization: token TOKEN" \ --data '{"body":"A new comment, :+1:"}' \ https://api.github.com/repos/schacon/blink/issues/6/comments { "id": 58322100, "html_url": "https://github.com/schacon/blink/issues/6#issuecomment-58322100", ... "user": { "login": "tonychacon", "id": 7874698, "avatar_url": "https://avatars.githubusercontent.com/u/7874698?v=2", "type": "User", }, "created_at": "2014-10-08T07:48:19Z", "updated_at": "2014-10-08T07:48:19Z", "body": "A new comment, :+1:" } Als je nu naar dat issue gaat, kan je de reactie zien dat we zojuist succesvol gepost hebben in Een commentaar gepost via de GitHub API. . Figuur 135. Een commentaar gepost via de GitHub API. Je kun de API gebruiken om zo ongeveer alles te doen wat je op website kunt doen — mijlpalen maken en zetten, mensen aan issues en pull requests toewijzen, labels maken en wijzigen, commit gegevens benaderen, nieuwe commits en branches maken, Pull Requests openen, sluiten of mergen, teams maken en wijzigen, reageren op regels code in een Pull Request, op de site zoeken enzovoorts, etcetera. De status van een Pull Request wijzigen We zullen nog een laatste voorbeeld bekijken omdat het echt handig is als je werkt met Pull Requests. Elke commit kan een of meerdere statussen hebben en er is een API om een status toe te voegen en deze uit te vragen. De meeste Continuous Integration en test services gebruiken deze API om op pushes te reageren door de code die is gepusht te testen, en dan terug te melden of die commit alle tests heeft gepasseerd. Je kunt dit ook gebruiken om te kijken of het commit bericht juist is geformatteerd, of de indiener al je bijdrage richtlijnen heeft gevolgd, of de commit juist getekend was — verzin het maar. Laten we stellen dat je een webhook op je repository ingericht hebt die een kleine webservice aanroept die controleert of er een Signed-off-by letterreeks in het commit bericht voorkomt. require 'httparty' require 'sinatra' require 'json' post '/payload' do push = JSON.parse(request.body.read) # parse the JSON repo_name = push['repository']['full_name'] # look through each commit message push["commits"].each do |commit| # look for a Signed-off-by string if /Signed-off-by/.match commit['message'] state = 'success' description = 'Successfully signed off!' else state = 'failure' description = 'No signoff found.' end # post status to GitHub sha = commit["id"] status_url = "https://api.github.com/repos/#{repo_name}/statuses/#{sha}" status = { "state" => state, "description" => description, "target_url" => "http://example.com/how-to-signoff", "context" => "validate/signoff" } HTTParty.post(status_url, :body => status.to_json, :headers => { 'Content-Type' => 'application/json', 'User-Agent' => 'tonychacon/signoff', 'Authorization' => "token #{ENV['TOKEN']}" } ) end end Hopelijk is dit redelijk eenvoudig te volgen. In deze webhook-verwerker kijken we door elke commit die zojuist is gepusht, we zoeken naar de reeks Signed-off-by in het commit bericht en tenslotten POSTen we via HTTP naar de /repos/<user>/<repo>/statuses/<commit_sha> API endpoint met de status. In dit geval kan je een status ( success , failure , error ) sturen, een omschrijving wat er gebeurd is, een doel URL waar de gebruiker heen kan gaan voor meer informatie en een “context” in geval er meerdere statussen voor een enkele commit zijn. Bijvoorbeeld, een test-service kan een status aangeven en een validatie service zoals deze kan ook een status aangeven — het “context” veld maakt hierin het onderscheid. Als iemand een nieuwe Pull Request op GitHub opent en deze hook is opgezet, kan je iets zoals Commit status via de API. zien. Figuur 136. Commit status via de API. Je kunt nu een klein groen vinkje zien naast de commit die “Signed-off-by” in het bericht heeft staan en een rode kruis door die waar de auteur is vergeten te tekenen. Je kunt ook zien dat de Pull Request de status krijgt van de laatste commit op de branch en waarschuwt je als het niet geslaagd is. Dit is erg handig als je deze API gebruikt voor test resultaten, zodat je niet per ongeluk iets merget waar de laatste commit tests laat falen. Octokit Alhoewel we ongeveer alles middels curl en eenvoudige HTTP requests hebben gedaan in deze voorbeelden, bestaan er diverse open source libraries die deze API in een meer taalspecifieke manier beschikbaar maken. Op het moment van schrijven, zijn de ondersteunde talen onder andere Go, Objective-C, Ruby en .NET. Kijk op https://github.com/octokit voor meer informatie hiervoor, omdat ze veel van de HTTP voor je afhandelen. Hopelijk kunnen deze instrumenten je helpen om GitHub aan te passen en te wijzigen zodat deze beter werkt voor jouw specifieke workflows. Voor volledige documentatie over de hele API zowel als handleidingen voor veelvoorkomende taken, verwijzen we je naar https://developer.github.com . prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-ngenius-business-analytics/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_full-click
nGenius Business Analytics | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in nGenius Business Analytics Business Intelligence (BI) Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About nGenius Business Analytics delivers a next-generation approach to analytics: unlimited, unchained, and unrestricted – allowing you to deliver subscriber analytics like a rock star. nGenius Business Analytics makes wired data consumable for Big Data analytics. This product is intended for Network Operations Center Service Provider Head of Product Information Technology Specialist Chief Information Officer Chief Executive Officer Head of Mobile Vice President of Business Development Network Security Engineer Head of Network Operations Similar products IBM Cognos Analytics IBM Cognos Analytics Business Intelligence (BI) Software Tableau Tableau Business Intelligence (BI) Software Zoho Analytics Zoho Analytics Business Intelligence (BI) Software Ascend Commercial Suite Ascend Commercial Suite Business Intelligence (BI) Software Tableau Tableau Business Intelligence (BI) Software Workday Analytics and Reporting Workday Analytics and Reporting Business Intelligence (BI) Software Sign in to see more Show more Show less NETSCOUT products Arbor Edge Defense Arbor Edge Defense DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software Arbor Threat Mitigation System (TMS) Arbor Threat Mitigation System (TMS) DDoS Protection Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/nl/v2/Git-op-de-server-Slimme-HTTP
Git - Slimme HTTP About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Aan de slag 1.1 Over versiebeheer 1.2 Een kort historisch overzicht van Git 1.3 Wat is Git? 1.4 De commando-regel 1.5 Git installeren 1.6 Git klaarmaken voor eerste gebruik 1.7 Hulp krijgen 1.8 Samenvatting 2. Git Basics 2.1 Een Git repository verkrijgen 2.2 Wijzigingen aan de repository vastleggen 2.3 De commit geschiedenis bekijken 2.4 Dingen ongedaan maken 2.5 Werken met remotes 2.6 Taggen (Labelen) 2.7 Git aliassen 2.8 Samenvatting 3. Branchen in Git 3.1 Branches in vogelvlucht 3.2 Eenvoudig branchen en mergen 3.3 Branch-beheer 3.4 Branch workflows 3.5 Branches op afstand (Remote branches) 3.6 Rebasen 3.7 Samenvatting 4. Git op de server 4.1 De protocollen 4.2 Git op een server krijgen 4.3 Je publieke SSH sleutel genereren 4.4 De server opzetten 4.5 Git Daemon 4.6 Slimme HTTP 4.7 GitWeb 4.8 GitLab 4.9 Hosting oplossingen van derden 4.10 Samenvatting 5. Gedistribueerd Git 5.1 Gedistribueerde workflows 5.2 Bijdragen aan een project 5.3 Het beheren van een project 5.4 Samenvatting 6. GitHub 6.1 Account setup en configuratie 6.2 Aan een project bijdragen 6.3 Een project onderhouden 6.4 Een organisatie beheren 6.5 GitHub Scripten 6.6 Samenvatting 7. Git Tools 7.1 Revisie Selectie 7.2 Interactief stagen 7.3 Stashen en opschonen 7.4 Je werk tekenen 7.5 Zoeken 7.6 Geschiedenis herschrijven 7.7 Reset ontrafeld 7.8 Mergen voor gevorderden 7.9 Rerere 7.10 Debuggen met Git 7.11 Submodules 7.12 Bundelen 7.13 Vervangen 7.14 Het opslaan van inloggegevens 7.15 Samenvatting 8. Git aanpassen 8.1 Git configuratie 8.2 Git attributen 8.3 Git Hooks 8.4 Een voorbeeld van Git-afgedwongen beleid 8.5 Samenvatting 9. Git en andere systemen 9.1 Git als een client 9.2 Migreren naar Git 9.3 Samenvatting 10. Git Binnenwerk 10.1 Binnenwerk en koetswerk (plumbing and porcelain) 10.2 Git objecten 10.3 Git Referenties 10.4 Packfiles 10.5 De Refspec 10.6 Uitwisseling protocollen 10.7 Onderhoud en gegevensherstel 10.8 Omgevingsvariabelen 10.9 Samenvatting A1. Bijlage A: Git in andere omgevingen A1.1 Grafische interfaces A1.2 Git in Visual Studio A1.3 Git in Visual Studio Code A1.4 Git in Eclipse A1.5 Git in Sublime Text A1.6 Git in Bash A1.7 Git in Zsh A1.8 Git in PowerShell A1.9 Samenvatting A2. Bijlage B: Git in je applicaties inbouwen A2.1 Commando-regel Git A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Bijlage C: Git Commando’s A3.1 Setup en configuratie A3.2 Projecten ophalen en maken A3.3 Basic Snapshotten A3.4 Branchen en mergen A3.5 Projecten delen en bijwerken A3.6 Inspectie en vergelijking A3.7 Debuggen A3.8 Patchen A3.9 Email A3.10 Externe systemen A3.11 Beheer A3.12 Binnenwerk commando’s (plumbing commando’s) 2nd Edition 4.6 Git op de server - Slimme HTTP Slimme HTTP We hebben nu geauthenticeerde toegang via SSH en ongeauthenticeerde toegang met git:// , maar er is een protocol die tot beide in staat is. Slimme HTTP opzetten is eigenlijk gewoon op de server een CGI script git-http-backend geheten activeren die met Git wordt geleverd. Deze CGI leest het pad en de headers die door een git fetch of een git push worden gestuurd aan een HTTP URL en bepaalt of de client via HTTP kan communiceren (wat elke client sinds versie 1.6.6 kan). Als de CGI ziet dat de client "slim" is, zal het op de slimme manier met deze communiceren, anders zal het terugvallen op het domme gedrag (dus het is "backward compatible" met lees acties van oudere clients). Laten we een heel eenvoudige opzet doornemen. We zullen het gaan opzetten met Apache als de CGI server. Als je geen Apache hebt geïnstalleerd, kan je dit op een Linux machine doen met iets wat lijkt op: $ sudo apt-get install apache2 apache2-utils $ a2enmod cgi alias env Dit zet mod_cgi , mod_alias en mod_env modules aan, die alle nodig zijn om dit goed te laten werken. Je zult ook nog de Unix user groep van de /srv/git directories naar www-data te zetten zodat je webserver lees- en schrijfrechten heeft op de repositories, omdat de Apache instantie die de CGI script draait (standaard) draait onder die user: $ chgrp -R www-data /srv/git Vervolgens moeten we een aantal dingen aan de Apache configuratie toevoegen om git-http-backend als afhandelaar te identificeren voor alles wat in het /git pad van je web server binnenkomt. SetEnv GIT_PROJECT_ROOT /opt/git SetEnv GIT_HTTP_EXPORT_ALL ScriptAlias /git/ /usr/libexec/git-core/git-http-backend/ Als je de GIT_HTTP_EXPORT_ALL uit je omgevingsvariabele laat, zal Git alleen de repositories met het git-daemon-export-ok bestand erin verspreiden, net zoals de Git daemon deed. Als laatste moet je Apache vertellen om verzoeken naar paden toe te staan die er zo uit zien, optioneel met een Auth block zoals hier: <Files "git-http-backend"> AuthType Basic AuthName "Git Access" AuthUserFile /srv/git/.htpasswd Require expr !(%{QUERY_STRING} -strmatch '*service=git-receive-pack*' || %{REQUEST_URI} =~ m#/git-receive-pack$#) Require valid-user </Files> Dat verplicht je een .htaccess bestand aan te maken met daarin de wachtwoorden van al de geldige gebruikers. Hier is een voorbeeld van hoe een “schacon” gebruiker toe te voegen aan het bestand: $ htpasswd -c /srv/git/.htpasswd schacon Er zijn tig manieren om geauthenticeerde gebruikers in Apache aan te geven, je zult een keuze moeten maken en een van deze implementeren. Dit is gewoon een van de eenvoudigste voorbeelden die we konden verzinnen. Je zult dit waarschijnlijk ook over SSL willen opzetten, zodat alle gegevens zijn versleuteld. We willen ons niet te veel bochten wringen om de specifieke zaken van Apache configuraties uit de doeken te doen, je gebruikt misschien een andere server of andere authenticatie behoeften hebben. De clou is dat Git met een CGI geleverd wordt die git-http-backend heet die, wanneer geactiveerd, alle onderhandelingen doet teneinde bestanden te sturen en te ontvangen over HTTP. Deze implementeert het authenticeren zelf niet, maar dat kan eenvoudig worden beheerd op het niveau van de server die dit aanroept. Je kunt dit met vrijwel elke web server met CGI capaciteiten, dus pas het vooral toe op hetgeen waar je het meest bekend mee bent. Noot Voor meer informatie over het configureren van authenticatie in Apache, verwijzen we je naar de volgende Apache documentatie: https://httpd.apache.org/docs/current/howto/auth.html prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://www.linkedin.com/uas/login?session_redirect=%2Fproducts%2Frd-station-rd-station-crm%3FviewConnections%3Dtrue&trk=products_details_guest_face-pile-cta
LinkedIn Login, Sign in | LinkedIn Sign in Sign in with Apple Sign in with a passkey By clicking Continue, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . or Email or phone Password Show Forgot password? Keep me logged in Sign in We’ve emailed a one-time link to your primary email address Click on the link to sign in instantly to your LinkedIn account. If you don’t see the email in your inbox, check your spam folder. Resend email Back New to LinkedIn? Join now Agree & Join LinkedIn By clicking Continue, you agree to LinkedIn’s User Agreement , Privacy Policy , and Cookie Policy . LinkedIn © 2026 User Agreement Privacy Policy Community Guidelines Cookie Policy Copyright Policy Send Feedback Language العربية (Arabic) বাংলা (Bangla) Čeština (Czech) Dansk (Danish) Deutsch (German) Ελληνικά (Greek) English (English) Español (Spanish) فارسی (Persian) Suomi (Finnish) Français (French) हिंदी (Hindi) Magyar (Hungarian) Bahasa Indonesia (Indonesian) Italiano (Italian) עברית (Hebrew) 日本語 (Japanese) 한국어 (Korean) मराठी (Marathi) Bahasa Malaysia (Malay) Nederlands (Dutch) Norsk (Norwegian) ਪੰਜਾਬੀ (Punjabi) Polski (Polish) Português (Portuguese) Română (Romanian) Русский (Russian) Svenska (Swedish) తెలుగు (Telugu) ภาษาไทย (Thai) Tagalog (Tagalog) Türkçe (Turkish) Українська (Ukrainian) Tiếng Việt (Vietnamese) 简体中文 (Chinese (Simplified)) 正體中文 (Chinese (Traditional))
2026-01-13T09:29:19
https://www.linkedin.com/products/netscout-arbor-threat-mitigation-system/?trk=products_details_guest_other_products_by_org_section_product_link_result-card_image-click
Arbor Threat Mitigation System (TMS) | LinkedIn Skip to main content LinkedIn NETSCOUT in Asan Expand search This button displays the currently selected search type. When expanded it provides a list of search options that will switch the search inputs to match the current selection. Jobs People Learning Clear text Clear text Clear text Clear text Clear text Join now Sign in Arbor Threat Mitigation System (TMS) DDoS Protection Software by NETSCOUT See who's skilled in this Add as skill Learn more Report this product About Internet Service Providers (ISPs), Cloud Providers and Enterprises face a common problem. Distributed Denial of Service (DDoS) attacks are a major risk to service availability. The power, sophistication and frequency of DDoS attacks continue to increase. Data center operators and network providers need a defense that is effective, cost-efficient and easily managed. Arbor Threat Mitigation System (TMS) is the acknowledged leader in DDoS protection. More Service Providers, Cloud Providers and large Enterprises use Arbor TMS for DDoS mitigation than any other solution. Working in conjunction with our Arbor Sightline solution, which provides visibility and threat detection, Arbor Threat Mitigation System (TMS) surgically removes DDoS attack traffic from your network without disrupting key network services. Media Products media viewer No more previous content Detecting and Mitigating a DDoS Attack Utilizing Arbor Sightline and TMS Learn how to detect and mitigate a DDoS Attack using NETSCOUT Arbor Sightline and TMS on a large enterprise or service provider network. Here, we demonstrate how DDoS Mitigation can be carried out by first detecting the attack and then automatically beginning mitigation using Arbor TMS. Multi-vector DDoS attacks are threats to all networks. In this example, NETSCOUT shows how Arbor Sightline and TMS can be used to detect and protect against an attack on a server farm. Arbor DDoS Protection for Carrier Service Providers - by NETSCOUT Arbor DDoS by NETSCOUT protection automates and orchestrates DDoS defense, using our unique Smart Data technology to extend visibility across any platform – physical, virtual or cloud - detecting a broader range of threats–before, during and after any attack. Advanced Detection and Mitigation Arbor Threat Mitigation System surgically removes up to 40Tbps of DDoS attack traffic, in a single deployment, while enabling the flow of legitimate traffic—without interrupting your network services. Arbor Threat Management System (TMS) Infrastructure No more next content Similar products Cloudflare DDoS Protection Cloudflare DDoS Protection DDoS Protection Software Cloudflare Spectrum Cloudflare Spectrum DDoS Protection Software Akamai Prolexic Routed Akamai Prolexic Routed DDoS Protection Software OVHcloud Anti-DDoS Protection OVHcloud Anti-DDoS Protection DDoS Protection Software Kona DDoS Defender Kona DDoS Defender DDoS Protection Software Kaspersky DDoS Protection Kaspersky DDoS Protection DDoS Protection Software Sign in to see more Show more Show less NETSCOUT products Arbor Edge Defense Arbor Edge Defense DDoS Protection Software Arbor Cloud DDoS Protection Arbor Cloud DDoS Protection DDoS Protection Software Arbor Sightline Arbor Sightline Network Monitoring Software InfiniStreamNG (ISNG) InfiniStreamNG (ISNG) Business Continuity Software nGenius Business Analytics nGenius Business Analytics Business Intelligence (BI) Software nGeniusONE nGeniusONE Application Performance Monitoring (APM) Software nGeniusPULSE nGeniusPULSE Network Management Software Omnis Threat Horizon Omnis Threat Horizon DDoS Protection Software Show more Show less LinkedIn © 2026 About Accessibility User Agreement Privacy Policy Cookie Policy Copyright Policy Brand Policy Guest Controls Community Guidelines English (English) Language
2026-01-13T09:29:19
https://git-scm.com/book/fa/v2/%d9%85%d9%82%d8%af%d9%85%d8%a7%d8%aa-%da%af%db%8c%d8%aa-git-basics-chapter-%d8%aa%da%af-%da%a9%d8%b1%d8%af%d9%86-Tagging
Git - تگ کردن (Tagging) About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. شروع به کار (getting started) 1.1 درباره ورژن کنترل (About Version Control) 1.2 تاریخچه کوتاهی از گیت (A Short History of Git) 1.3 گیت چیست؟ (What is Git) 1.4 نصب گیت (Installing Git) 1.5 ستاپ اولیه گیت (First-Time Git Setup) 1.6 دریافت کمک (Getting Help) 1.7 خلاصه (summary) 2. مقدمات گیت (git basics chapter) 2.1 گرفتن یک مخزن گیت (Getting a Git Repository) 2.2 ثبت تغییرات در مخزن (Recording Changes to the Repository) 2.3 مشاهده تاریخچه کامیت‌ها (Viewing the Commit History) 2.4 بازگرداندن تغییرات (Undoing Things) 2.5 کار کردن با ریموت ها (Working with Remotes) 2.6 تگ کردن (Tagging) 2.7 نام مستعار گیت (Git Aliases) 2.8 خلاصه (summary) 3. انشعاب‌گیری در گیت (Git Branching) 3.1 شاخه‌ها در یک نگاه (Branches in a Nutshell) 3.2 شاخه‌بندی و ادغام پایه‌ای (Basic Branching and Merging) 3.3 مدیریت شاخه‌ها (Branch Management) 3.4 روندهای کاری شاخه‌ها (Branching Workflows) 3.5 شاخه‌های راه دور (Remote Branches) 3.6 بازپایه‌گذاری (Rebasing) 3.7 خلاصه (Summary) 4. گیت روی سرور (Git on the server) 4.1 پروتکل‌ها (The Protocols) 4.2 راه‌اندازی گیت روی یک سرور (Getting Git on a Server) 4.3 ایجاد کلید عمومی SSH شما (Generating Your SSH Public Key) 4.4 نصب و راه‌اندازی سرور (Setting up server) 4.5 سرویس‌دهنده گیت (Git Daemon) 4.6 HTTP هوشمند (Smart HTTP) 4.7 گیت‌وب (GitWeb) 4.8 گیت‌لب (GitLab) 4.9 گزینه‌های میزبانی شخص ثالث (Third Party Hosted Options) 4.10 خلاصه (Summary) 5. گیت توزیع‌شده (Distributed git) 5.1 جریان‌های کاری توزیع‌شده (Distributed Workflows) 5.2 مشارکت در یک پروژه (Contributing to a Project) 5.3 نگهداری یک پروژه (Maintaining a Project) 5.4 خلاصه (Summary) 6. گیت هاب (GitHub) 6.1 راه‌اندازی و پیکربندی حساب کاربری (Account Setup and Configuration) 6.2 مشارکت در یک پروژه (Contributing to a Project) 6.3 نگهداری یک پروژه (Maintaining a Project) 6.4 مدیریت یک سازمان (Managing an organization) 6.5 اسکریپتنویسی در گیتهاب (Scripting GitHub) 6.6 خلاصه (Summary) 7. ابزارهای گیت (Git Tools) 7.1 انتخاب بازبینی (Revision Selection) 7.2 مرحله‌بندی تعاملی (Interactive Staging) 7.3 ذخیره موقت و پاک‌سازی (Stashing and Cleaning) 7.4 امضای کارهای شما (Signing Your Work) 7.5 جستجو (Searching) 7.6 بازنویسی تاریخچه (Rewriting History) 7.7 بازنشانی به زبان ساده (Reset Demystified) 7.8 ادغام پیشرفته (Advanced Merging) 7.9 بازاستفاده خودکار از حل تضادها (Rerere) 7.10 اشکال‌زدایی با گیت (Debugging with Git) 7.11 سابماژول ها (Submodules) 7.12 بسته‌بندی (Bundling) 7.13 جایگزینی (Replace) 7.14 ذخیره‌سازی اطلاعات ورود (Credential Storage) 7.15 خلاصه (Summary) 8. سفارشی‌سازی Git (Customizing Git) 8.1 پیکربندی گیت (Git Configuration) 8.2 ویژگی‌های گیت (Git Attributes) 8.3 هوک‌های گیت (Git Hooks) 8.4 یک نمونه سیاست اعمال شده توسط گیت (An Example Git-Enforced Policy) 8.5 خلاصه (Summary) 9. گیت و سیستم‌های دیگر (Git and Other Systems) 9.1 گیت به‌عنوان کلاینت (Git as a Client) 9.2 مهاجرت به گیت (Migrating to Git) 9.3 خلاصه (Summary) 10. مباحث درونی گیت (Git Internals) 10.1 ابزارها و دستورات سطح پایین (Plumbing and Porcelain) 10.2 اشیا گیت (Git Objects) 10.3 مراجع گیت (Git References) 10.4 فایل‌های بسته (Packfiles) 10.5 نگاشت (The Refspec) 10.6 پروتکل‌های انتقال (Transfer Protocols) 10.7 نگهداری و بازیابی داده‌ها (Maintenance and Data Recovery) 10.8 متغیرهای محیطی (Environment Variables) 10.9 خلاصه (Summary) A1. پیوست A: گیت در محیط‌های دیگر (Git in Other Environments) A1.1 رابط های گرافیکی (Graphical Interfaces) A1.2 گیت در ویژوال استودیو (Git in Visual Studio) A1.3 گیت در Visual Studio Code (Git in Visual Studio Code) A1.4 گیت در IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine (Git in IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine) A1.5 گیت در Sublime Text (Git in Sublime Text) A1.6 گیت در بش (Git in Bash) A1.7 گیت در Zsh (Git in Zsh) A1.8 گیت در PowerShell (Git in PowerShell) A1.9 خلاصه (Summary) A2. پیوست B: گنجاندن گیت در برنامه‌های شما (Embedding Git in your Applications) A2.1 خط فرمان گیت (Command-line Git) A2.2 کتابخانهٔ گیت به زبان سی (Libgit2) A2.3 کتابخانه گیت برای زبان جاوا (JGit) A2.4 کتابخانه گیت برای زبان گو (go-git) A2.5 کتابخانه گیت پایتون (Dulwich) A3. پیوست C: دستورات گیت (Git Commands) A3.1 تنظیم و پیکربندی (Setup and Config) A3.2 گرفتن و ایجاد پروژه‌ها (Getting and Creating Projects) A3.3 نمونه‌برداری پایه‌ای (Basic Snapshotting) A3.4 انشعاب‌گیری و ادغام (Branching and Merging) A3.5 به‌اشتراک‌گذاری و به‌روزرسانی پروژه‌ها (Sharing and Updating Projects) A3.6 بازرسی و مقایسه (Inspection and Comparison) A3.7 عیب‌یابی (Debugging) A3.8 اعمال تغییرات به صورت پچ (Patching) A3.9 ایمیل (Email) A3.10 سیستم‌های خارجی (External Systems) A3.11 مدیریت (Administration) A3.12 دستورات سطح پایین گیت (Plumbing Commands) 2nd Edition 2.6 مقدمات گیت (git basics chapter) - تگ کردن (Tagging) تگ کردن (Tagging) مانند اکثر VCS ها، گیت توانایی برچسب گذاری نقاط خاص در تاریخچه یک مخزن را به عنوان مهم دارد. به طور معمول، مردم از این قابلیت برای مشخص کردن نقاط انتشار ( v1.0 ، v2.0 و غیره) استفاده می کنند. در این بخش، شما یاد خواهید گرفت که چگونه برچسب های موجود را لیست کنید، چگونه برچسب ها را ایجاد و حذف کنید، و انواع مختلف برچسب ها چیست. لیست کردن تگ ها (Listing Your Tags) فهرست کردن تگ های موجود در گیت ساده است. فقط تایپ کنید git tag (با اختیاری -l یا --list ): $ git tag v1.0 v2.0 این دستور برچسب ها را به ترتیب الفبا لیست می کند؛ ترتیب نمایش آنها اهمیت واقعی ندارد. همچنین می توانید برچسب هایی را که با یک الگوی خاص مطابقت دارند، جستجو کنید. به عنوان مثال، منبع گیت شامل بیش از ۵۰۰ تگ است. اگر شما فقط علاقه مند به نگاه کردن به سری 1.8.5 هستید، می توانید این را اجرا کنید: $ git tag -l "v1.8.5*" v1.8.5 v1.8.5-rc0 v1.8.5-rc1 v1.8.5-rc2 v1.8.5-rc3 v1.8.5.1 v1.8.5.2 v1.8.5.3 v1.8.5.4 v1.8.5.5 یادداشت Listing tag wildcards requires -l or --list option اگر شما فقط کل لیست برچسب ها را می خواهید، اجرای دستور git tag به طور ضمنی فرض می کند که شما یک لیست می خواهید و یک لیست را ارائه می دهد؛ استفاده از -l یا --list در این مورد اختیاری است. با این حال، اگر شما در حال ارائه یک الگوی کارد جوینده برای مطابقت با نام های برچسب هستید، استفاده از -l یا --list اجباری است. ساخت تگ (Creating Tags) گیت از دو نوع تگ پشتیبانی می کند: lightweight و annotated . یک تگ سبک وزن بسیار شبیه به یک شاخه است که تغییر نمی کند — فقط یک اشاره کننده به یک commit خاص است. با این حال، برچسب های تشریح شده به عنوان اشیاء کامل در پایگاه داده گیت ذخیره می شوند. آنها چک جمع شده اند؛ شامل نام، ایمیل و تاریخ برچسب دار هستند؛ یک پیام برچسب دار دارند؛ و می توانند با GNU Privacy Guard (GPG) امضا و تأیید شوند. به طور کلی توصیه می شود که شما برچسب های تفسیر شده ایجاد کنید تا بتوانید تمام این اطلاعات را داشته باشید؛ اما اگر شما یک برچسب موقت می خواهید یا به دلایلی نمی خواهید اطلاعات دیگر را نگه دارید، برچسب های سبک وزن نیز در دسترس هستند. تگ‌های حاشیه‌گذاری‌شده (Annotated Tags) ایجاد یک برچسب با یادداشت در گیت ساده است. ساده ترین راه این است که -a را مشخص کنید وقتی دستور tag را اجرا می کنید: $ git tag -a v1.4 -m "my version 1.4" $ git tag v0.1 v1.3 v1.4 -m یک پیام برچسب گذاری را مشخص می کند که با برچسب ذخیره می شود. اگر شما یک پیام برای یک تگ با اشاره مشخص نکنید، Git ویرایشگر شما را راه اندازی می کند تا شما بتوانید آن را تایپ کنید. شما می توانید داده های تگ را همراه با commit که با استفاده از دستور `git show ` تگ شده است ببینید: $ git show v1.4 tag v1.4 Tagger: Ben Straub <ben@straub.cc> Date: Sat May 3 20:19:12 2014 -0700 my version 1.4 commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number این اطلاعات برچسب دار را نشان می دهد، تاریخ برچسب گذاری commit، و پیام تبصری قبل از نمایش اطلاعات commit. تگ‌های سبک (Lightweight Tags) راه دیگری برای برچسب گذاری commit ها با یک برچسب سبک وزن است. این اساساً چک سوم تاییدیه ای است که در یک فایل ذخیره شده است — هیچ اطلاعات دیگری حفظ نمی شود. برای ایجاد یک تگ سبک، هیچ یک از گزینه های -a ، -s یا -m را ارائه ندهید، فقط یک نام تگ را ارائه دهید: $ git tag v1.4-lw $ git tag v0.1 v1.3 v1.4 v1.4-lw v1.5 این بار، اگر شما git show را روی برچسب اجرا کنید، اطلاعات برچسب اضافی را نمی بینید. دستور فقط کامیت رو نشون میده: $ git show v1.4-lw commit ca82a6dff817ec66f44342007202690a93763949 Author: Scott Chacon <schacon@gee-mail.com> Date: Mon Mar 17 21:52:11 2008 -0700 Change version number برچسب‌گذاری بعدی (Tagging Later) شما همچنین می توانید بعد از اینکه از آنها گذشته اید، آنها را برچسب بزنید. فرض کنید که سابقه ی commit شما اینگونه باشد: $ git log --pretty=oneline 15027957951b64cf874c3557a0f3547bd83b3ff6 Merge branch 'experiment' a6b4c97498bd301d84096da251c98a07c7723e65 Create write support 0d52aaab4479697da7686c15f77a3d64d9165190 One more thing 6d52a271eda8725415634dd79daabbc4d9b6008e Merge branch 'experiment' 0b7434d86859cc7b8c3d5e1dddfed66ff742fcbc Add commit function 4682c3261057305bdd616e23b64b0857d832627b Add todo file 166ae0c4d3f420721acbb115cc33848dfcc2121a Create write support 9fceb02d0ae598e95dc970b74767f19372d61af8 Update rakefile 964f16d36dfccde844893cac5b347e7b3d44abbc Commit the todo 8a5cbc430f1a9c3d00faaeffd07798508422908a Update readme حالا فرض کنید فراموش کرده اید که پروژه را در نسخه 1.2 برچسب بزنید، که در "Update rakefile" commit بود. می تونی بعد از واقعه اضافهش کنی برای برچسب زدن به این کامیت، شما چک سوم کامیت (یا بخشی از آن) را در انتهای دستور مشخص می کنید: $ git tag -a v1.2 9fceb02 شما می توانید ببینید که شما تاگ commit: $ git tag v0.1 v1.2 v1.3 v1.4 v1.4-lw v1.5 $ git show v1.2 tag v1.2 Tagger: Scott Chacon <schacon@gee-mail.com> Date: Mon Feb 9 15:32:16 2009 -0800 version 1.2 commit 9fceb02d0ae598e95dc970b74767f19372d61af8 Author: Magnus Chacon <mchacon@gee-mail.com> Date: Sun Apr 27 20:43:35 2008 -0700 Update rakefile ... اشتراک گذاری تگ ها (Sharing Tags) به طور پیش فرض، دستور git push برچسب ها را به سرورهای از راه دور منتقل نمی کند. شما باید به طور صریح تگ ها را به یک سرور مشترک پس از ایجاد آنها ارسال کنید. این فرآیند درست مثل اشتراک گذاری شاخه های از راه دور است — شما می توانید git push origin <tagname> را اجرا کنید. $ git push origin v1.5 Counting objects: 14, done. Delta compression using up to 8 threads. Compressing objects: 100% (12/12), done. Writing objects: 100% (14/14), 2.05 KiB | 0 bytes/s, done. Total 14 (delta 3), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.5 -> v1.5 اگر تعداد زیادی tag دارید که می‌خواهید همزمان ارسال کنید، می‌توانید از گزینه --tags در دستور git push استفاده کنید. این کار تمام tag هایی را که هنوز در سرور راه‌دور وجود ندارند، منتقل می‌کند. $ git push origin --tags Counting objects: 1, done. Writing objects: 100% (1/1), 160 bytes | 0 bytes/s, done. Total 1 (delta 0), reused 0 (delta 0) To git@github.com:schacon/simplegit.git * [new tag] v1.4 -> v1.4 * [new tag] v1.4-lw -> v1.4-lw حالا، وقتی کسی از مخزن شما کلان یا استخراج می کند، همه ی برچسب های شما را هم می گیرد. یادداشت git push pushes both types of tags دستور git push <remote> --tags هر دو نوع lightweight و annotated tag را ارسال می‌کند. در حال حاضر گزینه‌ای برای ارسال فقط lightweight tags وجود ندارد، اما اگر از git push <remote> --follow-tags استفاده کنید، تنها annotated tags به سرور راه‌دور ارسال خواهند شد. حذف تگ ها (Deleting Tags) برای حذف یک برچسب در مخزن محلی خود، می توانید از git tag -d <tagname> استفاده کنید. به عنوان مثال، می توانیم برچسب سبک وزن بالا را به صورت زیر حذف کنیم: $ git tag -d v1.4-lw Deleted tag 'v1.4-lw' (was e7d5add) توجه داشته باشید که این کار برچسب را از هر سرور از راه دور حذف نمی کند. دو نوع رایج برای حذف یک برچسب از یک سرور از راه دور وجود دارد. اولین تغییر `git push <remote> :refs/tags/<tagname> ` است: $ git push origin :refs/tags/v1.4-lw To /git@github.com:schacon/simplegit.git - [deleted] v1.4-lw روش تفسیر این مطلب این است که آن را به عنوان مقدار صفر قبل از فشار دادن دو نقطه به نام تگ از راه دور بخوانیم، و به طور موثر آن را حذف کنیم. راه دوم (و بصری تر) برای حذف یک برچسب از راه دور این است: $ git push origin --delete <tagname> چک کردن تگ ها (Checking out Tags) اگر می خواهید نسخه های فایل هایی که یک برچسب به آنها اشاره دارد را مشاهده کنید، می توانید یک چکآوت `git از آن برچسب را انجام دهید، اگرچه این مخزن شما را در حالت “detached HEAD” قرار می دهد، که دارای برخی از عوارض جانبی بد است: $ git checkout v2.0.0 Note: switching to 'v2.0.0'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 99ada87... Merge pull request #89 from schacon/appendix-final $ git checkout v2.0-beta-0.1 Previous HEAD position was 99ada87... Merge pull request #89 from schacon/appendix-final HEAD is now at df3f601... Add atlas.json and cover image در حالت detached HEAD ، اگر تغییراتی ایجاد کنید و سپس یک commit بسازید، tag بدون تغییر می‌ماند، اما commit جدید به هیچ شاخه‌ای تعلق نخواهد داشت و تنها از طریق hash دقیق commit قابل دسترسی است. بنابراین، اگر نیاز به اعمال تغییرات دارید — برای مثال اصلاح یک باگ در نسخه‌ای قدیمی — معمولاً بهتر است یک branch ایجاد کنید: $ git checkout -b version2 v2.0.0 Switched to a new branch 'version2' اگر این کار را انجام دهید و یک commit را انجام دهید، شاخه version2 شما کمی متفاوت از برچسب v2.0.0 خواهد بود، زیرا با تغییرات جدید شما پیش خواهد رفت، بنابراین مراقب باشید. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19
https://git-scm.com/book/sl/v2/Osnove-Git-Pridobivanje-repozitorija-Git
Git - Pridobivanje repozitorija Git About Trademark Learn Book Cheat Sheet Videos External Links Tools Command Line GUIs Hosting Reference Install Community This book is available in English . Full translation available in azərbaycan dili , български език , Deutsch , Español , فارسی , Français , Ελληνικά , 日本語 , 한국어 , Nederlands , Русский , Slovenščina , Tagalog , Українська , 简体中文 , Partial translations available in Čeština , Македонски , Polski , Српски , Ўзбекча , 繁體中文 , Translations started for Беларуская , Indonesian , Italiano , Bahasa Melayu , Português (Brasil) , Português (Portugal) , Svenska , Türkçe . The source of this book is hosted on GitHub. Patches, suggestions and comments are welcome. Chapters ▾ 1. Začetek 1.1 O nadzoru različic 1.2 Kratka zgodovina Gita 1.3 Kaj je Git? 1.4 Ukazna vrstica 1.5 Namestitev Gita 1.6 Prva nastavitev Gita 1.7 Pridobivanje pomoči 1.8 Povzetek 2. Osnove Git 2.1 Pridobivanje repozitorija Git 2.2 Snemanje sprememb v repozitorij 2.3 Pregled zgodovine potrditev 2.4 Razveljavljanje stvari 2.5 Delo z daljavami 2.6 Označevanje 2.7 Aliasi Git 2.8 Povzetek 3. Veje Git 3.1 Veje na kratko 3.2 Osnove vej in združevanja 3.3 Upravljanje vej 3.4 Poteki dela z vejami 3.5 Oddaljene veje 3.6 Ponovno baziranje 3.7 Povzetek 4. Git na strežniku 4.1 Protokoli 4.2 Pridobitev Gita na strežniku 4.3 Generiranje vaših javnih ključev SSH 4.4 Nastavitev strežnika 4.5 Prikriti proces Git 4.6 Pametni HTTP 4.7 GitWeb 4.8 GitLab 4.9 Možnosti gostovanja pri tretjih ponudnikih 4.10 Povzetek 5. Porazdeljeni Git 5.1 Porazdeljeni poteki dela 5.2 Prispevek k projektu 5.3 Vzdrževanje projekta 5.4 Povzetek 6. GitHub 6.1 Namestitev in konfiguracija računa 6.2 Prispevek k projektu 6.3 Vzdrževanje projekta 6.4 Upravljanje organizacije 6.5 Skriptni GitHub 6.6 Povzetek 7. Orodja Git 7.1 Izbira revizije 7.2 Interaktivno pripravljanje 7.3 Shranjevanje na varno (angl. stashing) in čiščenje 7.4 Podpisovanje vašega dela 7.5 Iskanje 7.6 Prepisovanje zgodovine 7.7 Demistifikacija ponastavitve 7.8 Napredno združevanje 7.9 Rerere 7.10 Razhroščevanje z Gitom 7.11 Podmoduli 7.12 Povezovanje v pakete 7.13 Zamenjava 7.14 Shramba poverilnic 7.15 Povzetek 8. Prilagoditev Gita 8.1 Konfiguracija Git 8.2 Atributi Git 8.3 Kljuke Git 8.4 Primer pravilnika, ki ga uveljavlja Git 8.5 Povzetek 9. Git in ostali sistemi 9.1 Git kot odjemalec 9.2 Migracija na Git 9.3 Povzetek 10. Notranjost Gita 10.1 Napeljava in keramika 10.2 Objekti Git 10.3 Reference Git 10.4 Packfiles (datoteke zmanjšanih podatkov) 10.5 Refspec 10.6 Protokoli prenosa 10.7 Vzdrževanje in obnovitev podatkov 10.8 Spremenljivke okolja 10.9 Povzetek A1. Dodatek A: Git v drugih okoljih A1.1 Grafični vmesniki A1.2 Git v Visual Studio A1.3 Git v Visual Studio Code A1.4 Git v IntelliJ / PyCharm / WebStorm / PhpStorm / RubyMine A1.5 Git v Sublime Text A1.6 Git v Bashu A1.7 Git v Zsh A1.8 Git v Powershellu A1.9 Povzetek A2. Dodatek B: Vdelava Gita v vašo aplikacijo A2.1 Git v ukazni vrstici A2.2 Libgit2 A2.3 JGit A2.4 go-git A2.5 Dulwich A3. Dodatek C: Ukazi Git A3.1 Nastavitev in konfiguracija A3.2 Pridobivanje in ustvarjanje projektov A3.3 Osnove posnetkov A3.4 Veje in združevanje A3.5 Deljenje in posodabljanje projektov A3.6 Pregled in primerjava A3.7 Razhroščevanje A3.8 Popravljanje A3.9 E-pošta A3.10 Zunanji sistemi A3.11 Administracija A3.12 Orodja za sisteme napeljave 2nd Edition 2.1 Osnove Git - Pridobivanje repozitorija Git Če lahko preberete samo eno poglavje, da začnete z Gitom, je to to. To poglavje pokriva vsak osnovni ukaz, ki ga morate izvesti za glavnino stvari, za katere boste eventualno porabili čas pri opravljanju z Gitom. Do konca tega poglavja bi morali znati nastaviti in inicializirati repozitorij, začeti in ustaviti sledenje datotek, jih dati v področje priprave in potrditi spremembe. Pokazali vam bomo tudi, kako nastaviti Git, da ignorira določene datoteke in vzorce datotek, kako hitro in enostavno razveljaviti napake, kako brskati po zgodovini vašega projekta in pogledati spremembe med potrditvami ter kako potisniti in povleči iz oddaljenega repozitorija. Pridobivanje repozitorija Git Projekt Git običajno pridobite z dvema pristopoma: Vzamete lahko lokalni direktorij, ki trenutno ni pod nadzorom različic, in ga spremenite v repozitorij Git, ali pa Klonirate obstoječi repozitorij Git od drugod. V kateremkoli od obeh primerov dobite na vaši lokalni napravi repozitorij Git, ki je pripravljen za delo. Inicializacija repozitorija v obstoječem direktoriju Če imate projektni direktorij, ki trenutno ni pod nadzorom različic, in ga želite začeti nadzirati z Gitom, morate najprej iti v ta projektni direktorij. Če tega še niste počeli, je ukaz odvisen od sistema, ki ga poganjate: za Linux: $ cd /home/user/my_project za macOS: $ cd /Users/user/my_project za Windows: $ cd C:/Users/user/my_project in vpišete: $ git init To ustvari nov poddirektorij z imenom .git , ki vsebuje vse vaše potrebne datoteke repozitorija — skelet repozitorija Git. Na tej točki ni v vašem projektu še nič sledeno. Glejte poglavje Notranjost Gita za več informacij o tem, katere datoteke točno so vsebovane v direktoriju .git , ki ste ga ravno ustvarili. Če želite začeti nadzor različic obstoječih datotek (v primerjavi s praznim direktorijem), bi morali verjetno začeti slediti tem datotekam in narediti začetno potrditev. To lahko naredite z nekaj ukazi git add , ki določi datoteke, ki jim želite slediti, ter nato izvedete git commit : $ git add *.c $ git add LICENSE $ git commit -m 'Initial project version' Kaj te ukazi naredijo, bomo šli skozi čez minuto. Na tej točki imate repozitorij Git s sledenimi datotekami in začetno potrditvijo. Kloniranje obstoječega repozitorija Če želite dobiti kopijo obstoječega repozitorija Git, na primer projekt, ki mu želite prispevati, je ukaz, ki ga potrebujete, git clone . Če ste že seznanjeni z ostalimi sistemi VCS, kot je Subversion, boste opazili, da je ukaz clone in ne checkout . To je pomembna razlika — namesto da dobite samo delovno kopijo, Git dobi polno kopijo skoraj vseh podatkov, ki jih ima strežnik. Vsaka različica vsake datoteke zgodovine projekta je privzeto povlečena, ko poženete git clone . V bistvu, če se disk vašega strežnika pokvari, lahko pogosto uporabite skoraj katerikoli klon kateregakoli odjemalca, da strežnik povrnete v stanje, v katerem je bil, ko je bil kloniran (morda boste izgubili nekatere kljuke strežniške strani in podobno, vendar vsi podatki v različicah bi morali bili tam — za več podrobnosti si oglejte razdelek Pridobitev Gita na strežniku ). Repozitorij klonirate z git clone <url> . Na primer, če želite klonirati Git povezano knjižnico imenovano libgit2 , lahko to naredite na naslednji način: $ git clone https://github.com/libgit2/libgit2 To ustvari direktorij imenovan libgit2 , znotraj njega inicializira direktorij .git , povleče vse podatke za ta repozitorij in izvleče delovno kopijo zadnje različice. Če greste v novi direktorij libgit2 , boste tam videli projektne datoteke, ki so pripravljene za delo ali uporabo. Če želite klonirati repozitorij v direktorij, imenovan nekaj drugega kot libgit2 , lahko to določite kot naslednji argument v ukazni vrstici: $ git clone https://github.com/libgit2/libgit2 mylibgit Ukaz izvede enako stvar kot prejšnji, vendar ciljni direktorij je imenovan mylibgit . Git ima vrsto različnih prenosnih protokolov, ki jih lahko uporabite. Prejšnji primer uporablja protokol https:// , vendar videli boste lahko tudi git:// ali user@server:path/to/repo.git , ki uporablja prenosni protokol SSH. Razdelek Pridobitev Gita na strežniku vam bo predstavil vse možnosti, ki so na voljo, da jih strežnik lahko nastavi za dostopanje do vašega repozitorija Git ter prednosti in slabosti vsake. prev | next About this site Patches, suggestions, and comments are welcome. Git is a member of Software Freedom Conservancy
2026-01-13T09:29:19