hexsha
stringlengths
40
40
size
int64
5
1.04M
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
344
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
listlengths
1
11
max_stars_count
int64
1
368k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
344
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
listlengths
1
11
max_issues_count
int64
1
116k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
344
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
listlengths
1
11
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
1.04M
avg_line_length
float64
1.14
851k
max_line_length
int64
1
1.03M
alphanum_fraction
float64
0
1
lid
stringclasses
191 values
lid_prob
float64
0.01
1
912784033410cbea044d74258aefe4246b6b2685
886
md
Markdown
_posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md
tatudoquei/tatudoquei.github.io
a3a3c362424fda626d7d0ce2d9f4bead6580631c
[ "MIT" ]
null
null
null
_posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md
tatudoquei/tatudoquei.github.io
a3a3c362424fda626d7d0ce2d9f4bead6580631c
[ "MIT" ]
null
null
null
_posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md
tatudoquei/tatudoquei.github.io
a3a3c362424fda626d7d0ce2d9f4bead6580631c
[ "MIT" ]
1
2022-01-13T07:57:24.000Z
2022-01-13T07:57:24.000Z
--- layout: post item_id: 3351158114 title: >- Após 19 dias, Salles entrega celular a investigadores da operação Akuanduba author: Tatu D'Oquei date: 2021-06-07 18:45:32 pub_date: 2021-06-07 18:45:32 time_added: 2021-06-07 19:04:57 category: tags: [] image: https://www.cartacapital.com.br/wp-content/uploads/2021/05/ricardosalles.jpg --- O ministro do Meio Ambiente, Ricardo Salles, informou ao Supremo Tribunal Federal, nesta segunda-feira 7, que entregou seu celular aos os investigadores da Polícia Federal, no âmbito da Operação Akuanduba, que investiga sua possível atuação em prol de empresas na exportação ilegal de madeir **Link:** [https://www.cartacapital.com.br/politica/apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba/](https://www.cartacapital.com.br/politica/apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba/)
46.631579
291
0.784424
por_Latn
0.82637
9127a5ff9bdafd7c26d128524ed18eb560875fba
673
md
Markdown
sandstone/pattern-virtualgridlist-api/README.md
enyojs/enact-samples
e360b50eeed552613f8dc4f5b1d18d299ee5a7dd
[ "Apache-2.0" ]
null
null
null
sandstone/pattern-virtualgridlist-api/README.md
enyojs/enact-samples
e360b50eeed552613f8dc4f5b1d18d299ee5a7dd
[ "Apache-2.0" ]
null
null
null
sandstone/pattern-virtualgridlist-api/README.md
enyojs/enact-samples
e360b50eeed552613f8dc4f5b1d18d299ee5a7dd
[ "Apache-2.0" ]
null
null
null
## VirtualGridList add/remove/select/deselect pattern // My Gallery A sample Enact application that shows off how to add/remove/select/deselect items of VirtualGridList Run `npm install` then `npm run serve` to have the app running on [http://localhost:8080](http://localhost:8080), where you can view it in your browser. #### Enact Components Used - `sandstone/Button` - `sandstone/ImageItem` - `sandstone/Panels/Header` - `sandstone/VirtualList/VirtualGridList` In this app, you can add, remove, select, deselect, selectAll, and deselectAll items using redux. --- This project was bootstrapped with the Enact [cli](https://github.com/enactjs/cli).
37.388889
153
0.744428
eng_Latn
0.956692
912862e895a04c35a9cd6b8534c8800642e505ad
2,428
md
Markdown
README.md
es-shims/Promise.try
23546f147e25515660c9b20f346be3dca513dd74
[ "MIT" ]
5
2016-08-21T12:18:28.000Z
2019-10-27T22:28:44.000Z
README.md
es-shims/Promise.try
23546f147e25515660c9b20f346be3dca513dd74
[ "MIT" ]
2
2019-11-11T07:57:12.000Z
2019-11-19T09:48:45.000Z
README.md
es-shims/Promise.try
23546f147e25515660c9b20f346be3dca513dd74
[ "MIT" ]
2
2016-08-21T09:35:10.000Z
2018-01-17T09:33:54.000Z
# promise.try <sup>[![Version Badge][npm-version-svg]][package-url]</sup> [![Build Status][travis-svg]][travis-url] [![dependency status][deps-svg]][deps-url] [![dev dependency status][dev-deps-svg]][dev-deps-url] [![License][license-image]][license-url] [![Downloads][downloads-image]][downloads-url] [![npm badge][npm-badge-png]][package-url] [![browser support][testling-svg]][testling-url] ES Proposal spec-compliant shim for `Promise.try`. Invoke its "shim" method to shim `Promise.try` if it is unavailable or noncompliant. **Note**: a global `Promise` must already exist: the [es6-shim](https://github.com/es-shims/es6-shim) is recommended. This package implements the [es-shim API](https://github.com/es-shims/api) interface. It works in an ES3-supported environment that has `Promise` available globally, and complies with the [proposed spec](https://github.com/ljharb/proposal-promise-try). Most common usage: ```js var assert = require('assert'); var promiseTry = require('promise.try'); promiseTry(function () { throw 42; }).catch(function (e) { assert.equal(e, 42); }); promiseTry(function () { return Infinity; }).then(function (x) { assert.equal(x, Infinity); }); promiseTry.shim(); // will be a no-op if not needed Promise.try(function () { throw 42; }).catch(function (e) { assert.equal(e, 42); }); Promise.try(function () { return Infinity; }).then(function (x) { assert.equal(x, Infinity); }); ``` ## Tests Simply clone the repo, `npm install`, and run `npm test` [package-url]: https://npmjs.com/package/promise.try [npm-version-svg]: http://versionbadg.es/es-shims/Promise.try.svg [travis-svg]: https://travis-ci.org/es-shims/Promise.try.svg [travis-url]: https://travis-ci.org/es-shims/Promise.try [deps-svg]: https://david-dm.org/es-shims/Promise.try.svg [deps-url]: https://david-dm.org/es-shims/Promise.try [dev-deps-svg]: https://david-dm.org/es-shims/Promise.try/dev-status.svg [dev-deps-url]: https://david-dm.org/es-shims/Promise.try#info=devDependencies [testling-svg]: https://ci.testling.com/es-shims/Promise.try.png [testling-url]: https://ci.testling.com/es-shims/Promise.try [npm-badge-png]: https://nodei.co/npm/promise.try.png?downloads=true&stars=true [license-image]: http://img.shields.io/npm/l/promise.try.svg [license-url]: LICENSE [downloads-image]: http://img.shields.io/npm/dm/promise.try.svg [downloads-url]: http://npm-stat.com/charts.html?package=promise.try
36.238806
253
0.719522
kor_Hang
0.266022
9128a347536e4cd1119bce7afa140e842578a0cb
142
md
Markdown
content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md
technology-toolbox/website
9d845dc68e650ee164959da418fde24eacecf1c9
[ "MIT" ]
null
null
null
content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md
technology-toolbox/website
9d845dc68e650ee164959da418fde24eacecf1c9
[ "MIT" ]
109
2021-03-25T11:16:17.000Z
2022-01-23T20:55:51.000Z
content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md
technology-toolbox/website
9d845dc68e650ee164959da418fde24eacecf1c9
[ "MIT" ]
null
null
null
--- layout: popout title: Table 1 - MOSS 2007 Feature Definitions date: 2008-04-08T18:39:00-06:00 --- {{< include-html "../table-1.html" >}}
17.75
46
0.65493
kor_Hang
0.46492
912951379577e4f7b23fcae87b7916c8adf06b40
349
md
Markdown
README.md
Ccode-lang/pydrumplugin
ab65688288a6f2f464fe1ca69560c975a750f194
[ "MIT" ]
null
null
null
README.md
Ccode-lang/pydrumplugin
ab65688288a6f2f464fe1ca69560c975a750f194
[ "MIT" ]
null
null
null
README.md
Ccode-lang/pydrumplugin
ab65688288a6f2f464fe1ca69560c975a750f194
[ "MIT" ]
null
null
null
# pydrumplugin A template for a drumbash plugin in python # setup 1. Change filler in `build.sh` to whatever you want to name the plugin 2. Write code using the imported api # building Run: ```bash ./build.sh ``` The outputed drumfile will be in `./artifact`. # drum api look at how to use at https://github.com/Ccode-lang/drumbash/wiki/py-api-docs
24.928571
77
0.74212
eng_Latn
0.978943
912958e5c9c66c83392f1299f852e467080be1d4
738
md
Markdown
_posts/2020-01-21-cartagena-taller3.md
EducacionSiglo21/jekyll
f04b07efe35a5c2c61b290592e3b9b63d2e16552
[ "MIT" ]
3
2018-11-26T15:11:12.000Z
2019-02-12T06:51:17.000Z
_posts/2020-01-21-cartagena-taller3.md
EducacionSiglo21/startbootstrap-clean-blog-jekyll
db55bc500681697614b32c530ae29bdf58305b72
[ "MIT" ]
null
null
null
_posts/2020-01-21-cartagena-taller3.md
EducacionSiglo21/startbootstrap-clean-blog-jekyll
db55bc500681697614b32c530ae29bdf58305b72
[ "MIT" ]
3
2018-12-16T10:55:27.000Z
2021-01-22T21:19:25.000Z
--- layout: post title: "Realidad virtual y realidad aumentada" subtitle: "Taller" background: "/img/posts/bg-cartagena.jpg" eventdate: 2020-02-20 10:00:00 +0100 category: "local" tags: "cartagena" placeName: "CEIP Bethoven, Cartagena" speakers: - name: Paqui Rosique --- Descripción taller: La realidad virtual (VR) se coló en la vida de más de uno desde el mundo de los videojuegos. No obstante, sus aplicaciones van ya mucho más allá del propio entretenimiento y son algo más que útiles para distintos sectores en el día a día de cualquier persona o profesional. El taller de Realidad Virtual y Realidad Aumentada per - mitirá a los participantes sentir la inmersión de la Realidad Virtual a través de las gafas Oculus Ri
33.545455
117
0.768293
spa_Latn
0.993285
912a636db978d531b25d027896114e19feb331b1
227
md
Markdown
README.md
ShaneMcC/DMDirc-Util
ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b
[ "MIT" ]
null
null
null
README.md
ShaneMcC/DMDirc-Util
ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b
[ "MIT" ]
6
2015-01-17T21:58:27.000Z
2017-01-15T05:09:08.000Z
README.md
ShaneMcC/DMDirc-Util
ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b
[ "MIT" ]
2
2019-05-02T22:30:56.000Z
2019-05-08T05:51:13.000Z
# Util [![Codacy Badge](https://api.codacy.com/project/badge/Grade/8444740c995c403a9d559ffb0e3764f9)](https://www.codacy.com/app/DMDirc/Util?utm_source=github.com&utm_medium=referral&utm_content=DMDirc/Util&utm_campaign=badger)
113.5
220
0.823789
yue_Hant
0.886751
912b0439871eb5c3da9cb4c27038da1001ee31e6
204
md
Markdown
README.md
Roragok/namafia-anime
6d804163152bd002f198034734b84704356eb199
[ "MIT" ]
null
null
null
README.md
Roragok/namafia-anime
6d804163152bd002f198034734b84704356eb199
[ "MIT" ]
15
2019-12-17T16:49:32.000Z
2022-02-18T16:48:13.000Z
README.md
Roragok/namafia-anime
6d804163152bd002f198034734b84704356eb199
[ "MIT" ]
null
null
null
# namafia-anime Webpage for anime streamings To run this yourself install node and yarn, clone the repo and run yarn start. Useful Links: - https://www.typescriptlang.org/docs/handbook/basic-types.html
25.5
78
0.789216
eng_Latn
0.52797
912b1587f74aa7894fca0a1d7e118eab34590f4e
16,211
md
Markdown
_episodes/05-loop.md
statkclee/shell-novice-kr
c8c59be87a1c1f4e9310baca5f209eaac6801d15
[ "CC-BY-4.0" ]
null
null
null
_episodes/05-loop.md
statkclee/shell-novice-kr
c8c59be87a1c1f4e9310baca5f209eaac6801d15
[ "CC-BY-4.0" ]
null
null
null
_episodes/05-loop.md
statkclee/shell-novice-kr
c8c59be87a1c1f4e9310baca5f209eaac6801d15
[ "CC-BY-4.0" ]
null
null
null
--- title: "루프(Loops)" teaching: 40 exercises: 10 questions: - "다른 파일이 많은데 어떻게 동일한 동작을 수행시킬 수 있을까?" objectives: - "파일 집합의 각 파일에 따로 따로 나누어서 하나 혹은 명령어 다수를 적용하는 루프를 작성한다." - "루프가 실행되는 동안에 루프 변수가 취하는 값을 추적한다." - "변수명과 변수값 차이에 대해 설명한다." - "왜 공백과 일부 구두점 문자는 파일 이름에 사용되지 말아야 되는지 설명한다." - "어떤 명령어가 최근에 실행되었는지를 확인하는 방법을 시범으로 보여준다." - "명령어를 다시 타이핑하지 않고 최근에 실행된 명령어를 다시 실행한다." keypoints: - "`for` 루프는 리스트의 모든 원소에 대해서 명령어를 한번씩 모두 반복한다." - "모든 `for` 루프는 변수를 사용해서 현재 연산작업하는 것을 추적한다." - "`$name`을 사용해서 변수(즉, 변수값을 얻는데)를 확장한다. `${name}`도 사용될 수 있다." - "파일명에 공백, 인용부호, '*' 혹은 '?'와 같은 와일드카드 문자를 사용하지 않는다. 왜냐햐면, 변수 확장을 난해하게 된다." - "파일에 일관된 명칭을 부여해서 와일드카드 패턴으로 매칭되기 쉽게 하고 루프를 돌릴 때 선택도 쉽게 만든다." - "이전 명령어를 편집하고 반복실행하는데 키보드 윗방향 화살표를 사용한다." - "`Ctrl-R` 명령어를 사용해서 이전에 입력한 명령어를 검색한다." - "`history` 명령어를 사용해서 가장 최근 명령어를 화면에 출력하고, `!number`을 사용해서 해당 숫자 명령어를 반복실행한다." --- 반복적으로 명령어를 실행하게 함으로써 자동화를 통해서 **루프**는 생산성 향상에 핵심이 된다. 와일드카드와 탭 자동완성과 유사하게, 루프를 사용하면 타이핑 상당량(타이핑 실수)을 줄일 수 있다. 와일드카드와 탭 자동완성은 타이핑을 (타이핑 실수를) 줄이는 두가지 방법이다. 또다른 것은 쉘이 반복해서 특정 작업을 수행하게 하는 것이다. `basilisk.dat`, `unicorn.dat` 등으로 이름 붙여진 게놈 데이터 파일이 수백개 있다고 가정하자. 이번 예제에서, 단지 두개 예제 파일만 있는 `creatures` 디렉토리를 사용할 것이지만 동일한 원칙은 훨씬 더 많은 파일에 즉시 적용될 수 있다. 디렉토리에 있는 파일을 변경하고 싶지만, 원본 파일을 `original-basilisk.dat`와 `original-unicorn.dat`으로 이름을 변경해서 저장한다. 하지만 다음 명령어를 사용할 수 없다: ~~~ $ cp *.dat original-*.dat ~~~ {: .language-bash} 왜냐하면 상기 두 파일 경우에 전개가 다음과 같이 될 것이기 때문이다: ~~~ $ cp basilisk.dat unicorn.dat original-*.dat ~~~ {: .language-bash} 상기 명령어는 파일을 백업하지 않고 대신에 오류가 발생된다: ~~~ cp: target `original-*.dat' is not a directory ~~~ {: .error} `cp` 명령어는 입력값 두개 이상을 받을 때 이런 문제가 발생한다. 이런 상황이 발생할 때, 마지막 입력값을 디렉토리로 예상해서 모든 파일을 해당 디렉토리로 넘긴다. `creatures` 디렉토리에는 `original-*.dat` 라고 이름 붙은 하위 디렉토리가 없기 때문에, 오류가 생긴다. 대신에, 리스트에서 한번에 연산작업을 하나씩 수행하는 **루프(loop)**를 사용할 수 있다. 교대로 각 파일에 대해 첫 3줄을 화면에 출력하는 단순한 예제가 다음에 나와 있다: ~~~ $ for filename in basilisk.dat unicorn.dat > do > head -n 3 $filename # 루프 내부에 들여쓰기는 가독성에 도움을 준다. > done ~~~ {: .language-bash} ~~~ COMMON NAME: basilisk CLASSIFICATION: basiliscus vulgaris UPDATED: 1745-05-02 COMMON NAME: unicorn CLASSIFICATION: equus monoceros UPDATED: 1738-11-24 ~~~ {: .output} > ## `for` 루프 내부에 코드 들여쓰기 > `for` 루프 내부의 코드를 들여쓰는 것이 일반적인 관행이다. > 들여쓰는 유일한 목적은 코드를 더 읽기 쉽게 하는 것 밖에 없다 -- `for` 루프를 실행하는데는 꼭 필요하지는 않다. {: .callout} 쉘이 키워드 `for`를 보게 되면, 쉘은 리스트에 있는 각각에 대해 명령문 하나(혹은 명령문 집합)을 반복할 것이라는 것을 알게 된다. 루프를 반복할 때마다(iteration 이라고도 한다), 현재 작업하고 있는 파일 이름은 `filename`으로 불리는 **변수(variable)**에 할당된다. 리스트의 다음 원소로 넘어가기 전에 루프 내부 명령어가 실행된다. 루프 내부에서, 변수 이름 앞에 `$` 기호를 붙여 변수 값을 얻는다: `$` 기호는 쉘 해석기가 변수명을 텍스트나 외부 명령어가 아닌 **변수**로 처리해서 값을 해당 위치에 치환하도록 지시한다. 이번 경우에 리스트는 파일이름이 두개다: `basilisk.dat`, `unicorn.dat`. 매번 루프가 돌 때마다 파일명을 `filename` 변수에 할당하고 `head` 명령어를 실행시킨다. 즉, 루프가 첫번째 돌 때 `$filename` 은 `basilisk.dat`이 된다. 쉘 해석기는 `basilisk.dat` 파일에 `head` 명령어를 실행시켜서 `basilisk.dat` 파일의 첫 3줄을 화면에 출력시킨다. 두번째 반복에서, `$filename`은 `unicorn.dat`이 된다. 이번에는 쉘이 `head` 명령어를 `unicorn.dat` 파일에 적용시켜 `unicorn.dat` 파일 첫 3줄을 화면에 출력시킨다. 리스트에 원소가 두개라서, 쉘은 `for` 루프를 빠져나온다. 변수명을 분명히 구분하는데, 중괄호 내부에 변수명을 넣어서 변수로 사용하는 것도 가능하다: `$filename` 은 `${filename}`와 동치지만, `${file}name`와는 다르다. 이 표기법을 다른 사람 프로그램에서 찾아볼 수 있다. > ## 루프 내부의 변수 > > 이번 예제는 `data-shell/molecules` 디렉토리를 가정한다. > `ls` 명령어를 던지면 출력결과는 다음과 같다: > > ~~~ > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > ~~~ > {: .output} > > 다음 코드의 출력결과는 어떻게 나오는가? > > ~~~ > $ for datafile in *.pdb > > do > > ls *.pdb > > done > ~~~ > {: .language-bash} > > 이제 다음 코드의 출력결과는 무엇인가? > > ~~~ > $ for datafile in *.pdb > > do > > ls $datafile > > done > ~~~ > {: .language-bash} > > 왜 상기 두 루프 실행결과는 다를까? > > > ## 해답 > > 첫번째 코드 블록은 루프를 돌릴 때마다 동일한 출력결과를 출력한다. > > 배쉬는 루프 몸통 내부 와일드카드 `*.pdb`을 확장해서 `.pdb`로 끝나는 > > 모든 파일을 매칭시킨다. > > 확장된 루프는 다음과 같이 생겼다: > > ``` > > $ for datafile in cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > > do > > > ls cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > > done > > ``` > > {: .language-bash} > > > > ``` > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb > > ``` > > {: .output} > > > > 두번째 코드 블록은 루프를 돌 때마다 다른 파일을 출력한다. > > `datafile` 파일 변수값이 `$datafile`을 통해 평가되고 > > `ls` 명령어를 사용해서 파일 목록을 출력하게 된다. > > > > ``` > > cubane.pdb > > ethane.pdb > > methane.pdb > > octane.pdb > > pentane.pdb > > propane.pdb > > ``` > > {: .output} > {: .solution} {: .challenge} > ## 프롬프트 따라가기 > > 루프안에서 타이핑을 할 때, 쉘 프롬프트가 `$`에서 `>`으로 바뀐다. > 두번째 프롬프트는, `>`, 온전한 명령문 타이핑이 끝마치지 않았음을 상기시키려고 다르게 표기된다. > 세미콜론 `;` 을 사용해서 두 명령어로 구성된 문장을 단일 명령줄로 단순화한다. {: .callout} > ## 동일한 기호, 하지만 다른 의미 > > 쉘 프롬프트로 `>` 기호가 사용되는 것을 확인했지만, > `>` 기호는 출력결과를 방향변경(redirect) 하는데도 사용된다. > 유사하게 `$` 기호를 쉘 프롬프트로 사용했지만, 앞에서 살펴봤듯이, > 쉘로 하여금 변수값을 추출하는데도 사용된다. > > *쉘*이 `>` 혹은 `$` 기호를 출력하게 되면, 사용자가 뭔가 타이핑하길 기대하고 있다는 것으로 > 해당 기호는 프롬프트를 의미한다. > > *사용자 본인*이 `>` 혹은 `$` 기호를 타이핑하게 되면, > 출력결과를 방향변경하거나 변수 값을 끄집어내는 지시를 쉘에 전달하게 된다. {: .callout} `data-shell/creatures` 디렉토리의 예제로 돌아가자. 사람 코드를 읽는 독자에게 목적을 좀더 명확히 하기 위해서 루프의 변수명을 `filename`로 했다. 쉘 자체는 변수명이 어떻게 작명되든지 문제삼지 않는다. 만약 루프를 다음과 같이 작성하거나: ~~~ $ for x in basilisk.dat unicorn.dat > do > head -n 3 $x > done ~~~ {: .language-bash} 혹은: ~~~ $ for temperature in basilisk.dat unicorn.dat > do > head -n 3 $temperature > done ~~~ {: .language-bash} 둘다 정확하게 동일하게 동작한다. *이렇게는 절대 하지 마세요*. 사람이 프로그램을 이해할 수 있을 때만 프로그램이 유용하기 때문에, (`x`같은) 의미없는 이름이나, (`temperature`같은) 오해를 줄 수 있는 이름은 오해를 불러일으켜서 독자가 생각하기에 당연히 프로그램이 수행해야 할 작업을 프로그램이 수행하지 못하게 할 가능성을 높인다. > ## 파일 집합 제한걸기 > > `data-shell/molecules` 디렉토리에서 다음 루프를 실행하게 되면 출력결과는 어떻게 될까? > > ~~~ > $ for filename in c* > > do > > ls $filename > > done > ~~~ > {: .language-bash} > > 1. 어떤 파일도 출력되지 않는다. > 2. 모든 파일이 출력된다. > 3. `cubane.pdb`, `octane.pdb`, `pentane.pdb` 파일만 출력된다. > 4. `cubane.pdb` 파일만 출력된다. > > > ## 해답 > > 정답은 4. 와일드카드 `*` 문자는 0 혹은 그 이상 문자를 매칭하게 된다. > > 따라서, 문자 `c`로 시작하는 문자 다음에 0 혹은 그 이상 문자를 갖는 모든 파일이 매칭된다. > {: .solution} > > How would the output differ from using this command instead? > > ~~~ > $ for filename in *c* > > do > > ls $filename > > done > ~~~ > {: .language-bash} > > 1. The same files would be listed. > 2. All the files are listed this time. > 3. No files are listed this time. > 4. The files `cubane.pdb` and `octane.pdb` will be listed. > 5. Only the file `octane.pdb` will be listed. > > > ## 해답 > > 정답은 4. 와일드카드 `*` 문자는 0 혹은 그 이상 문자를 매칭하게 된다. > > 따라서, `c` 앞에 0 혹은 그 이상 문자가 올 수 있고, `c` 문자 다음에 0 혹은 그 이상 문자가 모두 매칭된다. > {: .solution} {: .challenge} `data-shell/creatures` 디렉토리에서 예제를 계속해서 학습해보자. 다음에 좀더 복잡한 루프가 있다: ~~~ $ for filename in *.dat > do > echo $filename > head -n 100 $filename | tail -n 20 > done ~~~ {: .language-bash} 쉘이 `*.dat`을 전개해서 쉘이 처리할 파일 리스트를 생성한다. 그리고 나서 **루프 몸통(loop body)** 부분이 파일 각각에 대해 명령어 두개를 실행한다. 첫 명령어 `echo`는 명령 라인 매개변수를 표준 출력으로 화면에 뿌려준다. 예를 들어: ~~~ $ echo hello there ~~~ {: .language-bash} prints: ~~~ hello there ~~~ {: .output} 이 사례에서, 쉘이 파일 이름으로 `$filename`을 전개했기 때문에, `echo $filename`은 단지 파일 이름만 화면에 출력한다. 다음과 같이 작성할 수 없다는 것에 주의한다: ~~~ $ for filename in *.dat > do > $filename > head -n 100 $filename | tail -n 20 > done ~~~ {: .language-bash} 왜냐하면, `$filename`이 `basilisk.dat`으로 전개될 때 루프 처음에 쉘이 프로그램으로 인식한 `basilisk.dat`를 실행하려고 하기 때문이다. 마지막으로, `head`와 `tail` 조합은 어떤 파일이 처리되든 81-100줄만 선택해서 화면에 뿌려준다. (파일이 적어도 100줄로 되었음을 가정) > ## 파일, 디렉토리, 변수 등 이름에 공백 > > 공백(whitespace)을 사용해서 루프를 돌릴 때 리스트의 각 원소를 구별했다. > 리스트 원소중 일부가 공백을 갖는 경우, 해당 원소를 인용부호로 감싸서 사용해야 된다. > 데이터 파일이 다음과 같은 이름으로 되었다고 가정하자: > > ~~~ > red dragon.dat > purple unicorn.dat > ~~~ > {: .source} > > 다음을 사용하여 파일을 처리하려고 한다면: > > ~~~ > $ for filename in "red dragon.dat" "purple unicorn.dat" > > do > > head -n 100 "$filename" | tail -n 3 > > done > ~~~ > {: .language-bash} > > 파일명에 공백(혹은 다른 특수 문자)를 회피하는 것이 더 단순하다. > > 상기 파일은 존재하지 않는다. 그래서 상기 코드를 실행하게 되면, `head` 명령어는 > 파일을 찾을 수가 없어서 예상되는 파일명을 보여주는 오류 메시지가 반환된다: > ``` > head: cannot open ‘red dragon.dat’ for reading: No such file or directory > head: cannot open ‘purple unicorn.dat’ for reading: No such file or directory > ``` > {: .output} > 상기 루프 내부 `$filename` 파일명 주위 인용부호를 제거하고 공백 효과를 살펴보자. > `creatures` 디렉토리에서 코드를 실행시키게 되면 `unicorn.dat` 파일에 대한 결과를 루프 명령어 실행 결과를 얻게 됨에 주목한다: > ``` > head: cannot open ‘red’ for reading: No such file or directory > head: cannot open ‘dragon.dat’ for reading: No such file or directory > head: cannot open ‘purple’ for reading: No such file or directory > CGGTACCGAA > AAGGGTCGCG > CAAGTGTTCC > ``` > {: . output} {: .callout} 원래 파일 복사문제로 되돌아가서, 다음 루프를 사용해서 문제를 해결해 보자: ~~~ $ for filename in *.dat > do > cp $filename original-$filename > done ~~~ {: .language-bash} 상기 루프는 `cp` 명령문을 각 파일이름에 대해 실행한다. 처음에 `$filename`이 `basilisk.dat`로 전개될 때, 쉘은 다음을 실행한다: ~~~ cp basilisk.dat original-basilisk.dat ~~~ {: .language-bash} 두번째에는 명령문은 다음과 같다: ~~~ cp unicorn.dat original-unicorn.dat ~~~ {: .language-bash} `cp` 명령어는 아무런 출력결과도 만들어내지 않기 때문에, 루프가 제대로 돌아가는지 확인하기 어렵다. `echo`로 명령문 앞에 위치시킴으로써, 명령문 각각이 제대로 동작되고 있는 확인하는 것이 가능하다. 다음 도표를 통해서 스크립트가 동작할 때 어떤 작업이 수행하고 있는지 상술하고 있다. 또한 `echo` 명령어를 사려깊이 사용하는 것이 어떻게 훌륭한 디버깅 기술이 되는지도 보여주고 있다. ![For Loop in Action](../fig/shell_script_for_loop_flow_chart.svg) ## Nelle의 파이프라인: 많은 파일 처리하기 Nelle은 이제 `goostats` 프로그램(논문 지도교수가 작성한 쉘 스크립트)을 사용해서 데이터 파일을 처리할 준비가 되었다. `goostats` 프로그램은 표본추출 단백질 파일에서 통계량을 산출하는데 인자를 두개 받는다: 1. 입력파일 (원본 데이터를 포함) 2. 출력파일 (산출된 통계량을 저장) 아직 쉘을 어떻게 사용하는지 학습단계에 있기 때문에, 단계별로 요구되는 명령어를 차근히 작성하기로 마음먹었다. 첫번째 단계는 적합한 파일을 선택했는지를 확인하는 것이다 --- 'Z'가 아닌 'A' 혹은 'B'로 파일이름이 끝나는 것이 적합한 파일이라는 것을 명심한다. 홈 디렉토리에서 시작해서, 박사과정 Nelle이 다음과 같이 타이핑한다: ~~~ $ cd north-pacific-gyre/2012-07-03 $ for datafile in NENE*[AB].txt > do > echo $datafile > done ~~~ {: .language-bash} ~~~ NENE01729A.txt NENE01729B.txt NENE01736A.txt ... NENE02043A.txt NENE02043B.txt ~~~ {: .output} 다음 단계는 `goostats` 분석 프로그램이 생성할 파일이름을 무엇으로 할지 결정하는 것이다. "stats"을 각 입력 파일에 접두어로 붙이는 것이 간단해 보여서, 루프를 변경해서 작업을 수행하도록 한다: ~~~ $ for datafile in NENE*[AB].txt > do > echo $datafile stats-$datafile > done ~~~ {: .language-bash} ~~~ NENE01729A.txt stats-NENE01729A.txt NENE01729B.txt stats-NENE01729B.txt NENE01736A.txt stats-NENE01736A.txt ... NENE02043A.txt stats-NENE02043A.txt NENE02043B.txt stats-NENE02043B.txt ~~~ {: .output} `goostats`을 아직 실행하지는 않았지만, 이제 확신할 수 있는 것은 올바른 파일을 선택해서, 올바른 출력 파일이름을 생성할 수 있다는 점이다. 명령어를 반복적으로 타이핑하는 것은 귀찮은 일이지만, 더 걱정이 되는 것은 Nelle이 타이핑 실수를 하는 것이다. 그래서 루프를 다시 입력하는 대신에 위쪽 화살표를 누른다. 위쪽 화살표에 반응해서 컴퓨터 쉘은 한줄에 전체 루프를 다시 보여준다. (스크립트 각 부분이 구분되는데 세미콜론이 사용됨): ~~~ $ for datafile in NENE*[AB].txt; do echo $datafile stats-$datafile; done ~~~ {: .language-bash} 왼쪽 화살표 키를 사용해서, Nelle은 `echo`명령어를 `bash goostats`으로 변경하고 백업한다: ~~~ $ for datafile in NENE*[AB].txt; do bash goostats $datafile stats-$datafile; done ~~~ {: .language-bash} 엔터키를 누를 때, 쉘은 수정된 명령어를 실행한다. 하지만, 어떤 것도 일어나지 않는 것처럼 보인다 --- 출력이 아무것도 없다. 잠시뒤에 Nelle은 작성한 스크립트가 화면에 아무것도 출력하지 않아서, 실행되고 있는지, 얼마나 빨리 실행되는지에 대한 정보가 없다는 것을 깨닫는다. 컨트롤+C(Control-C)를 눌러서 작업을 종료하고, 반복할 명령문을 위쪽 화살표로 선택하고, 편집해서 다음과 같이 작성한다: ~~~ $ for datafile in NENE*[AB].txt; do echo $datafile; bash goostats $datafile stats-$datafile; done ~~~ {: .language-bash} > ## 시작과 끝 > > 쉘에 ^A, 콘트롤+A(Control-A, `Ctrl-a`)를 타이핑해서 해당 라인 처음으로 가고, > ^E (`Ctrl-e`, Control-E)를 쳐서 라인의 끝으로 이동한다. {: .callout} 이번에 프로그램을 실행하면, 매 5초간격으로 한줄을 출력한다: ~~~ NENE01729A.txt NENE01729B.txt NENE01736A.txt ... ~~~ {: .output} 1518 곱하기 5초를 60으로 나누면, 작성한 스크립트를 실행하는데 약 2시간 정도 소요된다고 볼 수 있다. 마지막 점검으로, 또다른 터미널 윈도우를 열어서, `north-pacific-gyre/2012-07-03` 디렉토리로 가서, `cat stats-NENE01729B.txt`을 사용해서 출력파일 중 하나를 면밀히 조사한다. 출력결과가 좋아보인다. 그래서 커피를 마시고 그동안 밀린 논문을 읽기로 한다. > ## 역사(history)를 아는 사람은 반복할 수 있다. > > 앞선 작업을 반복하는 또다른 방법은 `history` 명령어를 사용하는 것이다. > 실행된 마지막 수백개 명령어 리스트를 얻고 나서, > 이들 명령어 중 하나를 반복실행하기 위해서 `!123`("123"은 명령 숫자로 교체된다.)을 사용한다. > 예를 들어 Nelle이 다음과 같이 타이핑한다면: > > ~~~ > $ history | tail -n 5 > ~~~ > {: .language-bash} > ~~~ > 456 ls -l NENE0*.txt > 457 rm stats-NENE01729B.txt.txt > 458 bash goostats NENE01729B.txt stats-NENE01729B.txt > 459 ls -l NENE0*.txt > 460 history > ~~~ > {: .output} > > 그리고 나서, 단순히 `!458`을 타이핑함으로써, > `NENE01729B.txt` 파일에 `goostats`을 다시 실행할 수 있게 된다. {: .callout} > ## 다른 이력(history) 명령어 > > 이력(history)에 접근하는 단축 명령어가 다수 존재한다. > > - `Ctrl-R` 탄축키는 "reverse-i-search" 이력 검색모드로 > 입력한 텍스트와 매칭되는 가장 최슨 명령어를 이력에서 찾아서 제시한다. > `Ctrl-R` 단축키를 한번 혹은 그 이상 누르게 되면 그 이전 매칭을 검색해 준다. > - `!!` 명령어는 바로 직전 명령어를 불러온다. > (키보드 윗화살표를 사용하는 것보다 더 편리할수도 편리하지 않을 수도 있다.) > - `!$` 명령어는 마지막 명령문의 마지막 단어를 불러온다. > 기대했던 것보다 훨씬 유용할 수 있다: > `bash goostats NENE01729B.txt stats-NENE01729B.txt` 명령문을 실행한 후에 > `less !$`을 타이핑하게 되면 `stats-NENE01729B.txt` 파일을 찾아준다. > 키보드 위화살표를 눌러 명령라인을 편집하는 것보다 훨씬 빠르다. {: .callout} > ## 루프 내부에서 파일에 저장하기 - 1부 > > `data-shell/molecules` 디렉토리에 있다고 가정하자. > 다음 루프의 효과는 무엇인가? > > ~~~ > $ for alkanes in *.pdb > > do > > echo $alkanes > > cat $alkanes > alkanes.pdb > > done > ~~~ > {: .language-bash} > > 1. `fructose.dat`, `glucose.dat`, `sucrose.dat`을 출력하고, `sucrose.dat`에서 나온 텍스트를 `xylose.dat`에 저장된다. > 2. `fructose.dat`, `glucose.dat`, `sucrose.dat`을 출력하고, 모든 파일 3개에서 나온 텍스트를 합쳐 `xylose.dat`에 저장된다. > 3. `fructose.dat`, `glucose.dat`, `sucrose.dat`, `xylose.dat`을 출력하고, `sucrose.dat`에서 나온 텍스트를 `xylose.dat`에 저장된다. > 4. 위 어느 것도 아니다. > > > ## 해답 > > 1. 순차적으로 각 파일의 텍스트가 `alkanes.pdb` 파일에 기록된다. > > 하지만, 루프가 매번 반복될 때마다 파일에 덮어쓰기가 수행되어서 마지막 `alkanes.pdb` 파일 텍스트만 > > `alkanes.pdb` 파일에 기록된다. > {: .solution} {: .challenge} > ## 루프 내부에서 파일에 저장하기 - 2부 > > 이번에도 `data-shell/molecules` 디렉토리에 있다고 가정하고, 다음 루프 실행 출력결과는 무엇일까? > > ~~~ > $ for datafile in *.pdb > > do > > cat $datafile >> all.pdb > > done > ~~~ > {: .language-bash} > > 1. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb` 파일에 나온 모든 모든 텍스트가 하나로 붙여져서 > `all.pdb` 파일에 저장된다. > 2. `ethane.pdb` 파일에 나온 텍스트만 `all.pdb` 파일에 저장된다. > 3. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb`, `propane.pdb` 파일에서 나온 모든 텍스트가 > 하나로 풑여져서 `all.pdb` 파일에 저장된다. > 4. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb`, `propane.pdb` 파일에서 나온 > 모든 텍스트가 화면에 출력되고 `all.pdb` 파일에 저장된다. > > > ## 해답 > > 정답은 3. 명령어 실행 출력결과를 방향변경하여 덮었는 것이 아니라 `>>` 기호는 파일에 덧붙인다. > > `cat` 명령어에서 나온 출력결과가 파일로 방향변경되어 어떤 출력결과도 화면에 출력되지는 않는다. > {: .solution} {: .challenge} > ## 시운전(Dry Run) > > 루프는 한번에 많은 작업을 수행하는 방식이다 --- 만약 잘못된 것이 있다면, > 한번에 실수를 대단히 많이 범하게 된다. > 루프가 수행하는 작업을 점검하는 한 방법이 실제로 루프를 돌리는 대신에 > `echo` 명령어를 사용하는 것이다. > > 실제로 명령어를 실행하지 않고, 다음 루프가 실행할 명령어를 머릿속으로 미리보고자 한다고 가정한다: > > ~~~ > $ for file in *.pdb > > do > > analyze $file > analyzed-$file > > done > ~~~ > {: .language-bash} > > 아래 두 루프 사이에 차이는 무엇이고, 어느 것을 시운전으로 실행하고 싶은가? > > ~~~ > # Version 1 > $ for file in *.pdb > > do > > echo analyze $file > analyzed-$file > > done > ~~~ > {: .language-bash} > > ~~~ > # Version 2 > $ for file in *.pdb > > do > > echo "analyze $file > analyzed-$file" > > done > ~~~ > {: .language-bash} > > > ## 해답 > > 두번째 버젼을 실행하면 좋을 것이다. > > 달러 기호로 접두명을 주었기 때문에 루프 변수를 확장해서 인용부호로 감싼 모든 것을 화면에 출력한다. > > > > 첫번째 버전은 `echo analyze $file` 명령을 수행해서 `analyzed-$file` 파일로 > > 출력결과를 방향변경하여 저장시킨다. 따라서 파일이 쭉 자동생성된다:`analyzed-cubane.pdb`, > > `analyzed-ethane.pdb` ... > > > > 두가지 버젼을 직접 실행해보고 출력결과를 살펴보자! > > `analyzed-*.pdb` 파일을 열어서 파일에 기록된 내용도 살펴본다. > {: .solution} {: .challenge} > ## 중첩루프(Nested Loops) > > 다른 화합물과 다른 온도를 갖는 조합을 해서, 각 반응율 상수를 측정하는 > 실험을 조직하도록 이에 상응하는 디렉토리 구조를 갖추고자 한다. > 다음 코드 실행결과는 어떻게 될까? > > ~~~ > $ for species in cubane ethane methane > > do > > for temperature in 25 30 37 40 > > do > > mkdir $species-$temperature > > done > > done > ~~~ > {: .language-bash} > > > ## 해답 > > 중첩 루프(루프 내부에 루프가 포함됨)를 생성하게 된다. > > 외부 루프에 각 화학물이, 내부 루프(중첩된 루프)에 온도 조건을 반복하게 되서, > > 화학물과 온도를 조합한 새로운 디렉토리가 쭉 생성된다. > > > > 직접 코드를 실행해서 어떤 디렉토리가 생성되는지 확인한다! > {: .solution} {: .challenge}
23.700292
114
0.637283
kor_Hang
1.00001
912b1e112548b60e6bc6100387aff62084aa29e9
3,646
md
Markdown
desktop-src/SecGloss/b-gly.md
velden/win32
94b05f07dccf18d4b1dbca13b19fd365a0c7eedc
[ "CC-BY-4.0", "MIT" ]
552
2019-08-20T00:08:40.000Z
2022-03-30T18:25:35.000Z
desktop-src/SecGloss/b-gly.md
velden/win32
94b05f07dccf18d4b1dbca13b19fd365a0c7eedc
[ "CC-BY-4.0", "MIT" ]
1,143
2019-08-21T20:17:47.000Z
2022-03-31T20:24:39.000Z
desktop-src/SecGloss/b-gly.md
velden/win32
94b05f07dccf18d4b1dbca13b19fd365a0c7eedc
[ "CC-BY-4.0", "MIT" ]
1,287
2019-08-20T05:37:48.000Z
2022-03-31T20:22:06.000Z
--- description: Contains definitions of security terms that begin with the letter B. ROBOTS: NOINDEX, NOFOLLOW ms.assetid: 2e570727-7da0-4e17-bf5d-6fe0e6aef65b title: B (Security Glossary) ms.topic: article ms.date: 05/31/2018 --- # B (Security Glossary) [A](a-gly.md) B [C](c-gly.md) [D](d-gly.md) [E](e-gly.md) F [G](g-gly.md) [H](h-gly.md) [I](i-gly.md) J [K](k-gly.md) [L](l-gly.md) [M](m-gly.md) [N](n-gly.md) [O](o-gly.md) [P](p-gly.md) Q [R](r-gly.md) [S](s-gly.md) [T](t-gly.md) [U](u-gly.md) [V](v-gly.md) [W](w-gly.md) [X](x-gly.md) Y Z <dl> <dt> <span id="_security_backup_authority_gly"></span><span id="_SECURITY_BACKUP_AUTHORITY_GLY"></span>**backup authority** </dt> <dd> A trusted application running on a secure computer that provides secondary storage for the session keys of its clients. The backup authority stores session keys as key BLOBs that are encrypted with the backup authority's public key. </dd> <dt> <span id="_security_base_content_type_gly"></span><span id="_SECURITY_BASE_CONTENT_TYPE_GLY"></span>**base content type** </dt> <dd> A type of data contained in a PKCS \#7 message. Base content types only contain data, no cryptographic enhancements such as hashes or signatures. Currently, the only base content type is the Data content type. </dd> <dt> <span id="_security_base_cryptographic_functions_gly"></span><span id="_SECURITY_BASE_CRYPTOGRAPHIC_FUNCTIONS_GLY"></span>**base cryptographic functions** </dt> <dd> The lowest level of functions in the CryptoAPI architecture. They are used by applications and other high-level CryptoAPI functions to provide access to CSP-provided cryptographic algorithms, secure key generation, and secure storage of secrets. See also [*cryptographic service providers*](c-gly.md). </dd> <dt> <span id="_security_basic_encoding_rules_gly"></span><span id="_SECURITY_BASIC_ENCODING_RULES_GLY"></span>**Basic Encoding Rules** </dt> <dd> (BER) The set of rules used to encode ASN.1 defined data into a stream of bits (zeros or ones) for external storage or transmission. A single ASN.1 object may have several equivalent BER encodes. BER is defined in CCITT Recommendation X.209. This is one of the two encoding methods currently used by CryptoAPI. </dd> <dt> <span id="_security_ber_gly"></span><span id="_SECURITY_BER_GLY"></span>**BER** </dt> <dd> See *Basic Encoding Rules*. </dd> <dt> <span id="_security_big_endian_gly"></span><span id="_SECURITY_BIG_ENDIAN_GLY"></span>**big-endian** </dt> <dd> A memory or data format in which the most significant byte is stored at the lower address or arrives first. See also [*little-endian*](l-gly.md). </dd> <dt> <span id="_security_blob_gly"></span><span id="_SECURITY_BLOB_GLY"></span>**BLOB** </dt> <dd> A generic sequence of bits that contain one or more fixed-length header structures plus context specific data. See also [*key BLOBs*](k-gly.md), [*certificate BLOBs*](c-gly.md), [*certificate name BLOBs*](c-gly.md), and [*attribute BLOBs*](a-gly.md). </dd> <dt> <span id="_security_block_cipher_gly"></span><span id="_SECURITY_BLOCK_CIPHER_GLY"></span>**block cipher** </dt> <dd> A cipher algorithm that encrypts data in discrete units (called blocks), rather than as a continuous stream of bits. The most common block size is 64 bits. For example, DES is a block cipher. See also [*stream cipher*](s-gly.md). </dd> <dt> <span id="_security_bulk_encryption_key_gly"></span><span id="_SECURITY_BULK_ENCRYPTION_KEY_GLY"></span>**bulk encryption key** </dt> <dd> A session key derived from a master key. Bulk encryption keys are used in [*Schannel*](s-gly.md) encryption. </dd> </dl>    
39.204301
310
0.729292
eng_Latn
0.872546
912b3493a48828811ee290d29f546625c794918f
705
md
Markdown
docs/api/ESCWalkContext.md
StraToN/unofficial-escoria-reloaded
ccb34e319b716b4d3afc540fbb970348d872ffbf
[ "MIT" ]
7
2021-03-09T08:13:45.000Z
2021-09-20T07:12:08.000Z
docs/api/ESCWalkContext.md
StraToN/unofficial-escoria-reloaded
ccb34e319b716b4d3afc540fbb970348d872ffbf
[ "MIT" ]
17
2021-05-15T16:10:14.000Z
2021-07-04T17:00:05.000Z
docs/api/ESCWalkContext.md
StraToN/unofficial-escoria-reloaded
ccb34e319b716b4d3afc540fbb970348d872ffbf
[ "MIT" ]
null
null
null
<!-- Auto-generated from JSON by GDScript docs maker. Do not edit this document directly. --> # ESCWalkContext **Extends:** [Object](../Object) ## Description The walk context describes the target of a walk command and if that command should be executed fast ## Property Descriptions ### target\_object ```gdscript var target_object: ESCObject ``` Target object that the walk command tries to reach ### target\_position ```gdscript var target_position: Vector2 = "(0, 0)" ``` The target position ### fast ```gdscript var fast: bool ``` Wether to move fast ## Method Descriptions ### \_init ```gdscript func _init(p_target_object: ESCObject, p_target_position: Vector2, p_fast: bool) ```
15.326087
93
0.719149
eng_Latn
0.903082
912ba768002369b6d06e5b9801803f10a0ad41ee
9,164
md
Markdown
content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md
lnds/lnds-site
c7d8483a764c91f1653c77ab6934c4f34d847f62
[ "MIT" ]
null
null
null
content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md
lnds/lnds-site
c7d8483a764c91f1653c77ab6934c4f34d847f62
[ "MIT" ]
null
null
null
content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md
lnds/lnds-site
c7d8483a764c91f1653c77ab6934c4f34d847f62
[ "MIT" ]
null
null
null
--- comments: true date: 2009-05-26 20:53:24 layout: post slug: las-maquinas-del-fin-del-mundo-intermedio title: Las máquinas del fin del mundo (intermedio) wordpress_id: 191 categories: - General - Paradigmas --- Ya hemos visto [una posición](http://www.lnds.net/2009/05/el-desafio-del-nuevo-ludita.html), en uno de los extremos, sobre que hacer ante las amenazas de la tecnología para nuestra sociedad. En 2006 escribí este artículo, que he decidido retocar y adaptar a mi serie actual sobre los desafíos que nos impone la tecnología, creo que es apropiado publicarlo en este momento, como una suerte de puente hacia la otra propuesta extrema, sobre que hacer ante estos escenarios, en que podemos ser sobrepasados por la tecnología. Las máquinas del fin del mundo Leí con asombro un post de mi hermano, sobre una cadena que [anuncia el fin del mundo para el 6 de junio de este año (6/6/6)](http://www.ricardodiaz.org/archives/2006/04/fin.html) [^1]. Aparte de ser una interesante coincidencia numérica, ¿qué razones hay para pensar de que el mundo debe acabar? ## El fin por culpa de la tecnología La verdad es que el fin del mundo ha sido anunciado muchas veces, y no sólo por fanáticos religiosos. ![atomic_bomb_explosion_2.jpg](atomic_bomb_explosion_2.jpg) Las visiones apocalípticas normalmente anuncian todo tipo de cataclismos, pero, como sería de esperar, en el último tiempo, se han agregado otras causas para el fin del mundo, muchas consecuencia de nuestro abuso de la tecnología: * El fin del mundo por el uso de armas bioquímicas * Guerra Nuclear total * La explosión de plantas nucleares * [Una super máquina que crea un vórtice](http://www.lnds.net/2008/09/buuum.html), o un agujero negro * El calentamiento global y una nueva era glacial * El surgimiento de inteligencias artificiales que destruyen a su creador Esta última "amenaza" es la que me interesa explorar ahora, porque tiene que ver con temas que domino más. ## El miedo a la inteligencia artificial ![hal.jpg](hal9000.jpg) El temor al alzamiento de las máquinas ha sido discutido en taquilleras películas como Terminator y Matrix, y de una forma más elegante en "2001 Odisea Espacial". Cuando se habla de esta visión de la máquina que se alza contra su creador, muchos comparan con el [Frankenstein](http://www.literature.org/authors/shelley-mary/frankenstein/), de Mary Shelley. En realidad no es lo mismo, pues la "criatura" es un humano vuelt o a la vida (bueno, un cuarpo humano reanimado, construido a partir de partes de muchos cadaveres). En Matrix no sabemos cómo son las "máquinas", vemos robots en forma de calamares, pero son sólo "trajes", para acomodar inteligencias artificiales. Las inteligencias artificiales viven en la matrix, un mundo virtual, que tiene un soporte físico (hardware), alimentado por la energía bioquímica de millones de seres humanos. En Terminator la idea es similar, una inteligencia artificial surge desde una red militar (skynet), y esta intelgencia decide destruir al hombre, para eso se sirve de robots, conocidos como exterminadores, pero nuevamente, las IA están en otra parte. John Connor sabe que debe destruir Skynet antes que se forme, porque probablemente el acceso a Skynet es imposible. Neo se dirige hacia una suerte de nucleo de la Matrix, porque sabe que aunque maten a todos los robots calamares estos no son el problema. ![hyperion_early_sm.jpg](hyperion_early_sm.jpg) La brillante saga de ciencia ficción Hyperión, de Dan Simmons, nos muestra una guerra entre las IA y las personas, con una interesante variante, las IA han decidido sacrificarse a si mismas para crear IA más evolucionadas (¿seríamos capaces de sacrificarnos por la especie que nos suceda?). En todos estos casos, ls enemigos, no vistos, son Inteligencias Artificiales que viven, flotan, por decirlo de alguna manera, en una red, o soporte físico. Son programas, algoritmos, heurísticas, en cierta manera, pero con capacidad de aprender y manipular su entorno con el fin de protegerse, por ejemplo, construyendo robots. No son los robots los peligrosos, estos son herramientas, son las inteligencias artificiales las que comandan y las peligrosas. Pero, por su naturaleza, las IA son escurridizas. En 2001, HAL es eliminado, pues tiene un soporte físico de fácil acceso. Pero en Matrix y Terminator, eso no es posible, y los heroes de esa sagas, deben tratar de evitar que surjan, en el caso de Terminator, o atacar el núcleo donde se encuentran, como en Matrix. Sin embargo, la estrategia de John Connor es más astuta que la de Neo, porque John Connor quiere evitar que Skynet llegue a existir, pero se produce una interesante paradoja, que si John Connor destruye Skynet, el dejará de existir. En el caso de Matrix, sabemos que Neo no puede destruir a las IA, que la lucha es eterna y se repetirá por siempre, porque Neo en cierta manera es un mecanismo de la misma matrix para reorganizarse. Quizás por eso Matrix como trilogía decepciona tanto. Porque esperabamos un desenlace más emocionante, pero en realidad nos llevan a un final decepcionante, en que las IA ganan. Pero también, nos queda claro, que en cierta manera, la matrix sigue dependiendo de humanos libres, una élite acaso, que viven en Zión. Hemos visto, que sólo en 2001 el héroe logra eliminar la amenaza de la IA. HAL falla porque es incapaz en protegerse a si mismo. ## Vida Artificial Actualmente se postula que antes de construir una IA, debemos contruir niveles inferiores de inteligencia. Estos rdimentos de inteligencia artificial ya están disponibles, [incluso al alcance de cualquiera de nosotros](http://www.lnds.net/2009/03/el-sueno-de-las-ovejas-electricas.html). Estos rudimentos, constituyen los "instintos" de las IA. Sin esos instintos básicos, que le permitan a la IA desarrollarse, no será posible que esta llegue a ser conciente. Una Inteligencia Artificial debería 1. Preservarse 2. Reproducirse 3. Evolucionar Estos tres requisitos básicos son las bases de la denominada [tesis fuerte de la vida artificia](http://vidaartificial.com/index.php?title=Vida_Artificial)l, en esencia esta dice que una vida es un modelo operacional, y si satisface las tres operaciones mencionadas, entonces da lo mismo el soporte físico de la vida. La satisfacción de estas tres necesidades de las IAs son las que justificarían su hostilidad hacia nosotros. Efectivamente, si hacemos precaria su preservación, si tienen que competir con nosotros por los recursos, o somos un estorbo dentro de su evolución, las IAs pueden volverse contra sus creadores. ¿Dónde surgirá la Inteligencia Artificial? Lo interesante, es saber si las inteligencias artificiales son capaces de desarrollarse fuera de su soporte físico, o antes de llegar a tenerlo, es decir, [¿pueden las IA surgir dentro de la red?](http://www.lnds.net/2008/12/y-si-las-inteligencias-artificiales-ya-e.html). En Terminator, lo primero que se desarrolla es Skynet, una red informática, que empieza a controlar las factorías necesarias para desarrollar las armas y robots exterminadores. Entonces, además de preocuparnos del mundo industrial, y de la amenazas de las máquinas, que principalmente son amenazas ambientales, es interesante concentrarse en entender las amenazas que representan las tecnologías de la información, es decir, la amenaza de que se desarrolle una IA hostil. Consideren el siguiente fragmento: > "¿Saben lo que es barato en el siglo 21? El tiempo de computación. ¿Saben lo que es caro? El juicio humano. Ambos no son intercambiables. Los humanos son buenos entendiendo cosas, los computadores son buenos contando cosas, pero los humanos apestan al contar, y os computadores apestan al entender." > "¿Conocen los algoritmos genéticos? Tomen cualquier problema y generen 10 billones de pogramas de computador al azar y póngalos a resolver el problema. Tomen el 10% de los que lo hacen mejor, usen variantes al azar de ellos y repítanlos otras 10 billones de veces. Repítanlo 10 billones de veces más, y vuelvan en un día o dos, y descubrirán que el computador ha "evolucionado" a una extraña y bizarra respuesta que a ningún humano jamás se le habría ocurrido." > > "Ningún humano va a lograr escribir el código para una Inteligencia Artificial (IA). La inteligencia es una propiedad emergente de factores evolutivos, no de la planificación central. Es anarquía, no estalinismo. ¿Lo entienden?"[^2]. ¿Que tal si el surgimiento de la inteligencia artificial es algo inevitable, y eso hace que tengamos que empezar a aceptar que tendremos que lidiar con seres potencialmente superiores a nosotros? Hay una propuesta extrema, que nos dice lo que debemos hacer en este caso, y que al contrario de lo propuesto por los neoluditas, y los que llaman a destruir, o abandonar la tecnología, toma un camino radicalmente opuesto, y es el que vamos a explorar en el siguiente post. [^1]: La primera parte de este post fue escrita en abril de 2006, bajo el título de [Las máquinas del fin del mundo](/2006/04/las-maquinas-del-fin-del-mundo.html). [^2]: Cory Doctorow, "Pester Power", relato publicado en Communications of ACM, diciembre 2008.
78.324786
463
0.791248
spa_Latn
0.997531
912c31695f32b5ae8ab45d6507060e2a7a205424
1,480
md
Markdown
intl.en-US/Product Introduction/Benefits.md
vlgnaw/emapreduce
37918944befffc3895b53fe8d8ae5d793331f401
[ "MIT" ]
null
null
null
intl.en-US/Product Introduction/Benefits.md
vlgnaw/emapreduce
37918944befffc3895b53fe8d8ae5d793331f401
[ "MIT" ]
null
null
null
intl.en-US/Product Introduction/Benefits.md
vlgnaw/emapreduce
37918944befffc3895b53fe8d8ae5d793331f401
[ "MIT" ]
null
null
null
# Benefits {#concept_j4w_dky_w2b .concept} E-MapReduce has some practical strength over self-built clusters. For example, it provides some convenient and controllable means to manage its clusters. In addition, it also has the following strengths: - Usability User can select the required ECS types and disks and select the required software for automatic deployment. Users can apply for cluster resources at the corresponding position according to the geographical location where users or the data source are located. Now, Alibaba Cloud ECS supports regions, including China East 1, China East 2, China North 1, China North 2, China South 1, Singapore, Hong Kong, US East 1 and US West 1. E-MapReduce supports regions including China North 2, China East 1, China East 2 and China South 1, and later it will extend to all the regions supported by Alibaba Cloud ECS. - Low price The user can create a cluster as needed, that is, it can release the cluster after an offline task running is completed and add a node dynamically when needed. - Deep integration E-MapReduce can be subject to deep integration with other Alibaba Cloud products, so that they can be used as the input source or output destination of Hadoop or Spark computing engine in E-MapReduce. - Security E-MapReduce integrates Alibaba Cloud RAM resource permission management system, so that it can isolate the service permissions through the primary account or sub-accounts.
61.666667
501
0.787162
eng_Latn
0.999365
912c7f22c8273129c8aa65f19f04180f50e966d9
1,549
md
Markdown
business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md
nschonni/dynamics365smb-docs
619182073e912c1373c58db16c20f0770aefc1b3
[ "CC-BY-4.0", "MIT" ]
null
null
null
business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md
nschonni/dynamics365smb-docs
619182073e912c1373c58db16c20f0770aefc1b3
[ "CC-BY-4.0", "MIT" ]
null
null
null
business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md
nschonni/dynamics365smb-docs
619182073e912c1373c58db16c20f0770aefc1b3
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: How to Print Direct Sales and Purchase Details Reports | Microsoft Docs description: The **Direct Sales Details** and **Direct Purchase Details** reports include headers with order numbers and descriptions from sales and purchase documents. services: project-madeira documentationcenter: '' author: SorenGP ms.service: dynamics365-business-central ms.topic: article ms.devlang: na ms.tgt_pltfrm: na ms.workload: na ms.search.keywords: ms.date: 10/01/2018 ms.author: sgroespe --- # Print Direct Sales and Purchase Details Reports The **Direct Sales Details** and **Direct Purchase Details** reports include headers with order numbers and descriptions from sales and purchase documents. You can filter or select data for these reports based on: - General ledger account number - Document number - Posting date ## To print direct sales and direct purchase details reports 1. Choose the ![Search for Page or Report](../../media/ui-search/search_small.png "Search for Page or Report icon") icon, enter **Direct Sales Details** or **Direct Purchase Details**, and then choose the related link. 2. On the **G/L Entry** FastTab, select the appropriate filters. 3. Choose the **Print** button to print the report or choose the **Preview** button to view it on the screen. ## See Also [United Kingdom Local Functionality](united-kingdom-local-functionality.md) [Sales](../../sales-manage-sales.md) [Purchasing](../../purchasing-manage-purchasing.md)
41.864865
221
0.723047
eng_Latn
0.96782
912cdb13d5de0c13eb0d4826d9af66f3f2c938dd
50
md
Markdown
README.md
jorge-matricali/jwt-crack
16262f521ee871a0e2581e8b2016c9477c884be0
[ "MIT" ]
1
2019-12-10T23:52:16.000Z
2019-12-10T23:52:16.000Z
README.md
jorge-matricali/jwt-crack
16262f521ee871a0e2581e8b2016c9477c884be0
[ "MIT" ]
null
null
null
README.md
jorge-matricali/jwt-crack
16262f521ee871a0e2581e8b2016c9477c884be0
[ "MIT" ]
1
2019-11-05T16:47:23.000Z
2019-11-05T16:47:23.000Z
# jwt-crack JWT brute force cracker written in C.
16.666667
37
0.76
eng_Latn
0.986595
912cde06c9d7598ad824441dddc898c518243f1a
573
md
Markdown
README.md
rmurai0610/DArgs
170b7d354c90c3212886eb57e04999c2f16d7de0
[ "MIT" ]
null
null
null
README.md
rmurai0610/DArgs
170b7d354c90c3212886eb57e04999c2f16d7de0
[ "MIT" ]
null
null
null
README.md
rmurai0610/DArgs
170b7d354c90c3212886eb57e04999c2f16d7de0
[ "MIT" ]
null
null
null
# DArgs - Dumb Argument Parser for C++ DArgs is a minimal, simple argument parser for C++. DArgs parses the arguments as the options are defined, enabling the user to use DArgs with minimal lines of code. ## Example ``` DArgs::DArgs dargs(argc, argv); std::string dataset = dargs("--dataset", "Path to the dataset to load", std::string("dataset/default.txt")); int n_iters = dargs("--n_iters", "Number of iterations", 200); double alpha = dargs("--alpha", "alpha value", 0.0001f); if (!dargs.check()) { dargs.print_help(); exit(EXIT_FAILURE); } ```
35.8125
113
0.675393
eng_Latn
0.714656
912d3f6be97e0a30352980eab51df33a4d60bfa1
2,930
md
Markdown
docs/framework/wcf/guidelines-and-best-practices.md
nicolaiarocci/docs.it-it
74867e24b2aeb9dbaf0a908eabd8918bc780d7b4
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/wcf/guidelines-and-best-practices.md
nicolaiarocci/docs.it-it
74867e24b2aeb9dbaf0a908eabd8918bc780d7b4
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/wcf/guidelines-and-best-practices.md
nicolaiarocci/docs.it-it
74867e24b2aeb9dbaf0a908eabd8918bc780d7b4
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Linee guida e suggerimenti ms.date: 03/30/2017 helpviewer_keywords: - WCF, guidelines - best practices [WCF], application design - Windows Communication Foundation, best practices - WCF, best practices - Windows Communication Foundation, guidelines ms.assetid: 5098ba46-6e8d-4e02-b0c5-d737f9fdad84 ms.openlocfilehash: 1c16c0936f729636aaa56e6e50c33e18c2b0f788 ms.sourcegitcommit: 5b6d778ebb269ee6684fb57ad69a8c28b06235b9 ms.translationtype: MT ms.contentlocale: it-IT ms.lasthandoff: 04/08/2019 ms.locfileid: "59181342" --- # <a name="guidelines-and-best-practices"></a>Linee guida e suggerimenti In questa sezione contiene argomenti che forniscono le linee guida per la creazione di applicazioni Windows Communication Foundation (WCF). ## <a name="in-this-section"></a>In questa sezione [Procedure consigliate: Controllo delle versioni dei contratti dati](../../../docs/framework/wcf/best-practices-data-contract-versioning.md) Viene illustrato come e quando creare contratti dati che non determinano interruzioni quando vengono create versioni successive. [Controllo delle versioni dei servizi](../../../docs/framework/wcf/service-versioning.md) Viene illustrato come controllare la versione in WCF. Dopo la distribuzione potrebbe essere necessario modificare i servizi (e gli endpoint che espongono) per soddisfare, ad esempio, la modifica di requisiti aziendali o di requisiti IT o per correggere problemi. Ogni modifica introduce una nuova versione del servizio. [Bilanciamento del carico](../../../docs/framework/wcf/load-balancing.md) Elenca le linee guida per il bilanciamento del carico con una Web farm. [Controllo dell'utilizzo di risorse e miglioramento delle prestazioni](../../../docs/framework/wcf/controlling-resource-consumption-and-improving-performance.md) Descrive le proprietà progettate per impedire il consumo indebito di risorse e per migliorare la sicurezza e fa riferimento a informazioni più complete riguardo al loro utilizzo. [Distribuzione di applicazioni WCF con ClickOnce](../../../docs/framework/wcf/deploying-wcf-applications-with-clickonce.md) Descrive le considerazioni da fare in caso di utilizzo della funzionalità ClickOnce. ## <a name="reference"></a>Riferimenti <xref:System.ServiceModel> <xref:System.Runtime.Serialization> ## <a name="related-sections"></a>Sezioni correlate [Panoramica dei concetti](../../../docs/framework/wcf/conceptual-overview.md) [Programmazione WCF di base](../../../docs/framework/wcf/basic-wcf-programming.md) ## <a name="see-also"></a>Vedere anche - [Informazioni su Windows Communication Foundation](../../../docs/framework/wcf/whats-wcf.md) - [Esempi di Windows Communication Foundation (WCF)](./samples/index.md) - [Panoramica dei concetti](../../../docs/framework/wcf/conceptual-overview.md) - [Creazione di client](../../../docs/framework/wcf/building-clients.md)
55.283019
322
0.768259
ita_Latn
0.943743
912db653b18767cd57ca5921077b47029a577796
1,947
md
Markdown
dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: SalesTransaction.ChannelReferenceId Property (Microsoft.Dynamics.Commerce.Runtime.DataModel) TOCTitle: ChannelReferenceId Property ms:assetid: P:Microsoft.Dynamics.Commerce.Runtime.DataModel.SalesTransaction.ChannelReferenceId ms:mtpsurl: https://technet.microsoft.com/en-us/library/microsoft.dynamics.commerce.runtime.datamodel.salestransaction.channelreferenceid(v=AX.60) ms:contentKeyID: 49834319 ms.date: 05/18/2015 mtps_version: v=AX.60 f1_keywords: - Microsoft.Dynamics.Commerce.Runtime.DataModel.SalesTransaction.ChannelReferenceId dev_langs: - CSharp - C++ - VB --- # ChannelReferenceId Property Gets or sets the channel reference identifier. **Namespace:**  [Microsoft.Dynamics.Commerce.Runtime.DataModel](microsoft-dynamics-commerce-runtime-datamodel-namespace.md) **Assembly:**  Microsoft.Dynamics.Commerce.Runtime.Entities (in Microsoft.Dynamics.Commerce.Runtime.Entities.dll) ## Syntax ``` vb 'Declaration <DataMemberAttribute> _ <ColumnAttribute("CHANNELREFERENCEID")> _ Public Property ChannelReferenceId As String Get Set 'Usage Dim instance As SalesTransaction Dim value As String value = instance.ChannelReferenceId instance.ChannelReferenceId = value ``` ``` csharp [DataMemberAttribute] [ColumnAttribute("CHANNELREFERENCEID")] public string ChannelReferenceId { get; set; } ``` ``` c++ [DataMemberAttribute] [ColumnAttribute(L"CHANNELREFERENCEID")] public: property String^ ChannelReferenceId { String^ get (); void set (String^ value); } ``` #### Property Value Type: [System.String](https://technet.microsoft.com/en-us/library/s1wwdcbf\(v=ax.60\)) Returns [String](https://technet.microsoft.com/en-us/library/s1wwdcbf\(v=ax.60\)). ## See Also #### Reference [SalesTransaction Class](salestransaction-class-microsoft-dynamics-commerce-runtime-datamodel.md) [Microsoft.Dynamics.Commerce.Runtime.DataModel Namespace](microsoft-dynamics-commerce-runtime-datamodel-namespace.md)
27.422535
146
0.787365
yue_Hant
0.800152
912db6eab825327c7e96e88ea5dd1ef514eb9007
1,171
md
Markdown
AlchemyInsights/plan-passwordless-deployment.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
1
2020-05-19T19:07:24.000Z
2020-05-19T19:07:24.000Z
AlchemyInsights/plan-passwordless-deployment.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
2
2022-02-09T06:52:18.000Z
2022-02-09T06:52:35.000Z
AlchemyInsights/plan-passwordless-deployment.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
1
2019-10-09T20:27:31.000Z
2019-10-09T20:27:31.000Z
--- title: Planowanie wdrożenia bez haseł ms.author: pebaum author: pebaum manager: scotv ms.date: 04/14/2021 audience: Admin ms.topic: article ms.service: o365-administration ROBOTS: NOINDEX, NOFOLLOW localization_priority: Priority ms.collection: Adm_O365 ms.custom: - "10394" - "9005762" ms.openlocfilehash: a167e33a5d9cb7900aabceda37466d731d784345 ms.sourcegitcommit: 8bc60ec34bc1e40685e3976576e04a2623f63a7c ms.translationtype: MT ms.contentlocale: pl-PL ms.lasthandoff: 04/15/2021 ms.locfileid: "51837434" --- # <a name="plan-your-passwordless-deployment"></a>Planowanie wdrożenia bez haseł Uaktualnianie do logowania alternatywnego przy użyciu metody uwierzytelniania bez użycia hasła. Weryfikacja dwuskładnikowa, logowanie telefoniczne lub generatory kodów mogą zwiększyć bezpieczeństwo i utrudnić innym osobom włamanie się na konto. Do wyboru są: funkcja Windows Hello dla firm, aplikacja uwierzytelniania firmy Microsoft lub klucze zabezpieczeń FIDO2. Aby ustalić, która metoda jest najlepsza i jak wdrożyć je na urządzeniu, skorzystaj z [przewodnika wdrażania bez haseł.](https://admin.microsoft.com/adminportal/home?#/modernonboarding/passwordlesssetup)
40.37931
324
0.824082
pol_Latn
0.993109
912dcc2e7a7029ce548e782c115d146f512c7804
237
md
Markdown
kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md
JonathanxD/Kores
236f7db6eeef7e6238f0ae0dab3f3b05fc531abb
[ "MIT-0", "MIT" ]
1
2019-04-16T10:42:02.000Z
2019-04-16T10:42:02.000Z
kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md
koresframework/Kores
b6ab31b1d376ab501fd9f481345c767cb0c37d04
[ "MIT-0", "MIT" ]
8
2020-12-12T06:48:34.000Z
2021-08-15T22:34:49.000Z
kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md
koresframework/Kores
b6ab31b1d376ab501fd9f481345c767cb0c37d04
[ "MIT-0", "MIT" ]
null
null
null
//[Kores](../../../index.md)/[com.github.jonathanxd.kores](../index.md)/[MutableInstructions](index.md)/[MutableInstructions](-mutable-instructions.md) # MutableInstructions [jvm]\ fun [MutableInstructions](-mutable-instructions.md)()
33.857143
151
0.734177
kor_Hang
0.255148
912dd20adea8785d1f902da7a972d6c78d75bfe8
2,966
md
Markdown
docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md
cawrites/sql-docs
58158eda0aa0d7f87f9d958ae349a14c0ba8a209
[ "CC-BY-4.0", "MIT" ]
2
2020-05-07T19:40:49.000Z
2020-09-19T00:57:12.000Z
docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md
cawrites/sql-docs
58158eda0aa0d7f87f9d958ae349a14c0ba8a209
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md
cawrites/sql-docs
58158eda0aa0d7f87f9d958ae349a14c0ba8a209
[ "CC-BY-4.0", "MIT" ]
2
2020-03-11T20:30:39.000Z
2020-05-07T19:40:49.000Z
--- title: "Lesson 3: Create a SQL Server Credential | Microsoft Docs" ms.custom: "" ms.date: "06/13/2017" ms.prod: "sql-server-2014" ms.reviewer: "" ms.technology: "database-engine" ms.topic: conceptual ms.assetid: 29e57ebd-828f-4dff-b473-c10ab0b1c597 author: MikeRayMSFT ms.author: mikeray manager: craigg --- # Lesson 3: Create a SQL Server Credential In this lesson, you will create a credential to store security information used to access the Azure storage account. A SQL Server credential is an object that is used to store authentication information required to connect to a resource outside of SQL Server. The credential stores the URI path of the storage container and the shared access signature key values. For each storage container used by a data or log file, you must create a SQL Server Credential whose name matches the container path. For general information about credentials, see [Credentials &#40;Database Engine&#41;](security/authentication-access/credentials-database-engine.md). > [!IMPORTANT] > The requirements for creating a SQL Server credential described below are specific to the [SQL Server Data Files in Azure](databases/sql-server-data-files-in-microsoft-azure.md) feature. For information on creating credentials for backup processes in Azure storage, see [Lesson 2: Create a SQL Server Credential](../tutorials/lesson-2-create-a-sql-server-credential.md). To create a SQL Server Credential, follow these steps: 1. Connect to SQL Server Management Studio. 2. In Object Explorer, connect to the instance of Database Engine installed. 3. On the Standard tool bar, click New Query. 4. Copy and paste the following example into the query window, modify as needed. The following statement will create a SQL Server Credential to store your storage container's Shared Access Certificate. ```sql USE master CREATE CREDENTIAL credentialname - this name should match the container path and it must start with https. WITH IDENTITY='SHARED ACCESS SIGNATURE', -- this is a mandatory string and do not change it. SECRET = 'sharedaccesssignature' -- this is the shared access signature key that you obtained in Lesson 2. GO ``` For detailed information, see [CREATE CREDENTIAL &#40;Transact-SQL&#41;](/sql/t-sql/statements/create-credential-transact-sql) in SQL Server Books Online. 5. To see all available credentials, you can run the following statement in the query window: ```sql SELECT * from sys.credentials ``` For more information on sys.credentials, see [sys.credentials &#40;Transact-SQL&#41;](/sql/relational-databases/system-catalog-views/sys-credentials-transact-sql) in SQL Server Books Online. **Next Lesson:** [Lesson 4: Create a database in Azure Storage](lesson-3-database-backup-to-url.md)
50.271186
384
0.726231
eng_Latn
0.964667
912e13cd99b8478e64737398e0079e8381342732
622
md
Markdown
src/pages/companies/2018-02-13-supermeat.md
arvenjadeaguilar/cellagri-cms
f177977e4d859f540ed1fc455594629b30f42bd2
[ "MIT" ]
null
null
null
src/pages/companies/2018-02-13-supermeat.md
arvenjadeaguilar/cellagri-cms
f177977e4d859f540ed1fc455594629b30f42bd2
[ "MIT" ]
null
null
null
src/pages/companies/2018-02-13-supermeat.md
arvenjadeaguilar/cellagri-cms
f177977e4d859f540ed1fc455594629b30f42bd2
[ "MIT" ]
null
null
null
--- templateKey: company-post path: /supermeat date: '2018-02-15T11:00:00-05:00' title: SuperMeat location: 'Tel Aviv, Israel' website: supermeat.com socialMedia: - media: Twitter url: 'https://twitter.com/_SuperMeat_' logo: /img/supermeat logo.jpg thumbnail: /img/supermeat logo.jpg description: >- SuperMeat is a cellular agriculture startup based in Tel Aviv, Israel. SuperMeat was founded by Ido Savir, Koby Barak and Shir Friedman in 2015 to make cultured chicken meat. Through cellular agriculture, SuperMeat hopes to provides a sustainable and healthy source of chicken meat for processed foods. ---
31.1
80
0.765273
eng_Latn
0.809162
912e708b1c2bf65f27991b730490981a9fed93a4
1,352
md
Markdown
catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md
htron-dev/baka-db
cb6e907a5c53113275da271631698cd3b35c9589
[ "MIT" ]
3
2021-08-12T20:02:29.000Z
2021-09-05T05:03:32.000Z
catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md
zzhenryquezz/baka-db
da8f54a87191a53a7fca54b0775b3c00f99d2531
[ "MIT" ]
8
2021-07-20T00:44:48.000Z
2021-09-22T18:44:04.000Z
catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md
zzhenryquezz/baka-db
da8f54a87191a53a7fca54b0775b3c00f99d2531
[ "MIT" ]
2
2021-07-19T01:38:25.000Z
2021-07-29T08:10:29.000Z
# Boku no Mama-chan (43) Kaihatsu Nikki ![boku-no-mama-chan-43-kaihatsu-nikki](https://cdn.myanimelist.net/images/manga/2/190695.jpg) - **type**: manga - **volumes**: 1 - **chapters**: 6 - **original-name**: 僕のママちゃん(43)開発日記 - **start-date**: 2016-10-19 ## Tags - yaoi ## Authors - Neko Hakushaku (Story & Art) ## Sinopse When seasoned actor Mamoru "Mama-chan" Maeda first stepped onto the set of his new crime drama, Dark Panther, he thought the only part he would have to play was that of experienced detective Inspector Kinoda. However, his rather persistent co-star, Christopher Shirasu, has other ideas. Shirasu is an avid follower of fan culture, spending hours scrolling through his social media tags, looking at art created by his fans, and obsessing over reading fanfiction. After spending so much time engaging with the desires of the drama's most enthusiastic demographic, Shirasu has become rather strict with Maeda about his expectations—he needs to give fans what they want. Ensuring that the two toe the perfect line between close friendship and homosexual romance, Shirasu is determined to make sure that the hashtag "#shiramama" will continue to trend—regardless of any "ship wars." [Source My Anime List] ## Links - [My Anime list](https://myanimelist.net/manga/103294/Boku_no_Mama-chan_43_Kaihatsu_Nikki)
42.25
379
0.756657
eng_Latn
0.99339
912ea325833ebaae5e7ab8b9a6dacec03587355c
813
md
Markdown
README.md
Phocacius/kmeans
3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d
[ "MIT" ]
null
null
null
README.md
Phocacius/kmeans
3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d
[ "MIT" ]
null
null
null
README.md
Phocacius/kmeans
3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d
[ "MIT" ]
null
null
null
# K-Means Simulator Provides a step-by-step visualisation of the k-means algorithm for unsupervised clustering of 2D Data. Click on the canvas to add points and choose your desired number of clusters (k). You can then run the algorithm step-by-step manually using "Assign Clusters" and "Recenter Centroids" or automatically where there's a half second delay between each step using "Autostart". "Place centroids randomly" resets the clusters and centroids. ![](screenshot.png) ## Required Software for Development The javascript files are compiled using browserify. - [NodeJS](https://nodejs.org/en/download/) - [Browserify](http://browserify.org/): `npm install -g browserify` ## Building - install dependencies (currently only jquery): `npm install` - compile JS: `browserify js/main.js -o js/bundle.js`
45.166667
436
0.771218
eng_Latn
0.968846
912f240e2130569d98515738df087ed19fd5f0ee
476
md
Markdown
guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md
SweeneyNew/freeCodeCamp
e24b995d3d6a2829701de7ac2225d72f3a954b40
[ "BSD-3-Clause" ]
10
2019-08-09T19:58:19.000Z
2019-08-11T20:57:44.000Z
guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md
SweeneyNew/freeCodeCamp
e24b995d3d6a2829701de7ac2225d72f3a954b40
[ "BSD-3-Clause" ]
2,056
2019-08-25T19:29:20.000Z
2022-02-13T22:13:01.000Z
guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md
SweeneyNew/freeCodeCamp
e24b995d3d6a2829701de7ac2225d72f3a954b40
[ "BSD-3-Clause" ]
5
2018-10-18T02:02:23.000Z
2020-08-25T00:32:41.000Z
--- title: Use Destructuring Assignment to Assign Variables from Arrays localeTitle: استخدم Destructuring Assignment لتعيين متغيرات من صفائف --- ## استخدم Destructuring Assignment لتعيين متغيرات من صفائف علينا اتخاذ بعض الاحتياطات في هذه الحالة. 1. لا حاجة للثابتة \[ب ، أ\] لأنها ستحافظ على تأثير الواجب المحلي. 2. const \[b، a\] = \[a، b\] سيؤدي إلى قيمة a ، مثل b غير محددة (قاعدة تخصيص بسيطة من اليسار إلى اليمين). ومن هنا حل هذه المشكلة \[b، a\] = \[a، b\]
34
106
0.707983
arb_Arab
0.997059
912f4e0ad3415be2311ea7f9f504919fd6bc3f70
2,554
md
Markdown
README.md
breglerj/cloud-foundry-tools-api
883de7da0c233fabc366c86b62c72a2720139295
[ "Apache-2.0" ]
null
null
null
README.md
breglerj/cloud-foundry-tools-api
883de7da0c233fabc366c86b62c72a2720139295
[ "Apache-2.0" ]
null
null
null
README.md
breglerj/cloud-foundry-tools-api
883de7da0c233fabc366c86b62c72a2720139295
[ "Apache-2.0" ]
null
null
null
![GitHub package.json version](https://img.shields.io/github/package-json/v/SAP/cloud-foundry-tools-api) [![CircleCI](https://circleci.com/gh/SAP/cloud-foundry-tools-api.svg?style=svg)](https://circleci.com/gh/SAP/cloud-foundry-tools-api) [![Coverage Status](https://coveralls.io/repos/github/SAP/cloud-foundry-tools-api/badge.svg?branch=master)](https://coveralls.io/github/SAP/cloud-foundry-tools-api?branch=master) [![Language grade: JavaScript](https://img.shields.io/lgtm/grade/javascript/g/SAP/cloud-foundry-tools-api.svg?logo=lgtm&logoWidth=18)](https://lgtm.com/projects/g/SAP/cloud-foundry-tools-api/context:javascript) [![Commitizen friendly](https://img.shields.io/badge/commitizen-friendly-brightgreen.svg)](http://commitizen.github.io/cz-cli/) [![REUSE status](https://api.reuse.software/badge/github.com/SAP/cloud-foundry-tools-api)](https://api.reuse.software/info/github.com/SAP/cloud-foundry-tools-api) [![dependentbot](https://api.dependabot.com/badges/status?host=github&repo=SAP/cloud-foundry-tools-api)](https://dependabot.com/) # Overview This package provides a set of APIs to help you develop applications in Cloud Foundry. You can use these APIs to manage apps, service instances, orgs, spaces, and users in your environment. Mostly, this is a wrapper of the CF command line client, which runs a particular command and parses the output to the suitable JSON file. If an error or failure occurs, the runtime exception throws with relevant problem information. ## Prerequisite _from version_ `v1.0.0` _and above_: Make sure you have installed the CF CLI [v7](https://github.com/cloudfoundry/cli/blob/master/doc/installation-instructions/installation-instructions-v7.md#installers-and-compressed-binaries) tool in your environment. _below version_ `v1.0.0`: Make sure you have installed the CF CLI [v6](https://github.com/cloudfoundry/cli/blob/master/doc/installation-instructions/installation-instructions-v6.md#installers-and-compressed-binaries) tool in your environment. ## Examples of usage Example 1 ``` try { const result = await cfLogin("https://api.cf.....com", "user", "password"); if (result === "OK") { // successful } } catch (e) { // display or/and log error } ``` Example 2 ``` try { const spaces = await cfGetAvailableSpaces("myOrg"); for (const space of spaces) { console.log("Space label is " + space.label + " guid is " + space.guid); } } catch (e) { // display or/and log error } ``` ## Contributing Contributing information can be found in the [CONTRIBUTING.md](CONTRIBUTING.md) file.
49.115385
423
0.755286
eng_Latn
0.619575
912fa5fd54b2114e2cd8ca1fdf098ceebabbb7ad
6,151
md
Markdown
README.md
DazaSeal/wrapper-infinite
7079f33791246d023ac204a42f89f1b9d42b39fc
[ "MIT" ]
null
null
null
README.md
DazaSeal/wrapper-infinite
7079f33791246d023ac204a42f89f1b9d42b39fc
[ "MIT" ]
null
null
null
README.md
DazaSeal/wrapper-infinite
7079f33791246d023ac204a42f89f1b9d42b39fc
[ "MIT" ]
1
2022-03-12T15:00:11.000Z
2022-03-12T15:00:11.000Z
## Wrapper: Infinite Wrapper: Infinite is a project **(unofficially)** carrying on the torch of [VisualPlugin's Wrapper project](https://github.com/GoAnimate-Wrapper) after it's shutdown in 2020. Unlike the original project, Infinite can not be shut down by Vyond. Why? It's because of our twist on the Wrapper formula! Everything is hosted and programmed to function locally, no internet access required, barring the initial download (and TTS at this time). This project is important for archival purposes, as the ability to use this legacy editor and themes would be irritating and unproductive without it. Besides the main feature of Infinite, we have also added other additional features and improvements compared to the original Wrapper project, and there will be more to come, as Infinite will continue to get additional updates in the future. ## Running / Installation To start Infinite on Windows, open start_wrapper.bat. It'll automate just about everything for you and, well, start Wrapper: Infinite. On your first run, you will likely need to right-click it and click "Run as Administrator". This allows it to properly install what it needs to run. After your initial run, you shouldn't need to do that again, you can start it as normal. If you want to import videos and characters from the original Wrapper or any other clones of it, open its folder and drag the "_SAVED" folder into Infinite's "wrapper" folder. If you have already made any videos or characters, this will not work. Please only import on a new install with no saved characters or videos, or take the "_SAVED" folder in Infinite out before dragging the old one in. If you want to import character IDs from the original LVM, you can paste `&original_asset_id=[ID HERE]` at the end of the link for the matching character creator. ## Updates & Support ### This information is outdated and does not pertain to Wrapper: Infinite. For support, the first thing you should do is read through faq.txt, it most likely has what you want to know. If you can't find what you need, you can join the [Discord server](https://discord.gg/Kf7BzSw). Joining the server is recommended, as there is a whole community to help you out. If you don't use Discord, you can email mailbenson@protonmail.com to get in contact with me directly, but don't expect nearly as quick of a response. ## Dependencies This program relies on Flash and Node.js to work properly. SilentCMD is also used to suppress all the extra logging noise you'd only need for troubleshooting and development. These all have been included with the project (utilities folder) to ensure full offline operation and will be installed if missing. The "wrapper" folder has it's own dependencies, but they are included as well. ## License Most of this project is free/libre software[1] under the MIT license. You have the freedom to run, change, and share this as much as you want. This includes: - Files in the "wrapper" folder - Batch files included with Wrapper: Infinite - Node.js - SilentCMD - Chromium Web Store ungoogled-chromium is under the BSD 3-Clause license, which grants similar rights, but has some differences from MIT. MediaInfo has a similar BSD 2-Clause license. 7zip's license is mostly LGPL, but some parts are under the BSD 3-clause License, and some parts have an unRAR restriction. Stylus is under the GNU GPLv3 license. These licenses can be found in each program's folder in utilities\sourcecode. The source code for compiled programs are all stored in utilities\sourcecode, and you can modify these as you wish. Parts of Infinite that run from their source code directly (such as batch scripts) are not included in that folder, for obvious reasons. Flash Player (utilities folder) and GoAnimate's original assets (server folder) are proprietary and do not grant you these rights, but if they did, this project wouldn't need to exist. Requestly, an addon included in Infinite's browser, is sadly proprietary software, but you're free to remove the Chromium profile and use a fresh one if this bothers you. Requestly is primarily included because of how popular it is with our community. While completely unnecessary, if you decide to use your freedom to change the software, it would be greatly appreciated if you sent it to me so I can implement it into the main program! With credit down here of course :) ## Credits **Please do not contact anyone on the list for support, use the Discord server.** Original Wrapper credits: | Name | Contribution | | ------------ | -------------------- | | VisualPlugin | GoAnimate Wrapper | | xomdjl_ | Custom/Modded Assets | | CLarramore | Bug fixes | | PoleyMagik | Asset Store Archive | No members of the original team are officially working on Infinite, even if they have contributed. Some members of the original team have asked to not be given credit, and they have been removed. Developers will have (D) on their name, to show they are officially working on W:I. Everyone else has simply contributed something at some point, or left the development team. Wrapper: Infinite credits: | Name | Contribution | | ------------- | -------------------- | | Octanuary | Wrapper: Infinite | | Daza (D) | Improved CSS | These are unaffiliated people that they haven't directly done anything for the project (and probably don't even know it exists) but still deserve credit for their things. Kinda like a shoutout but in a project's readme. ***Please do not contact them about Wrapper: Infinite.*** | Name | Contribution | | ------------------ | -------------------------------- | | Vyond | Creators of the themes we love | | Stephan Brenner | Creator of SilentCMD | | vocatus | Some of TronScript's batch code | | ss64.com | Incredible CMD info resource | | robvanderwoude.com | Also amazing CMD info resource | | VisualPlugin | Text to speech voices | ## Footnotes [1] - See <https://www.gnu.org/philosophy/free-sw.html> for a better definition of free software.
93.19697
828
0.738579
eng_Latn
0.999434
912fe6ae1c7ab206900cc994d8593fa76c63f95f
1,547
md
Markdown
results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md
eliMakeouthill/AutoEq
b16c72495b3ce493293c6a4a4fdf45a81aec9ca0
[ "MIT" ]
3
2022-02-25T08:33:08.000Z
2022-03-13T11:27:29.000Z
results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md
billclintonwong/AutoEq
aa25ed8e8270c523893fadbda57e9811c65733f1
[ "MIT" ]
null
null
null
results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md
billclintonwong/AutoEq
aa25ed8e8270c523893fadbda57e9811c65733f1
[ "MIT" ]
null
null
null
# DUNU DN16 Hephaes See [usage instructions](https://github.com/jaakkopasanen/AutoEq#usage) for more options and info. ### Parametric EQs In case of using parametric equalizer, apply preamp of **-7.5dB** and build filters manually with these parameters. The first 5 filters can be used independently. When using independent subset of filters, apply preamp of **-7.5dB**. | Type | Fc | Q | Gain | |:--------|:---------|:-----|:--------| | Peaking | 26 Hz | 0.14 | -7.8 dB | | Peaking | 289 Hz | 0.41 | -9.7 dB | | Peaking | 1149 Hz | 0.26 | 9.0 dB | | Peaking | 5784 Hz | 1.05 | -6.7 dB | | Peaking | 11495 Hz | 0.81 | 7.3 dB | | Peaking | 3005 Hz | 4.4 | 2.2 dB | | Peaking | 3494 Hz | 3.64 | -1.8 dB | | Peaking | 9276 Hz | 6.08 | 1.3 dB | | Peaking | 15919 Hz | 1.97 | 4.7 dB | | Peaking | 20191 Hz | 0.63 | -9.7 dB | ### Fixed Band EQs In case of using fixed band (also called graphic) equalizer, apply preamp of **-7.9dB** (if available) and set gains manually with these parameters. | Type | Fc | Q | Gain | |:--------|:---------|:-----|:--------| | Peaking | 31 Hz | 1.41 | -8.4 dB | | Peaking | 62 Hz | 1.41 | -6.6 dB | | Peaking | 125 Hz | 1.41 | -7.9 dB | | Peaking | 250 Hz | 1.41 | -7.6 dB | | Peaking | 500 Hz | 1.41 | -1.2 dB | | Peaking | 1000 Hz | 1.41 | 5.0 dB | | Peaking | 2000 Hz | 1.41 | 6.6 dB | | Peaking | 4000 Hz | 1.41 | -0.6 dB | | Peaking | 8000 Hz | 1.41 | 2.4 dB | | Peaking | 16000 Hz | 1.41 | 6.0 dB | ### Graphs ![](./DUNU%20DN16%20Hephaes.png)
38.675
98
0.553975
eng_Latn
0.746417
912ff8cf5d480d6cfe9dc42d772158a1b1b9c609
513
md
Markdown
_course_files/Chapter 6 Services/Step.md
boriphuth/k8s-fleetman
5bc31a72d7fadd7f0550d0390fcab9525055663b
[ "MIT" ]
1
2019-10-23T09:14:35.000Z
2019-10-23T09:14:35.000Z
_course_files/Chapter 6 Services/Step.md
boriphuth/k8s-fleetman
5bc31a72d7fadd7f0550d0390fcab9525055663b
[ "MIT" ]
null
null
null
_course_files/Chapter 6 Services/Step.md
boriphuth/k8s-fleetman
5bc31a72d7fadd7f0550d0390fcab9525055663b
[ "MIT" ]
null
null
null
$ kubectl apply -f first-pod.yaml $ kubectl apply -f webapp-service.yaml ## Connect to Service $ minikube service fleetman-webapp $ kubectl get po --show-labels NAME READY STATUS RESTARTS AGE LABELS webapp 1/1 Running 0 92m app=webapp,release=0 webapp-release-0-5 1/1 Running 0 48m app=webapp,release=0-5 $ kubectl get po --show-labels -l release=0 $ kubectl get po --show-labels -l release=1 No resources found in default namespace.
36.642857
78
0.647173
yue_Hant
0.764128
9130714c260b6a6b44a06f07c4f320519420ba98
5,310
md
Markdown
articles/cognitive-services/Content-Moderator/review-api.md
changeworld/azure-docs.nl-nl
bdaa9c94e3a164b14a5d4b985a519e8ae95248d5
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Content-Moderator/review-api.md
changeworld/azure-docs.nl-nl
bdaa9c94e3a164b14a5d4b985a519e8ae95248d5
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Content-Moderator/review-api.md
changeworld/azure-docs.nl-nl
bdaa9c94e3a164b14a5d4b985a519e8ae95248d5
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Recensies, werk stromen en taken concepten-Content Moderator titleSuffix: Azure Cognitive Services description: In dit artikel vindt u meer informatie over de basis concepten van het hulp programma voor beoordeling. Beoordelingen, werk stromen en taken. services: cognitive-services author: PatrickFarley manager: nitinme ms.service: cognitive-services ms.subservice: content-moderator ms.topic: conceptual ms.date: 03/14/2019 ms.author: pafarley ms.openlocfilehash: 1aba86efb9ea76fbf060e80b47f9f2f6cdf8ee71 ms.sourcegitcommit: 7b25c9981b52c385af77feb022825c1be6ff55bf ms.translationtype: MT ms.contentlocale: nl-NL ms.lasthandoff: 03/13/2020 ms.locfileid: "79221148" --- # <a name="content-moderation-reviews-workflows-and-jobs"></a>Beoordelingen, werk stromen en taken voor content toezicht Content Moderator combineert de door de machine ondersteunde toezicht met de mogelijkheden van Human-in-the-loop om een optimaal toezicht proces te maken voor praktijk scenario's. Dit wordt gedaan via het [beoordelings programma](https://contentmoderator.cognitive.microsoft.com)op basis van de Cloud. In deze hand leiding vindt u informatie over de belangrijkste concepten van het hulp programma voor beoordeling: Recensies, werk stromen en taken. ## <a name="reviews"></a>Beoordelingen Bij een beoordeling wordt inhoud geüpload naar het hulp programma beoordeling en wordt deze weer gegeven onder het tabblad **controleren** . Hier kunnen gebruikers de toegepaste labels wijzigen en hun eigen aangepaste labels Toep assen. Wanneer een gebruiker een beoordeling verzendt, worden de resultaten verzonden naar een opgegeven eind punt van de retour aanroep en wordt de inhoud verwijderd van de site. ![De website voor het controleren van het hulp programma open in een browser op het tabblad controleren](./Review-Tool-user-Guide/images/image-workflow-review.png) Raadpleeg de [hand leiding voor het beoordelings programma](./review-tool-user-guide/review-moderated-images.md) om te beginnen met het maken van recensies of raadpleeg de [rest API gids](./try-review-api-review.md) voor informatie over hoe u dit kunt doen. ## <a name="workflows"></a>Werkstromen Een werk stroom is een op de cloud gebaseerd gepersonaliseerd filter voor inhoud. Werk stromen kunnen verbinding maken met verschillende services om inhoud op verschillende manieren te filteren en vervolgens de juiste actie ondernemen. Met de Content Moderator-connector kunt u met een werk stroom automatisch toezicht Tags Toep assen en beoordelingen maken met verzonden inhoud. ### <a name="view-workflows"></a>Werk stromen weer geven Als u uw bestaande werk stromen wilt weer geven, gaat u naar het [hulp programma controleren](https://contentmoderator.cognitive.microsoft.com/) en selecteert u **instellingen** > **werk stromen**. ![Standaardwerk stroom](images/default-workflow-listed.PNG) Werk stromen kunnen volledig worden beschreven als JSON-teken reeksen, waardoor ze via een programma toegankelijk zijn. Als u de optie **bewerken** selecteert voor uw werk stroom en vervolgens het tabblad **JSON** selecteert, ziet u een JSON-expressie zoals het volgende: ```json { "Type": "Logic", "If": { "ConnectorName": "moderator", "OutputName": "isAdult", "Operator": "eq", "Value": "true", "Type": "Condition" }, "Then": { "Perform": [ { "Name": "createreview", "CallbackEndpoint": null, "Tags": [] } ], "Type": "Actions" } } ``` Raadpleeg de [hand leiding voor het hulp programma](./review-tool-user-guide/workflows.md) voor het maken en gebruiken van werk stromen of raadpleeg de [rest API gids](./try-review-api-workflow.md) voor informatie over hoe u dit kunt doen met programma code. ## <a name="jobs"></a>Taken Een toezicht taak fungeert als een soort wrapper voor de functionaliteit van toezicht op inhoud, werk stromen en Beoordelingen. De taak scant uw inhoud met behulp van de Content Moderator afbeeldings toezicht-API of tekst toezicht-API en controleert deze vervolgens op de aangewezen werk stroom. Op basis van de werk stroom resultaten kan de inhoud niet worden gecontroleerd in het [beoordelings programma](./review-tool-user-guide/human-in-the-loop.md). Hoewel zowel beoordelingen als werk stromen kunnen worden gemaakt en geconfigureerd met hun respectieve Api's, kunt u met de taak-API een gedetailleerd rapport verkrijgen van het hele proces (dat kan worden verzonden naar een opgegeven call back-eind punt). Raadpleeg de [rest API-hand leiding](./try-review-api-job.md) om aan de slag te gaan met taken. ## <a name="next-steps"></a>Volgende stappen * Test de [taak-API-console](try-review-api-job.md)en gebruik de rest API code voorbeelden. Als u bekend bent met Visual Studio en C#raadpleegt u ook de [taken .net Quick](moderation-jobs-quickstart-dotnet.md)start. * Ga voor recensies naar aan de slag met de [API-console voor controle](try-review-api-review.md)en gebruik de rest API code voorbeelden. Zie de sectie Recensies van [.net Quick](dotnet-sdk-quickstart.md)start. * Gebruik voor video revisies de [Snelstartgids voor video revisie](video-reviews-quickstart-dotnet.md)en leer hoe u [Transcripten kunt toevoegen aan de video beoordeling](video-transcript-reviews-quickstart-dotnet.md).
66.375
712
0.775141
nld_Latn
0.999047
91308e7fdc37435df478c923542ea311fc37b53f
7,397
md
Markdown
business-central/readiness/readiness-learning-sales.md
MicrosoftDocs/dynamics365smb-docs-pr.nb-no
f57ffe1865b515a2240b7e4d1401263a33d2a535
[ "CC-BY-4.0", "MIT" ]
2
2020-05-18T17:20:08.000Z
2021-04-20T21:13:47.000Z
business-central/readiness/readiness-learning-sales.md
MicrosoftDocs/dynamics365smb-docs-pr.nb-no
f57ffe1865b515a2240b7e4d1401263a33d2a535
[ "CC-BY-4.0", "MIT" ]
null
null
null
business-central/readiness/readiness-learning-sales.md
MicrosoftDocs/dynamics365smb-docs-pr.nb-no
f57ffe1865b515a2240b7e4d1401263a33d2a535
[ "CC-BY-4.0", "MIT" ]
2
2019-10-12T19:50:37.000Z
2020-09-30T16:51:21.000Z
--- title: Læringskatalog for salg og markedsføring for partner description: Finn alle tilgjengelige læringsressurser for salg og markedsføringsroller for partner i Business Central. author: loreleishannonmsft ms.date: 04/01/2021 ms.topic: conceptual ms.author: margoc ms.openlocfilehash: 9830e4e842cc7fe3febcbe809547ad26701f527d ms.sourcegitcommit: 766e2840fd16efb901d211d7fa64d96766ac99d9 ms.translationtype: HT ms.contentlocale: nb-NO ms.lasthandoff: 03/31/2021 ms.locfileid: "5781841" --- # <a name="partner-sales-and-marketing-learning-catalog"></a>Læringskatalog for salg og markedsføring for partner Er du ansvarlig for å hjelpe kundene med å kjøpe riktig løsning? Følgende katalog er ordnet fra kjernekunnskap til bestemte domener, og fra de mest grunnleggende til de fleste avanserte. Hvis innholdet finnes i flere formater, gir vi deg beskjed slik at du kan velge det opplæringsformatet som best dekker dine behov. ## <a name="business-value"></a>Forretningsverdi<a name="busvalue"></a> | Innhold | Beskrivelse | Format | Lengde | |-------------|----------------|---------------------------------------|---------------------| | [Dynamics 365 Business Central (Blitz)](https://mbspartner.microsoft.com/D365/Videos/101760) | (Bare partnere) Denne økten vil lage en disposisjon av verdien for Business Central. Business Central tilbyr god verdi fordi den er integrert med andre Microsoft-skytjenester, inkludert Microsoft 365, og kan tilpasses eller utvides for bestemte bransjebehov med Power Apps, Power Automate og Power BI. | Video | 8 minutter 5 sekunder | | [Utforske Dynamics 365 Business Central-funksjoner](https://dynamics.microsoft.com/business-central/capabilities/) | Utforske Dynamics 365 Business Central-funksjoner | Nettsted | | | [Dynamics 365 Business Central](/learn/modules/dynamics-365-business-central/) | Denne modulen viser programfunksjoner, fordeler og eksempler på Dynamics 365 Business Central i bruk. | Gratis, elektronisk opplæringsbane i eget tempo | 24 minutter | ## <a name="getting-started"></a>Komme i gang<a name="get-started"></a> | Innhold | Beskrivelse | Format | Lengde | |-----------|---------------|---------------------------------------|---------------------| | [Kom i gang med Microsoft Dynamics 365 Business Central](/learn/paths/get-started-dynamics-365-business-central/) | Er du interessert i Business Central? Hvis dette er tilfellet, får du en god innføring i denne opplæringsbanen. Den drøfter hvordan du konfigurerer en prøveversjon, gir grunnleggende informasjon om hva som er tilgjengelig, og viser noen av tilpasningsalternativene du kan bruke for å tilpasse Business Central for egen bruk. | Gratis, elektronisk opplæringsbane i eget tempo | 3 timer 4 minutter | | [Arbeide med Microsoft Dynamics 365 Business Central-brukergrensesnittet](/learn/paths/work-with-user-interface-dynamics-365-business-central/) | Vil du lære om brukergrensesnittet for Business Central? Denne opplæringsbanen viser brukergrensesnittet, og også hvordan du kan tilpasse det. Det drøfter også alternative grensesnitt, for eksempel en smarttelefon. | Gratis, elektronisk opplæringsbane i eget tempo | 2 timer 27 minutter | | [Arbeid som en profesjonell med data i Business Central](/learn/paths/work-pro-data-dynamics-365-business-central) | Har du nettopp kommet i gang med Business Central? Denne opplæringsbanen gjør det enklere for deg å forberede deg til datatypene i Business Central og finne informasjon du trenger i programmet. | Gratis, elektronisk opplæringsbane i eget tempo | 2 timer 27 minutter | | [Business Central YouTube-spilliste](https://www.youtube.com/playlist?list=PLcakwueIHoT-wVFPKUtmxlqcG1kJ0oqq4) | YouTube-målside for Dynamics 365 Business Central-videoer | YouTube-videoer | | ## <a name="grow-your-business"></a>Utvikle bedriften<a name="grow"></a> | Innhold | Beskrivelse | Format | Lengde | |-----------|--------------------|---------------------------------------|-----------------------| | [Dynamics 365 Business Central - Overbevisende samtaler med kundeemner (Blitz)](https://mbspartner.microsoft.com/D365/Videos/101761) | (Bare partnere) Bli med Errol Schoenfish, direktøren for produktmarkedsføring, mens han snakker om hvordan du kan ha overbevisende samtaler med kundeemner på Microsoft Dynamics 365 Business Central. | Video | 11 minutter 15 sekunder | | [Program for leverandør av skytjenester](https://partner.microsoft.com/membership/cloud-solution-provider) | (Bare partnere) Om program for leverandør av skyløsninger | Nettsted | | | [Markedsføring i en skyverden](https://mbspartner.microsoft.com/OSS/Topic/19) | (Bare partner) Hvordan gjøre det mulig å optimalisere markedsføringspraksisen og justere markedsføringsprinsippene for en skytjeneste er en primær faktor for å finne ut hvor vellykket eller mislykket organisasjonen går fremover. Denne serien med videoer og ressurser inneholder gode fremgangsmåter for markedsføring, verktøy for oppretting av innhold, nyttige tips og verktøy for å bruke prinsipper for markedsføring i skyen til å skape økt etterspørsel etter løsningene dine. | Emneside | 35 minutter | | [Program for leverandør av skyløsninger – selge etterspurte skyløsninger](/partner-center/csp-overview) | Programmet for leverandør av skyløsninger (CSP) hjelper deg med å gå lenger enn å videreselge lisenser til å være mer involvert i kundens virksomhet. | Nettsted | | | [Program for leverandør av skyløsninger - sammenligne partnerstøtteprogrammer](https://partner.microsoft.com/support/partnersupport) | Se denne videoen for å se hvordan du kan velge den riktige planen for partnerstøtte for bedriften. Deretter ser du en side ved side-sammenligning nedenfor. | Nettsted | | | [Cloud SureStep-akselerert salg](https://mbspartner.microsoft.com/OSS/Topic/20) | (Bare partnere) Prosessen Cloud SureStep SureStep-akselerert salg er bygget rundt behovene til moderne kjøpere og søker å utløse en emosjonell, erfaringsdrevet strategisk vendepunkt tidlig i salgsprosessen. Denne serien med videoer og ressurser forklarer en normativ salgsprosess som er spesielt utformet for å motvirke den tradisjonelle ressursintensive, demoorienterte tilnærmingen til salg som vanligvis fører til svært tilpassede produktdemonstrasjoner, komplekse salgssykluser og lange implementeringsprosjekter. Ressursblioteket omfatter en rekke tilpassbare maler, slik at du kan opprette ditt eget sett med salgsressurser for å støtte akselerert salg. | Nettsted | | | [Tillegg til strategiplanen for overgangspartner (juli 2019)](https://assetsprod.microsoft.com/d365-business-central-supplement-to-transition-partner-playbook.pptx) | Dette tillegget til strategiplanen for overgangspartner gir oppdatert og ny veiledning for partnere for å sikre den eksisterende kundebasen mot konkurransen og nye SaaS-konkurrenter. | PowerPoint-lysbildefremvisning | 15 lysbilder | [!INCLUDE[footer-include](../includes/footer-banner.md)]
145.039216
785
0.723807
nob_Latn
0.982268
91317037239b34b820e8c646595604362f7e4e07
5,130
md
Markdown
basics/plugin_structure/plugin_configuration_file.md
pettermahlen/intellij-sdk-docs
1fd57016f2bf34afb3277d7094da091e0d84876b
[ "Apache-2.0" ]
1
2021-08-18T09:44:04.000Z
2021-08-18T09:44:04.000Z
basics/plugin_structure/plugin_configuration_file.md
pettermahlen/intellij-sdk-docs
1fd57016f2bf34afb3277d7094da091e0d84876b
[ "Apache-2.0" ]
null
null
null
basics/plugin_structure/plugin_configuration_file.md
pettermahlen/intellij-sdk-docs
1fd57016f2bf34afb3277d7094da091e0d84876b
[ "Apache-2.0" ]
null
null
null
--- title: Plugin Configuration File - plugin.xml --- The following is a sample plugin configuration file. This sample showcases and describes all elements that can be used in the plugin.xml file. ```xml <!-- url="" specifies the URL of the plugin homepage (displayed in the Welcome Screen and in "Plugins" settings dialog) --> <idea-plugin url="http://www.jetbrains.com/idea"> <!-- Plugin name --> <name>VssIntegration</name> <!-- Unique identifier of the plugin. Cannot be changed between the plugin versions. If not specified, assumed to be equal to <name>. --> <id>VssIntegration</id> <!-- Description of the plugin. --> <description>Vss integration plugin</description> <!-- Description of changes in the latest version of the plugin. Displayed in the "Plugins" settings dialog and in the plugin repository Web interface. --> <change-notes>Initial release of the plugin.</change-notes> <!-- Plugin version --> <version>1.0</version> <!-- The vendor of the plugin. The optional "url" attribute specifies the URL of the vendor homepage. The optional "email" attribute specifies the e-mail address of the vendor.--> <vendor url="http://www.jetbrains.com" email="support@jetbrains.com" /> <!-- The unique identifiers of the plugins on which this plugin depends. --> <depends>MyFirstPlugin</depends> <!-- Optional dependency on another plugin. If the plugin with the "MySecondPlugin" ID is installed, the contents of mysecondplugin.xml (the format of this file conforms to the format of plugin.xml) will be loaded. --> <depends optional="true" config-file="mysecondplugin.xml">MySecondPlugin</depends> <!-- Allows a plugin to integrate its help system (in JavaHelp format) with the IDEA help system. The "file" attribute specifies the name of the JAR file in the "help" subdirectory of the plugin directory. The "path" attribute specifies the name of the helpset file within the JAR file.--> <helpset file="myhelp.jar" path="/Help.hs" /> <!-- Minimum and maximum build of IDEA compatible with the plugin --> <idea-version since-build="3000" until-build="3999"/> <!-- Resource bundle from which the text of plugin descriptions, action names and etc. will be loaded --> <resource-bundle>messages.MyPluginBundle</resource-bundle> <!-- Plugin's application components --> <application-components> <component> <!-- Component's interface class --> <interface-class>com.foo.Component1Interface</interface-class> <!-- Component's implementation class --> <implementation-class>com.foo.impl.Component1Impl</implementation-class> </component> </application-components> <!-- Plugin's project components --> <project-components> <component> <!-- Interface and implementation classes are the same --> <interface-class>com.foo.Component2</interface-class> <!-- If the "workspace" option is set "true", the component saves its state to the .iws file instead of the .ipr file. Note that the <option> element is used only if the component implements the JDOMExternalizable interface. Otherwise, the use of the <option> element takes no effect. --> <option name="workspace" value="true" /> <!-- If the "loadForDefaultProject" tag is present, the project component is instantiated also for the default project. --> <loadForDefaultProject> </component> </project-components> <!-- Plugin's module components --> <module-components> <component> <interface-class>com.foo.Component3</interface-class> </component> </module-components> <!-- Actions --> <actions> <action id="VssIntegration.GarbageCollection" class="com.foo.impl.CollectGarbage" text="Collect _Garbage" description="Run garbage collector"> <keyboard-shortcut first-keystroke="control alt G" second-keystroke="C" keymap="$default"/> </action> </actions> <!-- Extension points defined by the plugin. Extension points are registered by a plugin so that other plugins can provide this plugin with certain data. The "beanClass" attribute specifies the class the implementations of which can be used for the extension point. --> <extensionPoints> <extensionPoint name="testExtensionPoint" beanClass="com.foo.impl.MyExtensionBean"/> </extensionPoints> <!-- Extensions which the plugin adds to extension points defined by the IDEA core or by other plugins. The "defaultExtensionNs " attribute must be set to the ID of the plugin defining the extension point, or to "com.intellij" if the extension point is defined by the IDEA core. The name of the tag within the <extensions> tag matches the name of the extension point, and the "implementation" class specifies the name of the class added to the extension point. --> <extensions xmlns="VssIntegration"> <testExtensionPoint implementation="com.foo.impl.MyExtensionImpl"/> </extensions> </idea-plugin> ```
41.370968
146
0.696491
eng_Latn
0.96917
9131a9f6f961a049b6249b5cf547e4a13f841134
8,094
markdown
Markdown
_posts/python/2020-08-13-virtual-environment.markdown
daesungRa/namu
8a6e5b74a20189fb56d498155e81f55daeb03f52
[ "MIT" ]
null
null
null
_posts/python/2020-08-13-virtual-environment.markdown
daesungRa/namu
8a6e5b74a20189fb56d498155e81f55daeb03f52
[ "MIT" ]
78
2020-10-02T12:50:55.000Z
2022-03-27T08:08:46.000Z
_posts/python/2020-08-13-virtual-environment.markdown
daesungRa/namu
8a6e5b74a20189fb56d498155e81f55daeb03f52
[ "MIT" ]
null
null
null
--- title: Python 가상환경을 만드는 방법 date: 2020-08-13 20:45:48 +0900 author: namu categories: python permalink: "/python/:year/:month/:day/:title" image: https://cdn.pixabay.com/photo/2017/07/31/14/56/wall-2558279_1280.jpg image-view: true image-author: StockSnap image-source: https://pixabay.com/ko/users/stocksnap-894430/ --- --- [목차] 1. [들어가며](#들어가며) 2. [pip](#pip) 3. [virtualenv](#virtualenv) 4. [pipenv](#pipenv) 5. [gitignore 에 포함할 것들](#gitignore-에-포함할-것들) [참조] 1. [python.org, "pip and virtual environments"](https://packaging.python.org/guides/installing-using-pip-and-virtual-environments/) 2. [velog.io](https://velog.io/@doondoony/pipenv-101) --- <br> ### 들어가며 파이썬에서 **_가상환경 개념_**은 매우 중요하다. 여러 개의 프로젝트를 동시에 개발할 때 각각 고유한 파이썬 개발환경을 구축하고(의존성 포함) 상호간에 영향을 미치지 않도록 해야 하기 때문이다. 파이썬으로 개발한다면 언제나 가상환경부터 구축해야 한다는 점을 유념해야 한다. 먼저 os global 영역에 파이썬이 설치되어 있다고 가정하고 출발해보자. python 3.x 버전이라면 무엇이든 가능하다. ### pip pip 는 파이썬 패키지 매니저이다. 현재 사용하는 파이썬 버전에 의존적인 모든 패키지들을 [pypi.org](https://pypi.org/) 아카이브로부터 설치 및 업데이트 해준다. os global 영역에 파이썬이 설치되어 있다면 기본적으로 pip 도 설치되어 있다. - 파이썬과 pip 버전확인 및 pip 업그레이드 ```text # windows > py --version > py -m pip --version > py -m pip install --upgrade pip # linux $ python --version $ python -m pip --version $ python -m pip install --upgrade pip ``` ### virtualenv virtualenv 모듈은 **_프로젝트 내에서 가상환경을 만들어 파이썬 패키지들을 관리_**해준다. 위의 global pip 를 활용해 virtualenv 를 설치한 후, 내 프로젝트를 위한 가상환경을 생성해 보자. - virtualenv 설치 및 업그레이드 ```text # windows > py -m pip install virtualenv > py -m pip install --upgrade virtualenv # 만약 필요하다면 업그레이드! # linux $ python -m pip install virtualenv $ python -m pip install --upgrade virtualenv # 리눅스도 마찬가지. ``` > tip 1 > > 당연하겠지만 환경변수에 global python 인터프리터를 등록해 두자. pip 까지 자유롭게 쓰게. - 가상환경 생성 후 실행 <br>프로젝트 root 디렉토리로 이동한 후, 그곳에 가상환경을 만들자. ```text # windows > cd D:\project\root\ D:\project\root> py -m virtualenv venv D:\project\root> call .\venv\Scripts\activate # 가상환경 실행 (venv) D:\project\root> where pip # 가상환경 내 pip 경로 # linux $ cd /DATA/project/root/ /DATA/project/root$ python -m virtualenv venv /DATA/project/root$ source ./venv/bin/activate # 가상환경 실행 (venv) /DATA/project/root$ which pip # 가상환경 내 pip 경로 ``` - 가상환경 내에서 패키지 설치 > ```requests``` ```text # windows (venv) D:\project\root> pip install requests # ... Successfully installed ... (venv) D:\project\root> pip install requests==2.18.4 # 특정 버전 지정 # linux (venv) /DATA/project/root$ pip install requests # ... Successfully installed ... (venv) /DATA/project/root$ pip install requests==2.18.4 # 여기도. ``` 이로써 가상환경 내 파이썬 인터프리터에 ```requests``` 패키지가 설치되었다. python import 를 통해 확인해 보자. - 패키지 의존성 freezing! > requirements.txt 활용하기 <br>requirements.txt 는 현재 가상환경에 설치된 패키지들의 정적 의존성 정보를 텍스트 파일로 저장한다. 이렇게 해두면 나중에 일일히 설치할 필요가 없어진다. ```text # windows (venv) D:\project\root> pip freeze > requirements.txt # 이 파일이 생성된다. # linux (venv) /DATA/project/root$ pip freeze > requirements.txt ``` 텍스트 파일을 열어보면 다음과 같다. ```text asgiref==3.2.10 certifi==2020.6.20 chardet==3.0.4 Django==3.1 idna==2.10 pytz==2020.1 requests==2.24.0 sqlparse==0.3.1 urllib3==1.25.10 ``` ```requests``` 패키지가 버전 의존성까지 포함해서 기록되어 있다. 이제 이것을 활용해보면, ```text # windows (venv) D:\project\root> pip install -r requirements.txt # 위 정보를 기반으로 패키지 자동설치. # linux (venv) /DATA/project/root$ pip install -r requirements.txt ``` 이제 어디서든 가상환경을 쉽게 구축할 수 있다! - 가상환경 빠져나오기 ```text # windows (venv) D:\project\root> deactivate D:\project\root> # 빠져나왔다! # linux (venv) /DATA/project/root$ deactivate /DATA/project/root$ ``` ### pipenv 이것도 가상환경을 구축하고 관리한다는 점에서 동일하다.<br> 하나 차이점은, ```pip + virtualenv == pipenv``` 라는 것! > Tip 2 > > pipenv 는 더 편리하지만, 업데이트가 느리고 블랙박스 동작이 있다고 한다. 그래도 현업에 사용하기에 큰 무리는 없다. > > 내가 그렇게 사용하고 있으니.. 그러므로 pipenv 를 사용하고자 한다면, 먼저 pip, virtualenv 사용법을 확실히 알자. - 주요 명령어 <br>pip 와 virtualenv 가 합쳐진 만큼, pipenv 하나만으로 가상환경 생성과 가상환경 내 패키지 관리가 가능하다. 이를 위한 기본 명령어들을 살펴보자. - pipenv --python 3.6 : 3.6 버전으로 가상환경 생성(단순 생성!). - pipenv **shell** : 가상환경 쉘 진입. 없다면 생성 후 진입. - pipenv **install** [package-name] : 패키지 설치. ```--dev``` 옵션을 붙이면 개발용으로 설치(Pipfile). 같은 방식으로 버전정보를 추가할 수 있다```~=1.2```. - pipenv **lock** : 현재 버전 의존성 그대로 locking(Pipfile.lock). 이는 **결정론적 빌드**를 보장한다(어떤 상황에서도 같은 환경 보장). - pipenv **sync** : locking 된 버전 의존성 그대로 패키지 설치(최신 버전이 아닐 수 있음). - pipenv **update** : 설치된 패키지들을 최신 버전으로 업데이트. locking 이 자동으로 이루어진다. 업데이트를 원하는 특정 패키지 명시 가능. shell 진입 시 ```Pipfile``` 및 ```Pipfile.lock``` 파일이 생성되며, install, lock, sync, update 명령에 따라 패키지 및 패키지 의존성 정보가 기록된다. Pipfile 예시는 다음과 같다. ```text [[source]] name = "pypi" url = "https://pypi.org/simple" verify_ssl = true [dev-packages] # 개발용은 여기 django = "*" [packages] # 배포용은 여기 requests = "*" django = "*" [requires] # 파이썬 버전 및 환경정보 python_version = "3.6" [scripts] # scripts 섹션을 만들고 start = "python app.py" # pipenv run start ``` - pipenv 설치 및 업그레이드 <br>virtualenv 와 마찬가지로 일단은 global 영역에 설치 및 업그레이드한다. ```text # windows > py -m pip install pipenv > py -m pip install --upgrade pipenv # linux $ python -m pip install pipenv $ python -m pip install --upgrade pipenv ``` - 가상환경 쉘 진입 > 원하는 패키지 설치 > lock ```text # windows > cd D:\project\root\ D:\project\root> py -m pipenv shell # pipenv 쉘로 진입. 없다면 생성. (root-x8NlcEbx) D:\project\root> pipenv install requests (root-x8NlcEbx) D:\project\root> pipenv install django (root-x8NlcEbx) D:\project\root> pipenv install django --dev (root-x8NlcEbx) D:\project\root> pipenv lock # linux $ cd /DATA/project/root/ /DATA/project/root$ python -m pipenv shell (root-x8NlcEbx) /DATA/project/root$ pipenv install requests (root-x8NlcEbx) /DATA/project/root$ pipenv install django (root-x8NlcEbx) /DATA/project/root$ pipenv install django --dev (root-x8NlcEbx) /DATA/project/root$ pipenv lock ``` - 가상환경은 어디에 생성되나?? <br>다음을 실행해 보자. ```text # windows (root-x8NlcEbx) D:\project\root> pipenv --py C:\Users\[USER]\.virtualenvs\root-x8NlcEbx\Scripts\python.exe # linux (root-x8NlcEbx) /DATA/project/root$ pipenv --py $HOME/.local/share/virtualenvs/root-x8NlcEbx/bin/python ``` virtualenv 와 다르게(프로젝트 내 venv), pipenv 의 가상환경은 **현재 계정 home 의 숨겨진 디렉토리 내에 생성**된다. 그러므로 내가 실행하는 프로젝트는 그것이 연계된 home 하위 가상환경의 인터프리터를 활용한다고 생각하면 된다. - 가상환경 삭제 및 빠져나오기 ```text # windows (root-x8NlcEbx) D:\project\root> pipenv --rm Removing virtualenv (C:\Users\[USER]\.virtualenvs\root-x8NlcEbx) … (root-x8NlcEbx) D:\project\root> exit D:\project\root> # linux (root-x8NlcEbx) /DATA/project/root$ pipenv --py Removing virtualenv ($HOME/.local/share/virtualenvs/root-x8NlcEbx) … (root-x8NlcEbx) /DATA/project/root$ exit /DATA/project/root$ ``` ```pipenv --rm``` 이후 위에서 언급한 **현재 계정 home 의 숨겨진 디렉토리**에서 확인해보면 방금전까지 있던 가상환경이 삭제된 것을 볼 수 있다. 사실상 ```rm -rf``` 로 실제 경로를 삭제해도 되기는 하다. ### gitignore 에 포함할 것들 팀 전체가 공유하는 프로젝트 github repository 에 가상환경 전체가 올라가 버린다면 참으로 비효율적일 것이다. 그래서 사용하는 것이 최소한의 정보만 기록된 requirements.txt 나 Pipfile, Pipfile.lock 인 것이다. pipenv 는 전혀 다른 공간에 가상환경을 생성하므로 상관 없지만, virtualenv 를 사용한다면 ```.gitignore``` 에 ```venv``` 혹은 ```가상환경 디렉토리명```을 추가하자. 끝~! <br><br> --- <br> 아 그리고 혹시라도 [**결정론적 빌드**](https://reproducible-builds.org/docs/deterministic-build-systems/)가 무엇인지 궁금하다면 링크를 따라가 문서를 읽어보자. 짧게 설명하자면, 최신 버전의 특정 패키지가 현존하는 프로젝트 의존성을 파괴할지도 모르는 상황 속에서, 언제 어디서나 같은 의존성을 보장하는 재현 가능한 빌드를 상상해보면 된다. 결정론적 빌드 시스템! > **"Be able to get the exact same set of dependencies on multiple machines"**, >[Using locked packages](https://docs.npmjs.com/files/package-locks#using-locked-packages) 진짜끝!!
26.98
131
0.628614
kor_Hang
0.999892
9131ee56d21d543ec50a4735087372d2e2ee7840
1,376
md
Markdown
docs/NugetDocumentation.md
LorenzCK/Pseudo-i18n
08fae4570fb89a832020299a91d1620f703e467c
[ "MIT" ]
6
2016-12-09T01:31:18.000Z
2019-01-23T17:56:51.000Z
docs/NugetDocumentation.md
LorenzCK/Pseudo-i18n
08fae4570fb89a832020299a91d1620f703e467c
[ "MIT" ]
null
null
null
docs/NugetDocumentation.md
LorenzCK/Pseudo-i18n
08fae4570fb89a832020299a91d1620f703e467c
[ "MIT" ]
null
null
null
# Pseudo-i18n *Simple pseudo-internationalization utility library.* The library allows you to convert any latin alphabet string to a pseudo-language in order to test whether your application is localization-ready. The generated pseudo-string will try to respect links, tags, and other markup in your original strings. ## Usage Use the `Translator` class to translate strings at runtime: ```cs PseudoInternationalization.Translator.ConvertToFakeInternationalized("Hello world"); ``` This will return the string `[Ħęľľő ŵőřľđ !!! !!!]`. An extension method can also be used: ```cs using PseudoInternationalization; "Hello world".ToPseudo(); ``` The following rules will be applied: * Strings containing URLs will not be translated, * Strings shorter than 10 characters will grow by 400%, longer strings will grow by 30% (`!` is used as a padding character), * Translated strings will always start with `[` and end with `]`, * Tags, braces, and other markup will not be translated. ## Links A command-line utility that converts resource files is available from the [library’s Github page](https://github.com/LorenzCK/Pseudo-i18n). Based on [John Robbin’s Pseudoizer](http://msdn.microsoft.com/msdnmag/issues/04/04/Bugslayer/default.aspx) and [Scott Hanselman’s implementation](http://www.hanselman.com/blog/PsuedoInternationalizationAndYourASPNETApplication.aspx).
37.189189
249
0.77689
eng_Latn
0.967275
9132dcfe623428ec49b932a5bd286781295c16b8
30
md
Markdown
README.md
khairnaramol/Angular5
7915ef1dee3f908b323295f12ef9a588e3b3dbd5
[ "MIT" ]
null
null
null
README.md
khairnaramol/Angular5
7915ef1dee3f908b323295f12ef9a588e3b3dbd5
[ "MIT" ]
null
null
null
README.md
khairnaramol/Angular5
7915ef1dee3f908b323295f12ef9a588e3b3dbd5
[ "MIT" ]
null
null
null
# Angular5 angular 5 learning
10
18
0.8
eng_Latn
0.781067
9132fda89a07f12c1729e42ee318d1814c7fd396
131
md
Markdown
README.md
LiveTiles/PageGallery
829d187da64775795d09b0f6ea5d85fea117336c
[ "MIT" ]
null
null
null
README.md
LiveTiles/PageGallery
829d187da64775795d09b0f6ea5d85fea117336c
[ "MIT" ]
null
null
null
README.md
LiveTiles/PageGallery
829d187da64775795d09b0f6ea5d85fea117336c
[ "MIT" ]
null
null
null
# PageGallery A collection of demo pages intended to guide folks in using (and in some cases hacking) LiveTiles to fit their needs
43.666667
116
0.801527
eng_Latn
0.99683
913328e76928236576f7f3b53c8297c59213d60f
3,460
md
Markdown
posts/blog/2015/06/15-c88.en.md
danmaq/danmaq.article
ab5626c7a8053175d33044a38404bd8f873f79ef
[ "MIT" ]
null
null
null
posts/blog/2015/06/15-c88.en.md
danmaq/danmaq.article
ab5626c7a8053175d33044a38404bd8f873f79ef
[ "MIT" ]
null
null
null
posts/blog/2015/06/15-c88.en.md
danmaq/danmaq.article
ab5626c7a8053175d33044a38404bd8f873f79ef
[ "MIT" ]
null
null
null
--- title: Comic Market 88 Exhibition Information post_id: '6827' date: '2015-06-15T02:25:22+09:00' draft: true tags: [] --- We _successfully won the_ comic this summer _at Sunday East Q - 24a_ ! ˶\> ◡ <˶ Although it is quickly, I will announce the distribution etc! ## \[\[Newly released game\] MATH.SC (tentative name) \] ( [http://math.sc/](http://math.sc/) ) Pre-Alpha trial version 100 yen → FREE! _For the convenience of completeness this time, I will distribute it for free!_ This new publication is a quiz & card battle game original, smart new sense. It is for Windows (Vista SP 2 or later required). Currently working intensively, we will guide you on [special](http://math.sc/) occasions as soon as there is progress! I will be happy to have the new publication as it is going to bring a certain amount of bamboo shoots, so please come and visit the big players slowly. However, _due to various circumstances on_ that _day, we will withdraw after 13:00_ . ## \[\[Pre-released game\] Concert fantasy mirror - mythical mirror \] ( [http://kagaminer.in/](http://kagaminer.in/) ) 1,000 yen ※ This is the player's replay. [It](http://dangoya.moo.jp/) is a vertical barrage STG using cooperation with Mr. [DANGOYA](http://dangoya.moo.jp/) , the view of the world in the east. Familiar characters such as Cirno and Suwako are like Gensokyo and are not Gensokyo, but have gone into the mysterious world, but such a thing rampage without a doubt! ORIYARA will also appear! It is for Windows (XP SP3 or later required). ## \[\[Ranube Before\] EVIL MAGIC ~ The Witch of Witchcraft \] (/ evilmagic) 800 yen → 500 yen (Comiket special price) ![Special page](https://danmaq.com/wp-content/uploads/2012/11/em_POP.png) It is an authentic fantasy light novel of 180 pages. Girl Enna who can use magical abilities and evil cuckoo of brother and childhood rampage! But the worse man, the lord swept away a boy with a childhood friend! Knowing things, hero's heavenly enel challenges the lord against a game ...! What? [If you are interested please try the trial version which can read only the beginning 69 pages!](https://danmaq.com/filez/em_trial.pdf) (4 MB | PDF Attention) ## \[\[Pre-release music CD\] danmaq soundtrack \] (/! / dst /) ¥ 100 ![Special page](https://danmaq.com/wp-content/uploads/2012/11/dst_jacket.png) 30 songs, a large volume of 75 minutes! It is an _outtake collection_ of songs that danmaq made from 1999 to 2007. Mainly in the eastern direction BEMANI Arrange and many original collections. I recorded only Bocaro and just a little ...! * [This](https://danmaq.com/!/dst/Hey!.mp3) * [Songs and so on](https://danmaq.com/filez/music/new_psm.mp3) * [A lot](https://danmaq.com/filez/music/pcb.mp3) * [It is included!](http://lama.danmaq.com/lamarisa/mp3/15.mp3) **\[\[Previously released music CD\] Leila Prismriver** \] (/! / Leila /) 300 yen → 100 yen (Comiket special price) ![Special page](https://danmaq.com/wp-content/uploads/2012/11/leila1-300x296.jpg) It is an eastward arrangement music CD of 13 songs 64 minutes. We concentrated on secret club music and tried remixing with synthesizer KORG OASYS! * [Listen to all song cross fade demo](https://danmaq.com/!/leila/x.mp3) (7: 11 | 8 MB) The above announcement is _all scheduled_ . Please note that the distribution may be added or changed by the day. ※ Redistribution of summer in Dream Purple Dream PACK is severe due to circumstances of budget> <I am sorry!
104.848485
645
0.738439
eng_Latn
0.986061
913461ab034ea0fe626d2990b170c6da5050a9d6
2,320
md
Markdown
index.md
AliNite/spreadsheets-socialsci
ee98eadef71b9f45775e7115cffc397e0e5dbef5
[ "CC-BY-4.0" ]
null
null
null
index.md
AliNite/spreadsheets-socialsci
ee98eadef71b9f45775e7115cffc397e0e5dbef5
[ "CC-BY-4.0" ]
null
null
null
index.md
AliNite/spreadsheets-socialsci
ee98eadef71b9f45775e7115cffc397e0e5dbef5
[ "CC-BY-4.0" ]
null
null
null
--- layout: lesson root: . --- Good data organization is the foundation of any research project. Most researchers have data in spreadsheets, so it's the place that many research projects start. Typically we organize data in spreadsheets in ways that we as humans want to work with the data. However computers require data to be organized in particular ways. In order to use tools that make computation more efficient, such as programming languages like R or Python, we need to structure our data the way that computers need the data. Since this is where most research projects start, this is where we want to start too! In this lesson, you will learn: - Good data entry practices - formatting data tables in spreadsheets - How to avoid common formatting mistakes - Approaches for handling dates in spreadsheets - Basic quality control and data manipulation in spreadsheets - Exporting data from spreadsheets In this lesson, however, you will *not* learn about data analysis with spreadsheets. Much of your time as a researcher will be spent in the initial 'data wrangling' stage, where you need to organize the data to perform a proper analysis later. It's not the most fun, but it is necessary. In this lesson you will learn how to think about data organization and some practices for more effective data wrangling. With this approach you can better format current data and plan new data collection so less data wrangling is needed. > ## Getting Started > > Data Carpentry's teaching is hands-on, so participants are encouraged to use > their own computers to ensure the proper setup of tools for an efficient > workflow. <br>**These lessons assume no prior knowledge of the skills or tools.** > > To get started, follow the directions in the "[Setup](setup.html)" tab to > download data to your computer and follow any installation instructions. > > #### Prerequisites > > This lesson requires a working copy of spreadsheet software, such as Microsoft > Excel or LibreOffice or OpenOffice.org (see more details in "[Setup](setup/)"). > <br>To most effectively use these materials, please make sure to install > everything *before* working through this lesson. {: .prereq} > ## For Instructors > If you are teaching this lesson in a workshop, please see the > [Instructor notes](guide/). {: .prereq}
42.181818
105
0.775
eng_Latn
0.99958
91346216640df47b297cb0e4df4cc03f68a1e107
3,632
markdown
Markdown
_posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown
martinlowinski/halfthetruth.de
a9513ca95cb07dad58f625b3b56fa56fbf40e946
[ "MIT" ]
null
null
null
_posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown
martinlowinski/halfthetruth.de
a9513ca95cb07dad58f625b3b56fa56fbf40e946
[ "MIT" ]
null
null
null
_posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown
martinlowinski/halfthetruth.de
a9513ca95cb07dad58f625b3b56fa56fbf40e946
[ "MIT" ]
null
null
null
--- wordpress_id: 131 author_login: admin layout: post comments: [] author: martinlowinski title: How a certification authority handles whois data published: true tags: [] date: 2010-03-11 17:09:16 +01:00 categories: - Website author_email: martin@goldtopf.org wordpress_url: http://halfthetruth.de/2010/03/11/how-a-certification-authority-handles-whois-data/ author_url: http://goldtopf.org status: publish --- I just got a call from someone at <a href="http://comodo.com/">comodo.com</a> and the call went something like this<sup><a href="../2010/03/11/how-a-certification-authority-handles-whois-data/#fn12870367704caa288765c55">1</a></sup>: <em>co:</em> I am [some name here] from comodo and looking for the owner of halfthetruth.de. <em>me:</em> That&rsquo;s me. <em>co:</em> Great. First I want to let you know that the ssl-certificate for halfthetruth.de has expired, did you know that? <em>me:</em> Jep. Then a never ending story about what ssl-certs can do for you and your business and how many clients you&rsquo;ll get by buying a &ldquo;so great and cheap certificate from comodo.com&rdquo; &ndash; for only 300 euros (I forgot the exact number) per year. So I told her, that halfthetruth.de is, as you can see, no business and I don&rsquo;t have clients to that not existing business and I just don&rsquo;t need a certificate from comodo. <em>co:</em> Ok, can I ask you how much you paid for your certificate? <em>me:</em> Nothing, it&rsquo;s for free. <em>*pause</em> <em>co:</em> Ok well, and how much do you pay for renewing the certificate? <em>me:</em> It&rsquo;s for free as well, I can renew it anytime for free.[2] <em>*pause</em> <em>co:</em> Than you obviously don&rsquo;t need a certificate from comodo. <em>me:</em> Absolutely right, just one question: Where did you get my phone number? <em>co:</em> We have a research team doing that kind of stuff. <em>me:</em> But I didn&rsquo;t post my phone number on my website or somewhere else publicly available. <em>co:</em> Our research team is doing this, I got your phone number from them.. Then she asked me, if I want to give her my contact information so that she can forward some information regarding comodo and certificates and stuff. I declined politly and hang up. Apparently, they got my phone number from the whois service, but the information from the whois is not meant for advertising as you can see from the terms of use: <blockquote>Terms and Conditions of Use The data in this record is provided by DENIC for informational purposes only. DENIC does not guarantee its accuracy and cannot, under any circumstances, be held liable in case the stored information would prove to be wrong, incomplete or not accurate in any sense. All the domain data that is visible in the whois service is protected by law. It is not permitted to use it for any purpose other than technical or administrative requirements associated with the operation of the Internet. It is explicitly forbidden to extract, copy and/or use or re-utilise in any form and by any means (electronically or not) the whole or a quantitatively or qualitatively substantial part of the contents of the whois database without prior and explicit written permission by DENIC. It is prohibited, in particular, to use it for transmission of unsolicited and/or commercial and/or advertising by phone, fax, e-mail or for any similar purposes.</blockquote> <p id="fn2">Footnotes: <sup>1</sup> I cannot quote the exact words. <sup>2</sup> I&rsquo;m using a certificate from CAcert, which is free of charge and really great by the way.</p>
58.580645
349
0.755231
eng_Latn
0.998625
913475b577bba43e252b21252f8a2597af70fcf0
1,102
md
Markdown
ClassNotes/python_class_homework_0.md
jona-sassenhagen/python_for_psychologists
0604ff5c6382ae02ffeb2e078853b835dab03860
[ "BSD-3-Clause" ]
7
2018-09-19T20:53:55.000Z
2022-02-28T12:55:39.000Z
ClassNotes/python_class_homework_0.md
jona-sassenhagen/python_for_psychologists
0604ff5c6382ae02ffeb2e078853b835dab03860
[ "BSD-3-Clause" ]
null
null
null
ClassNotes/python_class_homework_0.md
jona-sassenhagen/python_for_psychologists
0604ff5c6382ae02ffeb2e078853b835dab03860
[ "BSD-3-Clause" ]
3
2019-03-10T09:25:33.000Z
2021-12-16T20:24:50.000Z
% Python Class Session 1 Homework # Repetitions - Open a new and empty iPython notebook - Create a list of strings that contains the first names of you and your close family members - Access the third entry in that list - Create a list of ages of family members - *Using these two lists (not manually!)*, create a new list containing the names of those family members older than 30 - *With a dict comprehension and the `zip` operator*, construct a dict where keys are family names and values are ages ## Something to actually mildly warm up your brains for - similarly to the list you have created above, and *using the dictionary you have just created*, construct a list of the ages of your family members whose last letter of their first name is an 'a'. # Something new Put some more thought into a simple experiment you want to conduct. It should be as boring and simple as possible, require only very few subjects, and simple stimuli (tip: words are very simple). Note that reaction times or ratings are the easiest to acquire, and in this context almost the only thing that can be made to work
68.875
326
0.779492
eng_Latn
0.999979
9134ac8f71966b0eeeabaf4bd6e1a0c4caf0ed32
211
md
Markdown
README.md
teloxide/teloxide-book
56d226ce5f3efd598365759f8596a3f158ab11a2
[ "BlueOak-1.0.0" ]
4
2021-09-21T09:51:47.000Z
2021-11-28T22:17:58.000Z
README.md
teloxide/teloxide-book
56d226ce5f3efd598365759f8596a3f158ab11a2
[ "BlueOak-1.0.0" ]
null
null
null
README.md
teloxide/teloxide-book
56d226ce5f3efd598365759f8596a3f158ab11a2
[ "BlueOak-1.0.0" ]
null
null
null
# Teloxide user guide This repository contains a user guide for the `teloxide` library. [`teloxide`]: https://github.com/teloxide/teloxide ## Note This book is very much work in progress. Use with caution.
19.181818
65
0.748815
eng_Latn
0.988509
9134c2d6582f2fa40d9dfb8ca520377331f0657a
49
md
Markdown
README.md
yuleihua/aircmn
ff29b25629dcacf65be4fba7fbefc7e7f624f939
[ "Apache-2.0" ]
null
null
null
README.md
yuleihua/aircmn
ff29b25629dcacf65be4fba7fbefc7e7f624f939
[ "Apache-2.0" ]
null
null
null
README.md
yuleihua/aircmn
ff29b25629dcacf65be4fba7fbefc7e7f624f939
[ "Apache-2.0" ]
1
2021-11-13T15:48:26.000Z
2021-11-13T15:48:26.000Z
# aircmn aircmn is common library in c language.
16.333333
39
0.77551
eng_Latn
0.999726
91354c3382e52e23e19ce5f43f3d7c9492305430
2,593
md
Markdown
doc/debugging.md
Chlorie/libunifex
9869196338016939265964b82c7244915de6a12f
[ "Apache-2.0" ]
1
2021-11-23T11:30:39.000Z
2021-11-23T11:30:39.000Z
doc/debugging.md
Chlorie/libunifex
9869196338016939265964b82c7244915de6a12f
[ "Apache-2.0" ]
null
null
null
doc/debugging.md
Chlorie/libunifex
9869196338016939265964b82c7244915de6a12f
[ "Apache-2.0" ]
1
2021-07-29T13:33:13.000Z
2021-07-29T13:33:13.000Z
# Async Stack Traces Unifex contains a prototype implementation of async stack-traces that allows you to traverse a chain/graph of async continuations. A stack-trace consists of a stack of `continuation_info` objects that describes the address of the "frame" and the type of the continuation as well as a mechanism to query what the next continuations in the chain are. This allows you to traverse from a leaf receiver back to the original task that launched it. If you are using structured concurrency and have represented your application as a structured set of tasks then this chain should progress all the way back to the root task of your application. Each receiver must customise the `visit_continuations()` CPO to be able to participate in the async stack-walk. Otherwise, the stack-walk will terminate when it reaches that receiver. Example: ```c++ template<typename Receiver> struct my_receiver { Receiver wrappedReceiver_; void set_value() && noexcept; void set_error(std::exception_ptr) && noexcept; void set_done() && noexcept; template <typename Func> friend void tag_invoke( tag_t<visit_continuations>, const my_receiver& r, Func&& func) { std::invoke(func, wrappedReceiver_); } }; ``` ## Capturing the current stack-trace There is a helper sender called `async_trace_sender` that you can use to get a dump of the async stack-trace at any point in a sender expression. It will produce a `std::vector<async_trace_entry>` that contains a description of the async stack at this point. For example: Some helpers to dump an async trace. ```c++ auto dump_async_trace(std::string tag = {}) { return transform( async_trace_sender{}, [tag = std::move(tag)](const std::vector<async_trace_entry>& entries) { std::cout << "Async Trace (" << tag << "):\n"; for (auto& entry : entries) { std::cout << " " << entry.depth << " [-> " << entry.parentIndex << "]: " << entry.continuation.type().name() << " @ 0x"; std::cout.setf(std::ios::hex, std::ios::basefield); std::cout << entry.continuation.address(); std::cout.unsetf(std::ios::hex); std::cout << "\n"; } }); } template <typename Sender> auto dump_async_trace_on_start(Sender&& sender, std::string tag = {}) { return unifex::sequence(dump_async_trace(std::move(tag)), (Sender &&) sender); } template <typename Sender> auto dump_async_trace_on_completion(Sender&& sender, std::string tag = {}) { return unifex::finally( (Sender &&) sender, dump_async_trace(std::move(tag))); } ```
35.520548
80
0.698419
eng_Latn
0.946728
9135674fda29402a227ebc0d7f0cfe81339cf0b4
82
md
Markdown
README.md
Bearzilasaur/ScholarScraper
ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164
[ "Unlicense" ]
null
null
null
README.md
Bearzilasaur/ScholarScraper
ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164
[ "Unlicense" ]
null
null
null
README.md
Bearzilasaur/ScholarScraper
ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164
[ "Unlicense" ]
1
2019-10-16T13:20:10.000Z
2019-10-16T13:20:10.000Z
# ScholarScraper Repository for a Google Scholar scraper for literature reviews.
27.333333
64
0.829268
eng_Latn
0.813123
9135823d4830ff5567be54c2a660f13205ae659c
295
md
Markdown
playbooks/openshift-monitor-availability/README.md
Roscoe198/Ansible-Openshift
b874bef456852ef082a27dfec4f2d7d466702370
[ "Apache-2.0" ]
164
2015-07-29T17:35:04.000Z
2021-12-16T16:38:04.000Z
playbooks/openshift-monitor-availability/README.md
Roscoe198/Ansible-Openshift
b874bef456852ef082a27dfec4f2d7d466702370
[ "Apache-2.0" ]
3,634
2015-06-09T13:49:15.000Z
2022-03-23T20:55:44.000Z
playbooks/openshift-monitor-availability/README.md
Roscoe198/Ansible-Openshift
b874bef456852ef082a27dfec4f2d7d466702370
[ "Apache-2.0" ]
250
2015-06-08T19:53:11.000Z
2022-03-01T04:51:23.000Z
# OpenShift Availability Monitoring This playbook runs the [OpenShift Availability Monitoring role](../../roles/openshift_monitor_availability). See the role for more information. ## GCP Development The `install-gcp.yml` playbook is useful for ad-hoc installation in an existing GCE cluster.
32.777778
121
0.8
eng_Latn
0.959277
91367433ac2c38ef6788dc22e9c3fbb7bda51853
36
md
Markdown
src/examples/subscript/simple.md
alinex/node-report
0798d2bacf8064875b3f54cd035aa154306f5a7e
[ "Apache-2.0" ]
1
2016-06-02T15:05:20.000Z
2016-06-02T15:05:20.000Z
src/examples/subscript/simple.md
alinex/node-report
0798d2bacf8064875b3f54cd035aa154306f5a7e
[ "Apache-2.0" ]
null
null
null
src/examples/subscript/simple.md
alinex/node-report
0798d2bacf8064875b3f54cd035aa154306f5a7e
[ "Apache-2.0" ]
null
null
null
You need H~2~O for this experiment.
18
35
0.75
eng_Latn
0.999731
9136b6596227434d736d133c7a82fabbf47a2493
964
md
Markdown
_posts/2019-01-19-Github Blog.md
stone8765/blog
90d6420aaef33eae21235075394dd7e332be071b
[ "MIT" ]
null
null
null
_posts/2019-01-19-Github Blog.md
stone8765/blog
90d6420aaef33eae21235075394dd7e332be071b
[ "MIT" ]
null
null
null
_posts/2019-01-19-Github Blog.md
stone8765/blog
90d6420aaef33eae21235075394dd7e332be071b
[ "MIT" ]
null
null
null
--- layout: post title: Github Blog 搭建 author: StoneLi description: 使用frp内网穿透工具可以让内网中的电脑能够像访问公网电脑一样方便,比如将公司或个人电脑里面的Web项目让别人能够访问、或进行电脑远程连接、或ssh连接 catalog: true tags: [jekyll,github pages] --- # 1. 安装Ruby https://rubyinstaller.org/ # 2. 下载安装gem (Ruby的包管理器) 下载:https://rubygems.org/pages/download 解压之后 在目录中执行以下命令 ruby setup.rb # 3. 安装jekyll 在命令行执行gem install jekyll # 4. 运行 jekyll new myblog cd myblog jekyll server 在浏览器输入http://127.0.0.1:4000/即可浏览刚刚创建的blog # 5. Jekyll 主题选择 进入网站 http://jekyllthemes.org/ 选择主题,下载对应的仓库代码到本地即可 # 6.文件中文名本地无法显示的问题 修改安装目录\Ruby22-x64\lib\ruby\2.2.0\webrick\httpservlet下的filehandler.rb文件,建议先备份。 找到下列两处,添加一句(+的一行为添加部分) ``` path = req.path_info.dup.force_encoding(Encoding.find("filesystem")) + path.force_encoding("UTF-8") # 加入编码 if trailing_pathsep?(req.path_info) ``` ``` break if base == "/" + base.force_encoding("UTF-8") # 加入编码 break unless File.directory?(File.expand_path(res.filename + base)) ``` 修改完重新jekyll serve即可支持中文文件名。
20.956522
89
0.763485
yue_Hant
0.648572
913752a80068d6d04d5c0e2c018fa9ab92efb7c0
6,182
md
Markdown
README.md
pecigonzalo/opta
0259f128ad3cfc4a96fe1f578833de28b2f19602
[ "Apache-2.0" ]
null
null
null
README.md
pecigonzalo/opta
0259f128ad3cfc4a96fe1f578833de28b2f19602
[ "Apache-2.0" ]
null
null
null
README.md
pecigonzalo/opta
0259f128ad3cfc4a96fe1f578833de28b2f19602
[ "Apache-2.0" ]
null
null
null
<p align="center"><img src="https://user-images.githubusercontent.com/855699/125824286-149ea52e-ef45-4f41-9579-8dba9bca38ac.png" width="250"><br/> Automated, secure, scalable cloud infrastructure</p> <p align="center"> <a href="https://github.com/run-x/opta/releases/latest"> <img src="https://img.shields.io/github/release/run-x/opta.svg" alt="Current Release" /> </a> <a href="https://github.com/run-x/opta/actions/workflows/ci.yml"> <img src="https://github.com/run-x/opta/actions/workflows/ci.yml/badge.svg" alt="Tests" /> </a> <a href="https://codecov.io/gh/run-x/opta"> <img src="https://codecov.io/gh/run-x/opta/branch/main/graph/badge.svg?token=OA3PXV0HYX"> </a> <a href="http://www.apache.org/licenses/LICENSE-2.0.html"> <img src="https://img.shields.io/badge/LICENSE-Apache2.0-ff69b4.svg" alt="License" /> </a> <img src="https://img.shields.io/github/commit-activity/w/run-x/opta.svg?style=plastic" alt="Commit Activity" /> <img src="https://img.shields.io/badge/downloads-500%2Fweek-blue" /> <a href="https://github.com/PyCQA/bandit"> <img src="https://img.shields.io/badge/security-bandit-yellow.svg" alt="Security" /> </a> </p> <p align="center"> <a href="https://docs.opta.dev/">Documentation</a> | <a href="https://slack.opta.dev"> Slack Community </a> | <a href="mailto:info@runx.dev"> Email: info@runx.dev </a> </p> # What is Opta? Opta is a new kind of Infrastructure-as-Code (IaC) framework that lets engineers work with high-level constructs instead of getting lost in low-level cloud configuration. Opta has a vast library of modules (like EKS, RDS, DynamoDB, GKE, Cloud SQL, and even third-party services like Datadog) that engineers can compose together to build their ideal infrastructure stack. It's built on top of Terraform, and designed so engineers aren’t locked in – anyone can write custom Terraform or even take the Opta-generated Terraform and work independently. Opta gives any engineering team, whether it’s a team of 2 or 200, the same infrastructure advantages that companies like Google or Facebook have, without having to invest in infrastructure or DevOps engineers. ### Why use Opta? Infrastructure-as-Code (IaC) solutions are now widely accepted as the standard for provisioning and managing cloud infrastructure, and Terraform is widely considered to be the best IaC platform on the market – and it is – but Terraform is also quite complex and requires deep Cloud/infrastructure expertise. We developed Opta to help eliminate this complexity. Opta is a simpler IaC framework with best practices built-in. It lets users set up automated, scalable and secure infrastructure on any cloud, without having to be an infrastructure expert, or getting lost in the minutiae of cloud configuration. We are confident it can drastically reduce the complexity and headaches that come with DevOps and infrastructure at most fast moving organizations. Opta is currently being used by dozens of companies of all sizes. To read more about the vision behind Opta, check out this [blog post](https://blog.runx.dev/infrastructure-as-code-for-everyone-7dad6b813cbc). If you'd like to try it out or have any questions - feel free to join our [Slack](https://slack.opta.dev/) or explore the [Getting Started Guide](https://docs.opta.dev/getting-started)! <p align="center"> <a href="https://www.youtube.com/watch?v=nja_EfpGexE"><img width="480" src="https://user-images.githubusercontent.com/855699/149367998-9f00a9f4-abaa-4abf-949c-5b470e7d410c.png"></a> </br> <span><i>Deploying a Ruby on Rails application to AWS</i></span> </p> # Features ### Cloud services Opta supports the 3 major clouds - AWS, GCP and Azure. It has modules for the most commonly used services in these clouds like: * Microservices (powered by [Kubernetes](https://docs.opta.dev/architecture/aws/)) * Databases - Postgres, MySQL, Redis * Serverless workloads * Networking - VPCs, Subnets, Load balancers * CDN (Content Delivery Network) * Object storage (S3, GCS) Additionally, Opta has cloud and security best practices built in, including: * [Observability](https://docs.opta.dev/features/observability/) (Datadog, LogDNA) * [SOC2 compliance](https://docs.opta.dev/compliance/) * [Continuous Deployment](https://docs.opta.dev/features/continuous_deployment/) * Hardened network and security configurations ([AWS](https://docs.opta.dev/architecture/aws/), [GCP](https://docs.opta.dev/architecture/gcp/), [Azure](https://docs.opta.dev/architecture/azure)) * Auto-scaling and high availability (HA) ### Coexistence with existing infrastructure Opta aims to be compatible with your existing infrastructure setup. You can: * Import existing Terraform infrastructure into Opta * Write [custom Terraform modules](https://docs.opta.dev/reference/aws/modules/custom-terraform/) (for services that Opta doesn't support yet) * Run Opta in existing VPCs (WIP) * Export the generated Terraform # Try out Opta Check out the [Getting Started Guide](https://docs.opta.dev/getting-started/). You can also explore some [examples](https://github.com/run-x/opta/tree/main/examples) to get a better idea of what you can do with Opta. # Development We love user contributions! Check out our [Contributing](https://github.com/run-x/opta/blob/main/CONTRIBUTING.md) and [Dev guide](https://github.com/run-x/opta/blob/main/development.md) to get started. # Important Resources * [The Opta Team](https://www.runx.dev/about) * [Check Out The Blog](https://blog.runx.dev/) * [How Opta delivers upgrades](https://github.com/run-x/opta/blob/main/UPGRADING.md) * [Bugfix / Feature request policy](https://github.com/run-x/opta/blob/main/CONTRIBUTING.md#bugfix-resolution-time-expectations) * Comparison with other tools (WIP) * [Our Public roadmap](https://github.com/orgs/run-x/projects/1/views/1) * Case studies - [Flyte](https://blog.flyte.org/how-opta-makes-deploying-flyte-much-easier), [Fastbreak Labs](https://blog.runx.dev/how-fast-break-labs-uses-opta-to-bring-basketball-to-the-blockchain-7556353d70ee), [Canvas app](https://blog.runx.dev/how-the-canvas-team-uses-opta-to-make-data-easier-to-explore-f5615647cc43)
56.2
324
0.752507
eng_Latn
0.77633
9137cde5130df8bd9d5762e1bad5d97d338af01f
5,520
md
Markdown
packages/speeddial/CHANGELOG.md
zamblas/ui-material-components
ec3a4203c0de76d56814e72cfd32e7bd3c077a40
[ "Apache-2.0" ]
null
null
null
packages/speeddial/CHANGELOG.md
zamblas/ui-material-components
ec3a4203c0de76d56814e72cfd32e7bd3c077a40
[ "Apache-2.0" ]
null
null
null
packages/speeddial/CHANGELOG.md
zamblas/ui-material-components
ec3a4203c0de76d56814e72cfd32e7bd3c077a40
[ "Apache-2.0" ]
null
null
null
# Change Log All notable changes to this project will be documented in this file. See [Conventional Commits](https://conventionalcommits.org) for commit guidelines. ## [5.2.8](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.7...v5.2.8) (2021-02-24) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.7](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.6...v5.2.7) (2021-02-22) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.5](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.4...v5.2.5) (2021-02-04) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.4](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.3...v5.2.4) (2021-02-04) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.3](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.2...v5.2.3) (2021-01-23) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.2](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.1...v5.2.2) (2021-01-20) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.2.1](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.0...v5.2.1) (2021-01-19) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial # [5.2.0](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.16...v5.2.0) (2021-01-18) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.16](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.15...v5.1.16) (2020-12-14) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.15](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.14...v5.1.15) (2020-12-13) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.14](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.13...v5.1.14) (2020-12-10) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.13](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.12...v5.1.13) (2020-12-07) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.12](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.11...v5.1.12) (2020-12-04) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.10](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.9...v5.1.10) (2020-12-03) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.9](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.8...v5.1.9) (2020-12-01) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.8](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.7...v5.1.8) (2020-11-27) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.7](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.6...v5.1.7) (2020-11-26) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.6](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.5...v5.1.6) (2020-11-23) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.5](https://github.com/Akylas/nativescript-material-components/compare/v5.1.4...v5.1.5) (2020-11-22) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.4](https://github.com/Akylas/nativescript-material-components/compare/v5.1.3...v5.1.4) (2020-11-13) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.3](https://github.com/Akylas/nativescript-material-components/compare/v5.1.2...v5.1.3) (2020-11-12) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial ## [5.1.1](https://github.com/Akylas/nativescript-material-components/compare/v5.1.0...v5.1.1) (2020-11-08) **Note:** Version bump only for package @nativescript-community/ui-material-speeddial # [5.1.0](https://github.com/Akylas/nativescript-material-components/compare/v5.0.30...v5.1.0) (2020-11-06) ### Features * speeddial basic package ([7e221bb](https://github.com/Akylas/nativescript-material-components/commit/7e221bbb8b84eced2df9e0d81a18d06b1fc5d370))
29.361702
147
0.747464
eng_Latn
0.201678
9137eb9f3047c740fcacf366eaa16c3c438e5a31
204
md
Markdown
README.md
yatace/agn
597a33faf167b31a7fb584f8bebe0842a94b2150
[ "MIT" ]
22
2019-03-01T04:47:56.000Z
2021-06-24T08:31:41.000Z
README.md
yatace/agn
597a33faf167b31a7fb584f8bebe0842a94b2150
[ "MIT" ]
3
2019-03-05T15:34:02.000Z
2020-05-23T03:38:44.000Z
README.md
yatace/agn
597a33faf167b31a7fb584f8bebe0842a94b2150
[ "MIT" ]
5
2019-03-01T07:53:49.000Z
2019-03-05T03:26:32.000Z
# AGN生成器 ## AGN 全称 Make Acfun Great Again Network ~~Acfun Green(GKD) Network~~ 简单的agn评分生成和查询系统 ### 特别鸣谢 [btboyhappy1993](https://github.com/btboyhappy1993) 以上acer为AGN事业作出的贡献 ### [更新记录](changelog.md)
15.692308
69
0.735294
yue_Hant
0.516636
91387a86e3b12c9bdc2df6da857094de4b41fff5
6,546
md
Markdown
src/posts/gsoc-week-3.md
isabelcosta/website
777d1a20c6ef45be87848f829e2935b302d5a65a
[ "MIT" ]
3
2020-06-29T11:36:10.000Z
2020-07-03T10:21:23.000Z
src/posts/gsoc-week-3.md
isabelcosta/isabelcosta.github.io
592ae44426e30c8cedbdbca83af5cc3ec07a71f1
[ "MIT" ]
97
2019-01-30T23:46:40.000Z
2022-02-26T01:59:47.000Z
src/posts/gsoc-week-3.md
isabelcosta/website
777d1a20c6ef45be87848f829e2935b302d5a65a
[ "MIT" ]
7
2019-05-24T11:42:57.000Z
2021-05-14T15:50:26.000Z
--- title: Google Summer of Code | Coding Period | Week 3 date: '2018-06-03' tags: - gsoc crossposts: medium: https://medium.com/isabel-costa-gsoc/google-summer-of-code-coding-period-week-3-349e08f7d998 --- ![](/images/gsoc-week-3-cover.png) This week — May 28 to June 3 — was the third week of the coding period of [Google Summer of Code (GSoC)](https://summerofcode.withgoogle.com/) with [Systers Open Source](https://github.com/systers). If you want to know more about this you can read the [introduction to my journey](https://medium.com/isabel-costa-gsoc/intro-to-google-summer-of-code-with-systers-open-source-dbdaa92bd189) , [my latest weekly blog posts](https://medium.com/isabel-costa-gsoc) or [my weekly status report](https://github.com/systers/mentorship-backend/wiki/GSoC-2018-Isabel-Costa#weekly-status-report-for-week-3) for this week. These first weeks, I’ve been balancing GSoC and my Master’s thesis work. Next week I’ll present my dissertation and then I can fully focus on GSoC work. Even with the thesis I still managed to get some work done. This week I had these 3 pull requests (PR) merged: - Add Swagger documentation for Login API [#20](https://github.com/systers/mentorship-backend/pull/20); - Add initial tests that cover the majority of database models and data access object (DAO) [#16](https://github.com/systers/mentorship-backend/pull/16); - Update pull request template, to have checklist items to update swagger and postman JSON files. These files must be updated every time someone updates the API functionalities [#27](https://github.com/systers/mentorship-backend/pull/27). I also created [some issues on GitHub and organized them](https://github.com/systers/mentorship-backend/issues?utf8=%E2%9C%93&q=is%3Aissue+created%3A2018-05-28..2018-06-03+author%3Aisabelcosta) on the [ZenHub](https://www.zenhub.com/) board. In case you’re unfamiliar with [ZenHub](https://www.zenhub.com/), this is an agile project management tool for [GitHub](https://github.com). Some of these issues were based on bugs found by one of my mentors, [Murad](https://github.com/m-murad), and other issues were discussed at this week’s project meeting. ### Challenges This week I had to squash commits on multiple occasions. For [my first PR](https://github.com/systers/mentorship-backend/pull/14), I had to squash some commits of a branch, which did not include the most recent commit ([here’s a stack overflow answer that helped me](https://stackoverflow.com/questions/24310554/how-do-i-squash-specific-commits-on-a-local-branch/24310701#24310701)). For some of my next PRs I had to squash all commits of a certain branch. I never find my self in any of these scenarios before, so it was quite challenging learning about this. I eventually learned how to squash commits, with some tips from the community and one of my mentors, [Dilushi](https://github.com/Dilu9218). After the second instance of squashing the commits, I was much more comfortable at doing this, for the case where I squashed all the commits of a branch. Because I was still getting acquainted with flask backend development, I wasn’t doing tests and applying [Test-Driven Development (TDD)](https://en.wikipedia.org/wiki/Test-driven_development) methodology yet. I was still trying to figure out what was the best way to structure the app to allow me to test each separate module of the app, from DAOs to data model objects and the API itself. So I dedicated a large portion of this week to learn how to test the code. I implemented some basic tests, just to have an idea of how to test the project from now on. Although tests are a very important part of the project development, for now doing all the possible combination of tests wasn’t a priority, this can be done gradually and implemented for the next features. Also, we now have a more stable structure, than in the first weeks, and I understand a bit more of test development for flask applications. Another point is that by having this project open source, anyone can help to add more tests to the project during the GSoC period as well. I fixed [Swagger](https://swagger.io/) documentation of the backend API so that anyone can test the whole API, using the [Swagger UI](https://swagger.io/tools/swagger-ui/) instead of just using [Postman](https://www.getpostman.com/) (how I usually to test the API). I learned how to use [flask-RESTPlus](http://flask-restplus.readthedocs.io) framework to both document header authorization field and request body. The authorization field is required for some endpoints that require login with an access token. ### Shout-outs to blog posts and open source projects This week, I discovered other great GitHub open source sample projects that helped me in the learning journey. One of these projects — [mjhea0/flaskr-tdd](https://github.com/mjhea0/flaskr-tdd) — shows how to approach TDD using flask applications. Another project that I found was [mjhea0/flask-basic-registration](https://github.com/mjhea0/flask-basic-registration). These two projects were done by [Michael Herman](https://github.com/mjhea0). I also searched about how to implement email verification for new users, which I’m aiming to start implementing next week. I found [this blog post](https://realpython.com/handling-email-confirmation-in-flask/) on this topic, which seems very useful, from [Real Python](https://realpython.com/) website. Here’s the [project from the blog post on GitHub](https://github.com/realpython/flask-registration). Another great article that I found was [“How to structure a Flask-RESTPlus web service for production builds”](https://medium.freecodecamp.org/structuring-a-flask-restplus-web-service-for-production-builds-c2ec676de563) by [Greg Obinna](https://medium.com/@gregobinna). I intend to look more into this in the next days. The GitHub project used for this blog post can be found in [cosmic-byte/flask-restplus-boilerplate](https://github.com/cosmic-byte/flask-restplus-boilerplate). ### Plans for next week - Fix bugs found by my mentor and solve small tasks, until thesis presentation; - Implement email verification; - Start implementing features related to users mentorship pairing (probably after thesis presentation); - Create some issues, approved by my mentors, for new contributors to help out. These will be labeled as “first timers only”, which is the Systers label used to guide newcomers into easy quick issues to get started with open source contributions for [Systers Open Source](https://github.com/systers).
139.276596
1,042
0.779713
eng_Latn
0.992185
9138911b243ecab75293bb6446d9668ffadda978
1,339
md
Markdown
docs/nest/README.md
hackycy/sf-admin-cli
4965d5741589a4589d2ce66824af383299008a17
[ "MIT" ]
3
2021-12-13T07:44:16.000Z
2022-03-11T17:59:02.000Z
docs/nest/README.md
hackycy/sf-admin-cli
4965d5741589a4589d2ce66824af383299008a17
[ "MIT" ]
1
2021-12-13T07:15:58.000Z
2021-12-13T07:42:59.000Z
docs/nest/README.md
hackycy/sf-admin-cli
4965d5741589a4589d2ce66824af383299008a17
[ "MIT" ]
1
2022-03-02T02:38:04.000Z
2022-03-02T02:38:04.000Z
# 介绍 ![](https://img.shields.io/github/commit-activity/m/hackycy/sf-nest-admin) ![](https://img.shields.io/github/license/hackycy/sf-nest-admin) ![](https://img.shields.io/github/repo-size/hackycy/sf-nest-admin) ![](https://img.shields.io/github/languages/top/hackycy/sf-nest-admin) **基于NestJs + TypeScript + TypeORM + Redis + MySql + Vue + Element-UI编写的一款简单高效的前后端分离的权限管理系统。具有:** - 前后端请求参数校验 - JWT 认证 - 基于 NestJs 框架,内置了基础的中间件支持(用户认证、访问日志、请求追踪等) - 用户权限动态刷新 - 代码简单,结构清晰 ::: tip 环境要求 - Node.js 12.x+ - Typescript 2.8+ - MYSQL 5.7+ - Redis 5.0+ ::: ## 演示地址 - [http://opensource.admin.si-yee.com](http://opensource.admin.si-yee.com/) - [Swagger Api文档](http://opensource.admin.si-yee.com/api/doc/admin/swagger-api/static/index.html) 演示环境账号密码: | 账号 | 密码 | 权限 | | :----------: | :----: | :----------------------: | | openadmin | 123456 | 仅只有各个功能的查询权限 | | monitoradmin | 123456 | 系统监控页面及按钮权限 | > 所有新建的用户初始密码都为123456 本地部署账号密码: | 账号 | 密码 | 权限 | | :-------: | :----: | :--------: | | rootadmin | 123456 | 超级管理员 | ## 模块列表 ``` bash ├─系统管理 │ ├─用户管理 │ ├─角色管理 │ ├─菜单管理 ├─系统监控 │ ├─在线用户 │ ├─登录日志 │ ├─请求追踪 ├─任务调度 │ ├─定时任务 │ └─任务日志 ├─网盘空间 │ ├─空间管理 │ └─空间概览 ``` ## 技术选型 ### 后端 - NestJs + TypeScript - TypeORM(MYSQL) - ioredis(Redis) - bull(队列) ### 前端 - Vue、Vue-Router、VueX - Element-UI
18.859155
277
0.585512
yue_Hant
0.701851
9139f94e9a00280f0762282414141a07cf115673
1,295
md
Markdown
_posts/2018/2018-03-26-apache-proxy-balancer.md
haijunsu/navysu.github.io
c5e5d39d4a3dae79a0750e136b6a22e743e60db9
[ "MIT" ]
null
null
null
_posts/2018/2018-03-26-apache-proxy-balancer.md
haijunsu/navysu.github.io
c5e5d39d4a3dae79a0750e136b6a22e743e60db9
[ "MIT" ]
null
null
null
_posts/2018/2018-03-26-apache-proxy-balancer.md
haijunsu/navysu.github.io
c5e5d39d4a3dae79a0750e136b6a22e743e60db9
[ "MIT" ]
null
null
null
--- title: Apache Proxy Balancer author: Haijun (Navy) Su layout: post tags: [proxy, balancer, apache, linux] --- ### Enable proxy models ```shell sudo a2enmod proxy_html sudo a2enmod proxy_http sudo a2enmod proxy_wstunnel sudo a2enmod proxy_ajp sudo a2enmod lbmethod_byrequests sudo a2enmod lbmethod_bytraffic sudo a2enmod lbmethod_bybusyness sudo a2enmod lbmethod_heartbeat ``` ### Modify configuration file ```xml <IfModule mod_ssl.c> <VirtualHost *:443> ServerName happynavy.tk ServerAlias www.happynavy.tk ServerAdmin me@gmail ProxyPreserveHost On Header add Set-Cookie "ROUTEID=.%{BALANCER_WORKER_ROUTE}e; path=/" env=BALANCER_ROUTE_CHANGED <Proxy "balancer://mycluster"> BalancerMember "http://swarm-manager01:3080" route=1 BalancerMember "http://swarm-manager02:3080" route=2 BalancerMember "http://swarm-manager03:3080" route=3 ProxySet stickysession=ROUTEID </Proxy> ProxyPass "/" "balancer://mycluster/" ProxyPassReverse "/" "balancer://mycluster/" SSLCertificateFile /etc/letsencrypt/live/happynavy.tk/fullchain.pem SSLCertificateKeyFile /etc/letsencrypt/live/happynavy.tk/privkey.pem Include /etc/letsencrypt/options-ssl-apache.conf </VirtualHost> </IfModule> ``` source: <https://httpd.apache.org/docs/2.4/mod/mod_proxy_balancer.html>
26.979167
95
0.769884
yue_Hant
0.292929
913a6aeb4f157a79c2fc779a640c8c7533aeb8b7
276
md
Markdown
blog/first-post.md
acrobertson/gatsby-netlify-test
39c544aa52f10091afe9276efc454a033e43b246
[ "MIT" ]
null
null
null
blog/first-post.md
acrobertson/gatsby-netlify-test
39c544aa52f10091afe9276efc454a033e43b246
[ "MIT" ]
4
2021-03-09T18:58:02.000Z
2022-02-26T18:17:35.000Z
blog/first-post.md
acrobertson/gatsby-netlify-test
39c544aa52f10091afe9276efc454a033e43b246
[ "MIT" ]
null
null
null
--- path: /blog/post-1 date: 2019-10-08T18:05:15.349Z title: First Post --- ## First Post Content The following is the content of the first post ### Here's a list - Item 1 - Item 2 - Item 3 [This](https://example.com/ "example") is a link ## Another header More content
13.142857
48
0.673913
eng_Latn
0.96378
913b38cd59d141fcb6d43b788b5ff59a13db35af
597
md
Markdown
includes/data-explorer-authentication.md
p770820/azure-docs.zh-tw
dd2bd917784a4df8b52787a299a3df42e05642fe
[ "CC-BY-4.0", "MIT" ]
null
null
null
includes/data-explorer-authentication.md
p770820/azure-docs.zh-tw
dd2bd917784a4df8b52787a299a3df42e05642fe
[ "CC-BY-4.0", "MIT" ]
null
null
null
includes/data-explorer-authentication.md
p770820/azure-docs.zh-tw
dd2bd917784a4df8b52787a299a3df42e05642fe
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- author: orspod ms.service: data-explorer ms.topic: include ms.date: 10/07/2019 ms.author: orspodek ms.openlocfilehash: a04f17ac809832b6fec51d1ffe0d9fcd6285b4ff ms.sourcegitcommit: f4d8f4e48c49bd3bc15ee7e5a77bee3164a5ae1b ms.translationtype: MT ms.contentlocale: zh-TW ms.lasthandoff: 11/04/2019 ms.locfileid: "73581832" --- ## <a name="authentication"></a>驗證 若要執行下列範例,您需要可存取資源的 Azure Active Directory (Azure AD)應用程式和服務主體。 若要建立免費 Azure AD 應用程式,並在訂用帳戶層級新增角色指派,請參閱[建立 Azure AD 應用程式](/azure/active-directory/develop/howto-create-service-principal-portal)。 您也需要 [目錄(租使用者)識別碼]、[應用程式識別碼] 和 [用戶端密碼]。
35.117647
232
0.798995
yue_Hant
0.265369
913bc31de79d1e9550e767cd21692f2a7d809a88
4,698
md
Markdown
_episodes/01-run-quit.md
lexnederbragt/python-novice-gapminder
6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5
[ "CC-BY-4.0" ]
null
null
null
_episodes/01-run-quit.md
lexnederbragt/python-novice-gapminder
6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5
[ "CC-BY-4.0" ]
null
null
null
_episodes/01-run-quit.md
lexnederbragt/python-novice-gapminder
6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5
[ "CC-BY-4.0" ]
null
null
null
--- title: "Running and Quitting" teaching: 15 exercises: 0 questions: - "How can I run Python programs?" objectives: - "Launch the Jupyter Notebook, create new notebooks, and exit the Notebook." - "Create Markdown cells in a notebook." - "Create and run Python cells in a notebook." keypoints: - FIXME --- ### Python Programs are plain text files. * They have the ".py" extension to let everyone know (including the operating system) it is a Python program. * It's common to write them using a text editor but we are going to use the Jupyter notebook. * The bit of extra setup is well worth it because the notebook provides code completion and other helpful features. * The notebooks have the extension ".ipynb" to distinguish them from plain text python programs. ### The Jupyter Notebook * The [Anaconda package manager][anaconda] is an automated way to install the Jupyter notebook. * It also installs all the extra libraries it needs to run. * Once you have installed Python and the Jupyter Notebook requirements, open a bash shell and type: ~~~ $ jupyter notebook . ~~~ {: .source} * This will start a Jupyter Notebook server and open your default web browser. * The server sends messages to your browser. * The server does the work and the web browser renders the notebook. * You can type code into the browser and see the result when the web page talks to the server. * FIXME: what advantages are there to this arrangement? * The notebook is stored as JSON. * Just like a webpage, the saved notebook looks different to what you see when it gets rendered by your browser. ### Control and Edit modes * Open a new notebook from the dropdown menu in the top right corner of the file browser page. * If you press "esc" and "return" alternately, you will see the surround of your code cell change from blue to green. * The difference in colour is subtle. * These are the control and edit modes of your notebook. * If you use the "esc" and "return" keys to make the surround blue and then press the "H" key, a list of all the shortcut keys will appear. * When in control mode (esc/Blue), * The "B" key will make a new cell below the currently selected cell. * The "A" key will make one above. * The "X" key will delete the current cell. * There are lots of shortcuts you can try out and most actions can done with the menus at the top of the page if you forget the shortcuts. * *If you first remember the "esc" and "H" shortcut, you will be able to find out all the rest.* * Pressing the "return" key turns the surround green to signify edit mode and you can type code into the cell. * Because we want to be able to write many lines of code in a single cell, the "return" key will do what it normally does. * It moves the cursor to the next line just like in a text editor. * We need some other way to tell the Notebook we want to execute the code in the cell. * Pressing the "return" key and the "shift" key together will execute the code in the cell. * Notice that the "return" and "shift" keys on the right of the keyboard are right next to each other. ### Markdown * Notebooks can also render Markdown. * Turn the current cell into a Markdown cell by entering the control mode (esc/blue) and press the "M" key. * The `In [ ]:` will disappear to show it is no longer a code cell and you will be able to write in Markdown. * FIXME: how do you turn a Markdown cell back into a code cell? > ## Creating Lists > > Create a numbered list in a Markdown cell in a notebook > that looks like this: > > 1. Get funding. > 2. Design experiment. > 3. Collect data. > 4. Analyze. > 5. Write up. > 6. Publish. {: .challenge} > ## More Math > > What is displayed when a Python cell in a notebook > that contains several calculations > is executed? > For example, > what happens when this cell is executed? > > ~~~ > 7 * 3 > 2 + 1 > ~~~ > {: .source} {: .challenge} > ## Change an Existing Cell from Code to Markdown > What if you already wrote some code in into your *code cell* and then you switch it to a *markdown cell*? > Give it a try; type the following into a cell in *Edit Mode*: > > ~~~ > x = 6 * 7 + 12 > print(x) > ~~~ > {: .python} > > Now run the cell to be sure that it works as a code cell (shift + return). > > Now go back to the cell, press escape to make sure you're in *Control Mode*, and press the "M" key. Now try running the cell (shift + return). > > What happened and how might this be useful? > > As an added challenge, try changing back to a code cell by pressing the "Y" key from the *Control Mode*. {: .challenge} [anaconda]: https://docs.continuum.io/anaconda/install
40.153846
145
0.714347
eng_Latn
0.999366
913be5f0be284f34ce20eb8ba7118593436c4637
741
md
Markdown
CHANGELOG.md
fizzed/java-jne
783226a1fb002d304d22f841870c5c73575fc994
[ "Apache-2.0" ]
null
null
null
CHANGELOG.md
fizzed/java-jne
783226a1fb002d304d22f841870c5c73575fc994
[ "Apache-2.0" ]
null
null
null
CHANGELOG.md
fizzed/java-jne
783226a1fb002d304d22f841870c5c73575fc994
[ "Apache-2.0" ]
null
null
null
Java Native Extractor by Fizzed =============================== #### 3.0.1 - 2017-08-18 - Only create temp dir a single time per JVM instance - Use UUID for temp dir #### 3.0.0 - 2017-07-17 - Bump parent to v2.1.0 - Add ANY enum for OS - New `findFile` feature to extract generic resources - Initial unit tests - Renamed Arch class to HardwareArchitecture - Renamed OS class to OperatingSystem #### 2.0.1 - 2016-08-23 - Verify temp dir exists (mpanthony) #### 2.0.0 - 2015-12-18 - Refactored package to com.fizzed - Added slf4j for logging #### 1.2.0 - 2015-12-11 - Refactored project layout #### 1.1.0 - 2014-04-16 - Added ability to rename extracted file (e.g. cat to mycat) #### 1.0.1 - 2014-03-18 - Initial release
23.15625
61
0.647773
eng_Latn
0.789744
913c1bb5a49af1546ba459f16a0be9c57c8797af
2,767
md
Markdown
apidoc/README.md
gitizenme/titanium_mobile
f9ebb757a7b78cc18b331cacc266cc5b0a02835f
[ "Apache-2.0" ]
2
2015-05-30T20:28:13.000Z
2021-01-08T17:02:41.000Z
apidoc/README.md
arnaudsj/titanium_mobile
4ed83dd6b355947a88f52efbf4ac82d86a2eeffd
[ "Apache-2.0" ]
6
2015-04-27T22:12:58.000Z
2020-05-23T01:14:06.000Z
apidoc/README.md
arnaudsj/titanium_mobile
4ed83dd6b355947a88f52efbf4ac82d86a2eeffd
[ "Apache-2.0" ]
1
2019-03-15T04:55:17.000Z
2019-03-15T04:55:17.000Z
# TDoc: The Titanium API Documentation Format _This documentation is a WIP_ The TDoc format follows a simple syntax for declaring Modules, Proxies, Methods, Properties, and Events for Titanium. ## Layout The documentation tree starts in the Titanium folder, and generally follows this pattern: <pre> Titanium/ -- Module/ ---- Module.tdoc ---- Proxy.tdoc ---- SubModule/ ------ SubModule.tdoc </pre> ## Format A TDoc file basically consists of sections, and properties, and allows for free-form [Markdown](http://daringfireball.net/projects/markdown/syntax) in most places. A section usually looks like: <pre>- SECTION section content</pre> ### Common sections * namespace (full namespace of the object) * type (module, proxy, view, property, method) * description (a full length description) * since (what version of Titanium this object was added in) * platforms (android,iphone,ipad) * notes (special notes for this object) * example (named code examples) ### Proxy and Module sections * methods (list of method names and descriptions) * method (section for a single method with parameters and return type) * properties (list of property names, types, and descriptions) * events (events fired by this proxy or module) * event (section for a single event with event object properties) ### Named sections The method, event, and example sections all support embedded names, and in the case of the method section also supports a return type. Some quick examples: <pre># returns void - method : helloWorld # returns string - method : methodName, string method documentation.. - event : eventName event documentation.. - example : My Cool Example &lt;code&gt; // here's some code.. &lt;/code&gt;</pre> ### Section Properties Section Properties generally take the form: <pre>name[type]: value</pre> ### Method parameters Method parameters are defined as properties under a Method section. Titanium types can be auto-link by surrounding the full API Name (using Titanium instead of Ti) using backticks. For example: <pre>- method : format, string formatString[string]: The format string args[array,...]: The format arguments - method : createTableView, `Titanium.UI.TableView` returns a new table view - method : updateRow row[`Titanium.UI.TableViewRow`]: row to update</pre> ### Properties Defined as properties under the Properties section, and can also auto-link Titanium types with backticks, For example: <pre>- properties name[string]: Name documentation myView[`Titanium.UI.View`]: my view</pre> ## AutoLinking Titanium types Except in the special cases of Properties and Method Parameters, you can generate a link to a Titanium type by surrounding it with double brackets, like so: <pre>- description Creates a [[Titanium.UI.TableView]]</pre>
31.089888
193
0.759668
eng_Latn
0.992044
913cc5179f86b964793fc14e1130d4ce66e802cb
2,093
md
Markdown
docs/internals/parameter-metadata.md
baileyherbert/reflection
bd161fa6ee32e296729f670b3fa915ea3b4361eb
[ "MIT" ]
1
2021-12-13T18:06:31.000Z
2021-12-13T18:06:31.000Z
docs/internals/parameter-metadata.md
baileyherbert/reflection
bd161fa6ee32e296729f670b3fa915ea3b4361eb
[ "MIT" ]
null
null
null
docs/internals/parameter-metadata.md
baileyherbert/reflection
bd161fa6ee32e296729f670b3fa915ea3b4361eb
[ "MIT" ]
null
null
null
# Parameter Metadata ## Introduction Parameter metadata can be stored in countless ways. This reflection library uses a specific format which enables the `getMetadata()` method to work on parameters. If you are writing your own decorators, consider invoking the `Meta.Parameter` function like below to easily set metadata in the correct structure. ```ts import { Meta } from '@baileyherbert/reflection'; export function Decorator(target: Object, propertyKey: string | symbol, parameterIndex: number) { Meta.Parameter('key', 'value')(target, propertyKey, parameterIndex); } ``` If your decorator only needs to set metadata, you can also directly wrap the `Meta.Parameter` function like below. ```ts import { Meta } from '@baileyherbert/reflection'; // With static data export const Decorator = Meta.Parameter('key', 'value'); // With variable data export const Decorator = (value: string) => Meta.Parameter('key', value); ``` The rest of this page will discuss how metadata is structured internally for this library in case of advanced usage. ## Storage structure - Parameter metadata is stored on the parent method under the `#!ts "reflection:params"` key - The value of this key must be `undefined` or an instance of `#!ts Array<Map<any, any>>` - The array indices are the indices of the parameters - The array values are `Map<any, any>` objects containing the metadata Refer to the following code sample which sets metadata on a parameter just like the `@Meta.Parameter()` decorator: ## Example ```ts function ParameterMeta(metadataKey: any, metadataValue: any) { return function(target: Object, propertyKey: string | symbol, parameterIndex: number) { let parameters = Reflect.getOwnMetadata('reflection:params', target, propertyKey); if (parameters === undefined) { parameters = new Array<Map<any, any>>(); Reflect.defineMetadata('reflection:params', parameters, target, propertyKey); } if (!(parameterIndex in parameters)) { parameters[parameterIndex] = new Map(); } parameters[parameterIndex].set(metadataKey, metadataValue); }; } ```
33.758065
116
0.74343
eng_Latn
0.944263
913cd90eb4548fd2518d1d730eb3973ceab4c6e3
1,290
md
Markdown
src/tech/erlang.md
joelwallis/log
e17f945ce253c3cc62ea215700e8de5c0b4955c8
[ "0BSD" ]
5
2022-01-21T00:43:50.000Z
2022-02-14T21:47:42.000Z
src/tech/erlang.md
joelwallis/knowledge
d9fa6d957fd1641b12eb968300cd4e8ca5a18e1d
[ "0BSD" ]
null
null
null
src/tech/erlang.md
joelwallis/knowledge
d9fa6d957fd1641b12eb968300cd4e8ca5a18e1d
[ "0BSD" ]
null
null
null
# erlang ⓔ My adventures on this amazing distributed computing platform. ## Erlang on macOS through asdf Installing Erlang through asdf is probably the easiest way to get it up and running on macOS. You'll need OpenSSL to run the installation, and the easiest way to get it is through Homebrew: ``` brew install openssl ``` Homebrew's OpenSSL formula is keg-only, and it means that even after installed it won't be globally available in your system. [There's an useful note about it on asdf-erlang's README.md](https://github.com/asdf-vm/asdf-erlang#dealing-with-openssl-issues-on-macos), but I ended up with a command that would use `brew` to get the path of the OpenSSL installation: ```sh # first, export this Kerl-specific variable with the following content export KERL_CONFIGURE_OPTIONS="--without-javac --with-ssl=$(brew --prefix openssl@1.1)" # then, install the Erlang version you want (e.g.: 22.1.5) asdf install erlang 22.1.5 ``` These options will disable Java-related features and point to the correct OpenSSL paths. You'll find more information about setting up Erlang in your machine in the [Setup](https://adoptingerlang.org/docs/development/setup) chapter of the [Adopting Erlang](https://adoptingerlang.org) book (a must-read if you're adopting Erlang just now).
49.615385
361
0.771318
eng_Latn
0.99035
913ce20b781b5af252be1dc41ce8ef7b9729b82b
4,971
md
Markdown
_posts/2019-12-24-hackerrank.md
aSquare14/aSquare14.github.io
740af53840bdc656bc0cbb3c722b069fc8e1c674
[ "MIT" ]
4
2019-08-17T21:05:14.000Z
2021-02-23T20:04:19.000Z
_posts/2019-12-24-hackerrank.md
asquare14/aSquare14.github.io
63a5ba98b80f062af12850da10c4a60701e80f33
[ "MIT" ]
3
2018-03-08T20:23:32.000Z
2021-04-26T13:06:31.000Z
_posts/2019-12-24-hackerrank.md
aSquare14/aSquare14.github.io
740af53840bdc656bc0cbb3c722b069fc8e1c674
[ "MIT" ]
8
2018-06-09T07:29:52.000Z
2020-10-21T22:22:29.000Z
--- title: "My 2019 Summer Internship at Hackerrank Bangalore" layout: post date: 2019-12-24 13:30 tag: - Internship Experiences category: blog author: atibhi description: Weekly Blogs --- I had the opportunity to intern at Hackerrank, Bangalore during the summer of 2019 and in this blog I’d like to tell you about my experience ! For those of you who haven’t heard about Hackerrank, it is the market–leading technical inteview platform to identify and hire developers with the right skills. It also has a community website where you can practice problems, solve challenges and become a better developer. It has offices in Mountain View, London and Bengaluru :) **Applying to Hackerrank** I heard about the internship opening from a friend and applied through their careers page. They hire two kinds of interns, Technical Content Engineers(TCE) and Software Engineering(SWE) Interns. The TCEs are responsible for curating and making problems on algorithms, data structures etc for hackerrank while the SWEs work on DevOps, Back-end, Front-end etc. I applied for an SWE position and had an interview scheduled with the CTO of hackerrank Harishankaran Karunanidhi. After the interview, he said they would get back to me and within a week I had an offer from Hackerrank. I was really excited to join ! ![Image-1-1](https://asquare14.github.io/assets/images/hr-1.jpeg) **The internship and the work culture** The internship started sometime in May and we were given an introductory session where all the interns set up their laptops and finished some formalities. The next day we were assigned tasks and mentors. We were also told and go to interact with the Sales team so we would get to know about how the sales works. This was really fun because as developers we often do not understand how customers buy our software or how the whole marketing and sales department works ! The best thing about hackerrank is that there is no hierarchy, we never felt like interns, we were included in every event, discussion and could give our views and opinions freely. Moreover, we had a weekly meeting with Hari where we explained whatever work we did during that week to him AND also to all the other interns. This was really good as we could learn by listening to other interns as well. During the course of the internship, all the interns(10 of us) were put up in a hotel nearby the office. This helped us to foster great friendships and we helped each other grow technically by having many intellectual discussions . You can check out all the interns here. **What I learnt during my internship** The main tech stack of hackerrank is ruby on rails and react. I had already worked on ruby on rails in my previous internship, so I got the opportunity to work on React and Redux. My mentor was Aakansha Doshi, who is a front-end developer. I learnt so much from her, from the basics to best practices, finding test-cases etc. She also gave me feedback at the end of the internship which I am grateful for. I also worked with Shiv Deepak from the Mountain View office. I worked on an interesting DevOps related task with him. This was also a good experience as I had never worked on DevOps before nor had any idea about it. Apart from the technical learning, I also learnt about stand-ups,team work, how to communicate effectively, test driven development, to not be scared to ask for help but also ask for help only when we have tried enough. **Fun times…** Apart from work, we also had a lot of fun in the office. The office has a foosball table, table tennis room as well as XBox. We would play foosball whenever we wanted to take a break, infact on weekends we would come to office just to play ! The office also has a very well stocked pantry :P ![Image-2](https://asquare14.github.io/assets/images/hr-2.jpeg) ![Image-3](https://asquare14.github.io/assets/images/hr-3.jpeg) Moreover, Hackerrank has tied up with KidsWhoKode and the office of HackerRank in Koramangala, Bengaluru is transformed into a classroom every friday. Underprivileged students visit to learn to code and are taught by the employees of this tech company. The interns volunteered and took turns to teach the children. Also,the full-time employees go out of their way to teach the children ! **Amazing summer came to an end…** All good things must come to an end and so did this internship. I learnt a lot and made great friendships. I would like to thank everyone at Hackerrank for making this one of the best summers of my life ! I would like to thank Ishani, Nalini and Kenny from the HR team for taking great care of us , and Hari, Aakansha, Shiv and so many others for the amazing technical experience ! 😄 P.S — Highly recommend students to apply for an internship. Feel free to reach out to me if you have any questions :) ------ *[This blog post was originally published on [Medium](https://medium.com/@atibhiagrawal/my-2019-summer-internship-at-hackerrank-bengaluru-8f3163612c3a)]*
97.470588
842
0.788171
eng_Latn
0.999894
913dcfd946ce37ed415fb0684c61d92e84cb4a37
4,433
md
Markdown
README.md
mkxml/glc-tratamento-simbolos-inuteis
4a8b089154b8a35cc2245f6271137e025de7c936
[ "MIT" ]
null
null
null
README.md
mkxml/glc-tratamento-simbolos-inuteis
4a8b089154b8a35cc2245f6271137e025de7c936
[ "MIT" ]
null
null
null
README.md
mkxml/glc-tratamento-simbolos-inuteis
4a8b089154b8a35cc2245f6271137e025de7c936
[ "MIT" ]
1
2020-04-22T14:35:00.000Z
2020-04-22T14:35:00.000Z
Remoção dos símbolos inúteis em uma GLC ======================================= Este é um pequeno projeto escrito em JavaScript em cima de Node.JS que implementa um algoritmo de remoção de símbolos inúteis em uma gramática livre do contexto (GLC). O programa foi desenvolvido para a disciplina de **Linguagens formais e Autômatos** da [Universidade Feevale](http://feevale.br). ## Como utilizar **Para utilizar o programa é simples, basta seguir os passos:** - Faça um `git clone` do projeto; - Se você ainda não tem Node.JS instalado no seu computador [instale-o](http://nodejs.com/download); - Junto com o Node.JS foi instalado um software chamado NPM, precisamos dele; - Abra seu Terminal (prompt de comando) e vá até o diretório do clone feito; - Execute `npm install` e as dependências para o programa serão instaladas; - Agora basta rodar o programa com `node app.js -i input.txt`; **O programa suporta alguns parâmetros úteis para customizar seu funcionamento:** - `-i` para indicar qual o arquivo TXT de input você quer utilizar; - `-o` para indicar o arquivo a ser criado para a saída do programa em TXT por padrão o software vai criar um arquivo chamado `output.txt` no diretório local; - `-f` para indicar o formato de quebra de linha, são suportados o `LF` do Unix e o `CRLF` do Windows; - `-h` exibe o help; ## Regras Para ser compatível com o programa o layout deve respeitar algumas regras, são elas: - A primeira linha do input deve declarar as **variáveis** utilizadas na GLC; - A segunda linha do input deve declarar os **terminais** utilizados na GLC; - A terceira linha do input deve indicar a **varíavel inicial** da GLC; - Da quarta linha em diante devem estar dispostas as **regras de produção** da GLC; O programa, funcionando corretamente, deve salvar um outro TXT seguindo as mesmas regras mostradas anteriormente. ### Exemplo de input ``` expr fator termo id nil x y + * ( ) PI expr expr -> termo expr -> expr + termo termo -> fator termo -> termo * fator fator -> id fator -> PI fator -> ( expr ) id -> x id -> y nil -> ``` ### Output ``` expr fator termo id x y + * ( ) PI expr expr -> termo expr -> expr + termo termo -> fator termo -> termo * fator fator -> id fator -> PI fator -> ( expr ) id -> x id -> y ``` ## Tecnologias e estruturas usadas O software é baseado em [Node.JS](http://nodejs.com) e usa a [NPM](http://npmjs.com) para controle de dependências e versionamento. É 100% escrito em JavaScript dentro das regras do projeto de não utilizar funções de parsing da linguagem. Para conseguir fazer o parsing do input fornecido utilizamos arrajnos (*arrays*) em cascata com loops validando cada símbolo, separando-os pelo espaço em branco e quebra de linha. Além disso utilizamos [Git](http://git-scm.com) para controle de versões do código e hospedamos o projeto aqui no GitHub. ## Como funciona O programa funciona em quatro etapas. 1) Primeiramente ele faz o parsing do TXT, separando cada linha em um vetor e depois procura por espaços em branco separando palavras e coloca cada palavra em um célula do vetor da linha. O parser também separa os significados dos símbolos utilizando as regras definidas acima. Criando variáveis para os grupos: **terminais**, **variáveis**, **símbolo inicial** e **regras de produção**. 2) Depois o programa executa a primeira parte do algoritmo que garante que todas as varáveis geram, direta ou indiretamente, terminais. Para fazer isso ele segue o seguinte procedimento: - 1) Considera `v1` todas as variáveis que geram terminais **diretamente**; - 2) Junta ao `v1` todas as variáveis que geram alguma variável do conjunto `v1`; - 3) Repete 2 até que a cardinalidade de `v1` não mude; 3) Aí o programa executa um segundo algoritmo, para garantir que todas as variáveis `v1` são atingidas apartir do **símbolo inicial**. Para fazer isso ele faz o seguinte: - 1) Considera `v2` todas as variáveis que são atingidas pelo **símbolo incial**; - 2) Junta ao `t2` todas os terminais atingidos pelas variáveis do `v2`; - 3) Junta ao `v2` todas as variáveis atingidas por `v2`; - 4) Repete 2 e 3 até que as cardinalidades de `t2` e `v2` não mudem; 4) O último passo é a impressão do TXT de volta para o usuário, seguindo o mesmo layout de entrada imposto. ## Autores - [Matheus Kautzmann](http://github.com/mkautzmann) - [Rômulo Alves](http://github.com/romuloalves) ## Licença [MIT LICENSE](LICENSE)
33.330827
92
0.730205
por_Latn
0.999678
913e56ffe2d482687bad8d76adbc4c6166854441
1,292
md
Markdown
ralbot/uvmgen/README.md
jiacaiyuan/uvm-generator
63f4c7bd0dad43b357d1cc859b61011718c597f8
[ "MIT" ]
13
2020-04-15T09:11:53.000Z
2022-03-13T02:05:53.000Z
ralbot/uvmgen/README.md
jiacaiyuan/uvm-generator
63f4c7bd0dad43b357d1cc859b61011718c597f8
[ "MIT" ]
null
null
null
ralbot/uvmgen/README.md
jiacaiyuan/uvm-generator
63f4c7bd0dad43b357d1cc859b61011718c597f8
[ "MIT" ]
4
2020-11-27T08:11:24.000Z
2022-02-19T09:11:36.000Z
# RALBot-uvm Generate UVM register model from compiled SystemRDL input ## Installing(left blank) Install from [PyPi](https://pypi.org/project/ralbot-uvm) using pip: python3 -m pip install ralbot-uvm -------------------------------------------------------------------------------- ## Exporter Usage Pass the elaborated output of the [SystemRDL Compiler](http://systemrdl-compiler.readthedocs.io) to the exporter. ```python import sys from systemrdl import RDLCompiler, RDLCompileError from ralbot.uvmgen import uvmGenExporter rdlc = RDLCompiler() try: rdlc.compile_file("path/to/my.rdl") root = rdlc.elaborate() except RDLCompileError: sys.exit(1) file = "test.svh" exporter = uvmGenExporter() exporter.export(root, file) ``` -------------------------------------------------------------------------------- ## Reference ### `uvmGenExporter(**kwargs)` Constructor for the uvmGen exporter class **Optional Parameters** * `indentLvl` * Three spaces or four spaces. Default three space. ### `uvmGenExporter.export(node, path)` Perform the export! **Parameters** * `node` * Top-level node to export. Can be the top-level `RootNode` or any internal `AddrmapNode`. * `path` * Output file. Can be (dir+filename without suffix. such as "output/test_uvmgen")
24.377358
96
0.640093
eng_Latn
0.334483
913f5c906602c5786f8954ae075278a2e8b5e72a
16
md
Markdown
README.md
inventioncorps/icb-website-2
f42d8dbe80cb4bb0863b594121ec125e9aca8099
[ "MIT" ]
1
2021-08-29T17:24:08.000Z
2021-08-29T17:24:08.000Z
README.md
inventioncorps/icb-website-2
f42d8dbe80cb4bb0863b594121ec125e9aca8099
[ "MIT" ]
2
2020-12-07T07:21:27.000Z
2020-12-08T08:21:11.000Z
README.md
inventioncorps/icb-website-2
f42d8dbe80cb4bb0863b594121ec125e9aca8099
[ "MIT" ]
null
null
null
# icb-website-2
8
15
0.6875
kor_Hang
0.191865
9140c4a7bccab176e0cd360ee21ee8c5bc08551e
1,040
md
Markdown
docs/goto.md
codemedic/bash-ninja
133c2d8e23c09770618ed0318b37380c35e8b9f8
[ "MIT" ]
9
2018-02-15T03:06:48.000Z
2020-09-21T11:35:13.000Z
docs/goto.md
codemedic/bash-ninja
133c2d8e23c09770618ed0318b37380c35e8b9f8
[ "MIT" ]
null
null
null
docs/goto.md
codemedic/bash-ninja
133c2d8e23c09770618ed0318b37380c35e8b9f8
[ "MIT" ]
1
2019-05-18T07:23:02.000Z
2019-05-18T07:23:02.000Z
## `goto` - bookmarks for the shell `goto` is a path bookmark utility that would help you navigate within the filesystem using bookmarks. It comes with auto-completion and bookmark addition command `goto_add`. Once you `cd` yourself into a path, you can run `goto_add bookmarkName` to add the current working directory into the bookmark. Once added, you can use `goto bookmarkName` to `cd` into the location, from elsewhere. The bookmark-name as well as any sub-directories under the location can be auto-completed by the usual bash means. The bookmark-name and its subpath(s) has to be separated by `#`. You can also go a level above (and auto-completed) using `goto bookmarkName#../`. See [`go_bookmarks.conf`](go_bookmarks.conf) for some examples. The config is re-read and executed each time you invoke `goto` or the auto-completion, so that the changes are instantaneous. In order to install this into your profile, edit your `.bashrc` and add the below lines. go_projects_conf=$HOME/go_bookmarks.conf source $HOME/go.sh
69.333333
261
0.770192
eng_Latn
0.999494
9140fb44218cc8fd1923ae736ee3241d3d884fca
501
md
Markdown
README.md
albovy/PadelGest
32d464cbb3aefa85450b41d67d0212c8c10efc9d
[ "MIT" ]
null
null
null
README.md
albovy/PadelGest
32d464cbb3aefa85450b41d67d0212c8c10efc9d
[ "MIT" ]
null
null
null
README.md
albovy/PadelGest
32d464cbb3aefa85450b41d67d0212c8c10efc9d
[ "MIT" ]
null
null
null
# PADELGEST ## Instalación ### NodeJS Descarga de NodeJs a partir del siguiente enlace: <https://nodejs.org/en/download/> ### ExpressJS Instalación de ExpressJS como dependencia utilizando el gestor de paquetes ` cd PadelGest ` `npm install express -save` ### MongoDB Descarga de MongoDB a partir del siguiente enlace: <https://docs.mongodb.com/manual/administration/install-community/> function test() { console.log("notice the blank line before this function?"); } ´´´
11.651163
75
0.716567
spa_Latn
0.573129
91417cb3dced2d77381fcf7297dab7498abc5e41
1,355
md
Markdown
README.md
net2cn/Real-ESRGAN_GUI
2190f499345546293c6fca0f3de09f747753781b
[ "MIT" ]
16
2021-12-30T05:31:24.000Z
2022-03-30T13:23:39.000Z
README.md
net2cn/Real-ESRGAN_GUI
2190f499345546293c6fca0f3de09f747753781b
[ "MIT" ]
1
2022-03-30T10:07:07.000Z
2022-03-30T11:51:20.000Z
README.md
net2cn/Real-ESRGAN_GUI
2190f499345546293c6fca0f3de09f747753781b
[ "MIT" ]
1
2022-03-20T04:59:59.000Z
2022-03-20T04:59:59.000Z
# Real-ESRGAN_GUI A C# GUI inference implementation of [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN). PRs are welcomed. --- ## Usage You know how a GUI works. ![UI](./assets/screenshot_2021-09-07_194309.png) ## Result From ![256px image](./assets/avatar_256px.png) to ![1024px image](./assets/avatar_256px_realesrgan-x4plus_anime_6B.png) with the magic of [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN). ## Build Prerequisites - Visual Studio 2019 or higher. ## Known Issue - GPU support is not working. - Directory input is not implemented yet. - Alpha channel will be ignored. - Huge memory consumption when handling large image (~1000x1000, eats up ~18.5G memory easily). ## Acknowledgements This repository contains ONNX models converted from [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo. All copyrights and trademarks of the materials used belong to their respective owners and are not being sold. This repository is created only for learning purpose. I DO NOT take any responsibilities for any possible damages. Image [upscale example](./assets/avatar_256px.png) and [result](./assets/avatar_256px_realesrgan-x4plus_anime_6B.png) attached in assets folder are derivatives of my personal artwork for my own SNS avatar. Please do not use without permission, especially for commercial purposes. --- 2021, net2cn.
41.060606
279
0.77786
eng_Latn
0.958297
91423960a711564dc3ae76ec1bdaf8f4765152ae
125
md
Markdown
README.md
krismuniz/pr-geo
7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea
[ "Unlicense" ]
null
null
null
README.md
krismuniz/pr-geo
7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea
[ "Unlicense" ]
null
null
null
README.md
krismuniz/pr-geo
7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea
[ "Unlicense" ]
null
null
null
# pr-geo Data collection of Puerto Rico's Political geography. Colección de datos de la geografía política de Puerto Rico.
25
59
0.792
spa_Latn
0.629213
914261f77470639a9c6c9be24e11a27e093872b5
6,068
md
Markdown
articles/supply-chain/asset-management/setup-for-objects/counters.md
MicrosoftDocs/Dynamics-365-Operations.fr-fr
9f97b0553ee485dfefc0a57ce805f740f4986a7e
[ "CC-BY-4.0", "MIT" ]
2
2020-05-18T17:14:08.000Z
2021-04-20T21:13:46.000Z
articles/supply-chain/asset-management/setup-for-objects/counters.md
MicrosoftDocs/Dynamics-365-Operations.fr-fr
9f97b0553ee485dfefc0a57ce805f740f4986a7e
[ "CC-BY-4.0", "MIT" ]
6
2017-12-13T18:31:58.000Z
2019-04-30T11:46:19.000Z
articles/supply-chain/asset-management/setup-for-objects/counters.md
MicrosoftDocs/Dynamics-365-Operations.fr-fr
9f97b0553ee485dfefc0a57ce805f740f4986a7e
[ "CC-BY-4.0", "MIT" ]
1
2019-10-12T18:19:20.000Z
2019-10-12T18:19:20.000Z
--- title: Mesures d’actifs description: La rubrique explique comment créer des types de mesures d’actifs dans le module Gestion des actifs. author: johanhoffmann ms.date: 10/15/2019 ms.topic: article ms.prod: '' ms.technology: '' ms.search.form: EntAssetObjectCounterPart, EntAssetObjectCounterLookup, EntAssetCounterType, EntAssetObjectCounterTotals audience: Application User ms.reviewer: kamaybac ms.custom: '' ms.assetid: '' ms.search.region: Global ms.author: johanho ms.search.validFrom: 2019-09-30 ms.dyn365.ops.version: 10.0.5 ms.openlocfilehash: fc515615afaa172e1832508d79e202b166f134a9171a0a35ea4f372f9d19b7e2 ms.sourcegitcommit: 42fe9790ddf0bdad911544deaa82123a396712fb ms.translationtype: HT ms.contentlocale: fr-FR ms.lasthandoff: 08/05/2021 ms.locfileid: "6723573" --- # <a name="counters"></a>Compteurs [!include [banner](../../includes/banner.md)] La rubrique explique comment créer des types de compteurs dans le module Gestion des actifs. Les types de compteurs permettent de faire des enregistrements de compteurs sur les actifs, par exemple en ce qui concerne le nombre d’heures de production ou la quantité produite sur l’actif. Les types d’actifs sont liés aux types de compteurs. Cela signifie qu’un compteur ne peut être utilisé sur un actif que si le compteur est configuré sur le type d’actif utilisé sur l’actif. Avant de pouvoir effectuer des enregistrements de compteurs, commencez par créer les types de compteurs à utiliser dans **Compteurs**. Ensuite, vous pouvez créer des enregistrements de compteurs sur les actifs dans **Compteurs**. Les compteurs peuvent être utilisés sur les plans de maintenance. Une ligne de plan de maintenance peut être de type « compteur », par exemple, en relation avec le nombre d’heures ou la quantité de production produite. Un enregistrement de compteur peut être mis à jour manuellement ou automatiquement en fonction des heures ou de la quantité de production produite. Un compteur peut être paramétré pour utiliser une des trois méthodes de mise à jour (que vous sélectionnez dans le champ **Mettre à jour** dans **Compteurs**) : - Manuel - Vous devez enregistrer manuellement des valeurs de compteurs. - Heures de production - le compteur est automatiquement mis à jour en fonction du nombre d’heures de production. - Quantité de production - le compteur est automatiquement mis à jour en fonction des quantités produites. >[!NOTE] >Si la quantité produite est utilisée, *tous* les articles enregistrés sont inclus dans l’enregistrement de compteurs, bonne quantité, ainsi que quantité d’erreur. Il reste possible d’effectuer des enregistrements de compteurs manuels, le cas échéant. ## <a name="create-counter-types-for-asset-counter-registrations"></a>Créez les contre-types pour les enregistrements de compteurs d’actifs 1. Sélectionnez **Gestion des actifs** > **Paramétrage** > **Types d’actifs** > **Compteurs**. 2. Sélectionnez **Nouveau** pour créer un type de compteur. 3. Insérez un ID dans le champ **Compteur**, et un nom de compteur dans le champ **Nom**. 4. Dans l’organisateur **Général**, sélectionnez un compteur dans le champ **Unité**. 5. Dans le champ **Mettre à jour**, sélectionnez la méthode de mise à jour à utiliser pour le compteur. 6. Sélectionnez « Oui » sur le bouton de basculement **Hériter des contre-valeurs** si les actifs enfants d’une structure d’actif doivent hériter automatiquement des enregistrements de compteurs effectués sur l’actif parent. 7. Dans le champ **Agrégat total**, sélectionnez la méthode de synthèse à utiliser pour un compteur à l’aide de ce type de compteur. La « somme » est la sélection standard utilisée pour ajouter en continu des valeurs enregistrées à la valeur totale. La « moyenne » peut être utilisée si un compteur est configuré pour surveiller un seuil, par exemple en ce qui concerne la température, les vibrations ou l’usure d’un actif. 8. Dans le champ **Écart supérieur à**, insérez le niveau supérieur en pourcentage pour vérifier si les enregistrements du compteur manuel se situent dans une plage prévue. La validation est basée sur une augmentation linéaire du nombre d’enregistrements existants de compteurs. 9. Dans le champ **Écart inférieur à**, insérez le niveau inférieur en pourcentage pour vérifier si les enregistrements du compteur manuel se situent dans une plage prévue. La validation est basée sur une diminution linéaire du nombre d’enregistrements existants de compteurs. 10. Dans le champ **Type**, sélectionnez le type de message (informations, avertissement, erreur) à afficher si des écarts en dehors de la plage définie se produisent lorsque vous enregistrez manuellement les compteurs. 11. Dans l’organisateur **Types d’actifs**, ajoutez des types d’actifs qui doivent pouvoir utiliser le compteur. 12. Dans l’organisateur **Compteurs d’actifs associés**, ajoutez le compteur que vous souhaitez mettre à jour automatiquement lorsque ce compteur est mis à jour. >[!NOTE] >Un compteur associé est automatiquement mis à jour uniquement si le compteur associé a le type d’actif auquel elle est liée dans la configuration du compteur. Par exemple : Vous paramétrez un compteur pour les « heures de production » et ajoutez le type « Moteur de camion ». Lorsque ce compteur est mis à jour, un compteur associé à « Huile » est également mis à jour avec les mêmes valeurs de compteurs. Le paramétrage dans **Compteurs** inclut le paramétrage des « Heures ». En outre, pour le compteur « Pétrole », le type d’actif « Moteur de camion » doit être ajouté dans l’organisateur **Types d’actifs** pour garantir la relation de compteur. Voir les captures d’écran ci-dessous pour obtenir un exemple du paramétrage des compteurs Heures et Pétrole. Lorsque des types d’actifs sont ajoutés à un type de compteur dans **Compteurs**, ce compteur est automatiquement ajouté aux types d’actifs dans l’organisateur **Compteurs** dans [Types d’actifs](../setup-for-objects/object-types.md). ![Figure 1.](media/071-setup-for-objects.png) [!INCLUDE[footer-include](../../../includes/footer-banner.md)]
87.942029
759
0.786915
fra_Latn
0.996779
9143286e29be442a8892d1dabb94fee74437b28b
28,208
md
Markdown
articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md
YutongTie-MSFT/azure-docs.de-de
f7922d4a0ebfb2cbb31d7004d4f726202f39716b
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md
YutongTie-MSFT/azure-docs.de-de
f7922d4a0ebfb2cbb31d7004d4f726202f39716b
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md
YutongTie-MSFT/azure-docs.de-de
f7922d4a0ebfb2cbb31d7004d4f726202f39716b
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Streamen von Azure-Diagnosedaten an Event Hubs description: Erfahren Sie, wie Sie eine End-to-End-Konfiguration von Azure-Diagnose mit Event Hubs durchführen. Außerdem werden Anleitungen für gängige Szenarien bereitgestellt. services: azure-monitor author: rboucher ms.service: azure-monitor ms.devlang: dotnet ms.topic: conceptual ms.date: 07/13/2017 ms.author: robb ms.subservice: diagnostic-extension ms.openlocfilehash: c2d577bd4c89046136a3465ff554e9662dd0ce19 ms.sourcegitcommit: cf88cf2cbe94293b0542714a98833be001471c08 ms.translationtype: HT ms.contentlocale: de-DE ms.lasthandoff: 01/23/2019 ms.locfileid: "54478126" --- # <a name="streaming-azure-diagnostics-data-in-the-hot-path-by-using-event-hubs"></a>Streamen von Azure-Diagnosedaten im heißen Pfad mithilfe von Event Hubs Die Azure-Diagnose bietet flexible Methoden zum Erfassen von Metriken und Protokollen virtueller Computer in Clouddiensten sowie zur Übertragung der Ergebnisse an Azure Storage. Seit März 2016 (SDK 2.9) haben Sie die Möglichkeit, benutzerdefinierte Datenquellen als Sendeziel für Diagnose zu verwenden und heiße Daten mithilfe von [Azure Event Hubs](https://azure.microsoft.com/services/event-hubs/) innerhalb von Sekunden zu übertragen. Zu den unterstützten Datentypen gehören: * Ereignisablaufverfolgung für Windows-Ereignisse (ETW) * Leistungsindikatoren * Windows-Ereignisprotokolle * Anwendungsprotokolle * Infrastrukturprotokolle der Azure-Diagnose In diesem Artikel erfahren Sie, wie Sie eine End-to-End-Konfiguration der Azure-Diagnose mit Event Hubs durchführen. Zusätzlich finden Sie Anleitungen für die folgenden gängigen Szenarien: * Anpassen der Protokolle und Metriken, die an Event Hubs gesendet werden * Ändern der Konfigurationen in den einzelnen Umgebungen * Anzeigen von Event Hubs-Streamdaten * Behandlung von Verbindungsproblemen ## <a name="prerequisites"></a>Voraussetzungen Die Verwendung von Event Hubs für den Empfang von Daten von Azure-Diagnose wird von Cloud Services, VMs, VM-Skalierungsgruppen und Service Fabric unterstützt. Die Unterstützung beginnt mit Azure SDK 2.9 und den entsprechenden Azure-Tools für Visual Studio. * Azure-Diagnoseerweiterung 1.6 (standardmäßiges Ziel im[Azure SDK für .NET 2.9 oder höher](https://azure.microsoft.com/downloads/) ) * [Visual Studio 2013 oder höher](https://www.visualstudio.com/downloads/download-visual-studio-vs.aspx) * Vorhandene Konfigurationen der Azure-Diagnose in einer Anwendung mithilfe einer *.wadcfgx* -Datei und einer der folgenden Methoden: * Visual Studio: [Konfigurieren der Diagnose für Azure Cloud Services und Virtual Machines](/visualstudio/azure/vs-azure-tools-diagnostics-for-cloud-services-and-virtual-machines) * Windows PowerShell: [Aktivieren der Diagnose mithilfe von PowerShell in Azure Cloud Services](../../cloud-services/cloud-services-diagnostics-powershell.md) * Es muss ein Event Hubs-Namespace bereitgestellt werden, wie im Artikel [Erste Schritte mit Event Hubs](../../event-hubs/event-hubs-dotnet-standard-getstarted-send.md) ## <a name="connect-azure-diagnostics-to-event-hubs-sink"></a>Herstellen einer Verbindung zwischen der Azure-Diagnose und der Event Hubs-Senke Azure-Diagnose sendet Protokolle und Metriken standardmäßig stets an ein Azure Storage-Konto. Eine Anwendung kann auch Daten an Event Hubs senden, indem der *WADCFGX*-Datei im **PublicConfig** / **WadCfg**-Element ein neuer **Sinks**-Abschnitt hinzugefügt wird. In Visual Studio wird die *WADCFGX*-Datei im folgenden Pfad gespeichert: **Cloud Service-Projekt** > **Rollen** > **(RoleName)** > **diagnostics.wadcfgx**. ```xml <SinksConfig> <Sink name="HotPath"> <EventHub Url="https://diags-mycompany-ns.servicebus.windows.net/diageventhub" SharedAccessKeyName="SendRule" /> </Sink> </SinksConfig> ``` ```JSON "SinksConfig": { "Sink": [ { "name": "HotPath", "EventHub": { "Url": "https://diags-mycompany-ns.servicebus.windows.net/diageventhub", "SharedAccessKeyName": "SendRule" } } ] } ``` In diesem Beispiel wird die Event Hub-URL auf den vollqualifizierten Namespace des Event Hubs festgelegt: Event Hubs-Namespace + „/“ + Event Hub-Name. Die Event Hub-URL wird im [Azure-Portal](https://go.microsoft.com/fwlink/?LinkID=213885) auf dem Event Hubs-Dashboard angezeigt. Der Name der **Senke** kann auf eine beliebige gültige Zeichenfolge festgelegt werden, solange der Wert in der gesamten Konfigurationsdatei konsistent verwendet wird. > [!NOTE] > In diesem Abschnitt können zusätzliche Senken konfiguriert werden, beispielsweise *applicationInsights* . Für die Azure-Diagnose können mehrere Senken definiert werden, wenn jede Senke ebenfalls im Abschnitt **PrivateConfig** deklariert ist. > > Die Event Hubs-Senke muss ebenfalls im Abschnitt **PrivateConfig** der *.wadcfgx* -Konfigurationsdatei deklariert und definiert werden. ```XML <PrivateConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration"> <StorageAccount name="{account name}" key="{account key}" endpoint="{optional storage endpoint}" /> <EventHub Url="https://diags-mycompany-ns.servicebus.windows.net/diageventhub" SharedAccessKeyName="SendRule" SharedAccessKey="{base64 encoded key}" /> </PrivateConfig> ``` ```JSON { "storageAccountName": "{account name}", "storageAccountKey": "{account key}", "storageAccountEndPoint": "{optional storage endpoint}", "EventHub": { "Url": "https://diags-mycompany-ns.servicebus.windows.net/diageventhub", "SharedAccessKeyName": "SendRule", "SharedAccessKey": "{base64 encoded key}" } } ``` Der `SharedAccessKeyName`-Wert muss einem SAS-Schlüssel (Shared Access Signature) und einer Richtlinie entsprechen, die im **Event Hubs**-Namespace definiert wurde. Navigieren Sie zum Event Hubs-Dashboard im [Azure-Portal](https://portal.azure.com), klicken Sie auf die Registerkarte **Konfigurieren**, und richten Sie eine benannte Richtlinie (z.B. „SendRule“) mit Berechtigungen zum *Senden* ein. Das **StorageAccount** wird ebenfalls in **PrivateConfig** konfiguriert. Wenn die Werte hier funktionieren, muss nichts geändert werden. In diesem Beispiel lassen wir die Werte leer. Dies bedeutet, dass ein Downstream-Asset später die Werte festlegt. Beispielsweise legt die Umgebungskonfigurationsdatei *ServiceConfiguration.Cloud.cscfg* die geeigneten Namen und Schlüssel für die Umgebung fest. > [!WARNING] > Der Event Hubs-SAS-Schlüssel wird im Nur-Text-Format in der *.wadcfgx* -Datei gespeichert. Häufig wird dieser Schlüssel in der Quellcodeverwaltung eingecheckt oder ist als Asset in Ihrem Buildserver verfügbar, deshalb sollten die Daten in geeigneter Weise geschützt werden. Es wird empfohlen, hier einen SAS-Schlüssel mit der Berechtigung *Nur senden* zu verwenden, sodass böswillige Benutzer Schreibvorgänge im Event Hub durchführen können, aber nicht am Event Hub lauschen und den Event Hub nicht verwalten können. > > ## <a name="configure-azure-diagnostics-to-send-logs-and-metrics-to-event-hubs"></a>Konfigurieren von Azure-Diagnose für das Senden von Protokollen und Metriken an Event Hubs Wie bereits erläutert, werden alle standardmäßigen und benutzerdefinierten Diagnosedaten (d.h. Metriken und Protokolle) in den konfigurierten Intervallen automatisch an Azure Storage gesendet. Mit Event Hubs und allen zusätzlichen Senken können Sie einen beliebigen Stamm- oder Blattknoten in der Hierarchie an Event Hub senden. Hierzu gehören ETW-Ereignisse, Leistungsindikatoren, Windows-Ereignisprotokolle und Anwendungsprotokolle. Es ist wichtig abzuwägen, wie viele Datenpunkte tatsächlich nach Event Hubs übertragen werden sollten. In der Regel übertragen Entwickler Daten mit niedriger Latenz im heißen Pfad, die schnell genutzt und interpretiert werden müssen. Beispiele dafür sind Systeme, die Warnungen oder Regeln für die automatische Skalierung überwachen. Entwickler können auch einen alternativen Analyse- oder Suchspeicher konfigurieren – beispielsweise Azure Stream Analytics, ElasticSearch, ein benutzerdefiniertes Überwachungssystem oder ein bevorzugtes Überwachungssystem eines Drittanbieters. Nachfolgend sehen Sie einige Beispielkonfigurationen. ```xml <PerformanceCounters scheduledTransferPeriod="PT1M" sinks="HotPath"> <PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\Bytes Total/Sec" sampleRate="PT3M" /> </PerformanceCounters> ``` ```JSON "PerformanceCounters": { "scheduledTransferPeriod": "PT1M", "sinks": "HotPath", "PerformanceCounterConfiguration": [ { "counterSpecifier": "\\Processor(_Total)\\% Processor Time", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Memory\\Available MBytes", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec", "sampleRate": "PT3M" } ] } ``` Im vorangehenden Beispiel wird die Senke auf den übergeordneten **PerformanceCounters**-Knoten in der Hierarchie angewendet. Dies bedeutet, dass alle untergeordneten **PerformanceCounters** an Event Hubs gesendet werden. ```xml <PerformanceCounters scheduledTransferPeriod="PT1M"> <PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Queued" sampleRate="PT3M" sinks="HotPath" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Rejected" sampleRate="PT3M" sinks="HotPath"/> <PerformanceCounterConfiguration counterSpecifier="\Processor(_Total)\% Processor Time" sampleRate="PT3M" sinks="HotPath"/> </PerformanceCounters> ``` ```JSON "PerformanceCounters": { "scheduledTransferPeriod": "PT1M", "PerformanceCounterConfiguration": [ { "counterSpecifier": "\\Processor(_Total)\\% Processor Time", "sampleRate": "PT3M", "sinks": "HotPath" }, { "counterSpecifier": "\\Memory\\Available MBytes", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec", "sampleRate": "PT3M" }, { "counterSpecifier": "\\ASP.NET\\Requests Rejected", "sampleRate": "PT3M", "sinks": "HotPath" }, { "counterSpecifier": "\\ASP.NET\\Requests Queued", "sampleRate": "PT3M", "sinks": "HotPath" } ] } ``` Im vorherigen Beispiel wird die Senke nur auf drei Indikatoren angewendet: **Anforderungen in Warteschlange**, **Zurückgewiesene Anforderungen** und **Prozessorzeit (%)**. Das folgende Beispiel zeigt, wie ein Entwickler die Menge der gesendeten Daten auf die kritischen Metriken für die Dienstintegrität einschränken kann. ```XML <Logs scheduledTransferPeriod="PT1M" sinks="HotPath" scheduledTransferLogLevelFilter="Error" /> ``` ```JSON "Logs": { "scheduledTransferPeriod": "PT1M", "scheduledTransferLogLevelFilter": "Error", "sinks": "HotPath" } ``` In diesem Beispiel wird die Senke auf Protokolle angewendet und auf die Ablaufverfolgung auf Fehlerebene gefiltert. ## <a name="deploy-and-update-a-cloud-services-application-and-diagnostics-config"></a>Bereitstellen und Aktualisieren einer Cloud Services-Anwendung und der Diagnosekonfiguration Visual Studio bietet die einfachste Methode, um die Anwendung und die Event Hubs-Senkenkonfiguration bereitzustellen. Öffnen Sie zum Anzeigen und Bearbeiten der Datei die *.wadcfgx* -Datei in Visual Studio, und bearbeiten und speichern Sie sie. Der Pfad lautet **Cloud Services-Projekt** > **Rollen** > **(RoleName)** > **diagnostics.wadcfgx**. Zu diesem Zeitpunkt ist **.wadcfgx** im Verpackungsprozess aller Aktionen für die Bereitstellung und Bereitstellungsaktualisierung in Visual Studio, Visual Studio Team System sowie in allen Befehlen oder Skripts enthalten, die auf MSBuild basieren und das Ziel */t:publish* verwenden. Darüber hinaus wird die Datei bei allen Bereitstellungen und Aktualisierungen mit der entsprechenden Agent-Erweiterung der Azure-Diagnose auf Ihren virtuellen Computern in Azure bereitgestellt. Nach der Bereitstellung der Anwendung und der Konfiguration der Azure-Diagnose wird die Aktivität sofort auf dem Dashboard des Event Hubs angezeigt. Anschließend können Sie die heißen Daten im Listenerclient oder in einem Analysetool Ihrer Wahl anzeigen. In der folgenden Abbildung zeigt das Event Hubs-Dashboard, dass Diagnosedaten nach 23:00 Uhr ohne Fehler an den Event Hub gesendet wurden. Zu dieser Zeit wurde die Anwendung mit einer aktualisierten *.wadcfgx* -Datei bereitgestellt, und die Senke wurde ordnungsgemäß konfiguriert. ![][0] > [!NOTE] > Wenn Sie an der Konfigurationsdatei für die Azure-Diagnose (.wadcfgx) Änderungen vornehmen, wird empfohlen, die Aktualisierungen per Push an die gesamte Anwendung sowie die Konfiguration zu übertragen. Verwenden Sie hierzu entweder eine Veröffentlichung über Visual Studio oder ein Windows PowerShell-Skript. > > ## <a name="view-hot-path-data"></a>Anzeigen von Daten im heißen Pfad Wie bereits erläutert, gibt es viele Anwendungsfälle für das Überwachen und Verarbeiten von Event Hubs-Daten. Ein einfacher Ansatz ist das Erstellen einer kleinen Testkonsolenanwendung, die am Event Hub lauscht und den Ausgabestream druckt. Sie können den folgenden Code (dieser wird in [Erste Schritte mit Event Hubs](../../event-hubs/event-hubs-dotnet-standard-getstarted-send.md) ausführlicher erläutert) in einer Konsolenanwendung platzieren. Beachten Sie, dass die Konsolenanwendung das [NuGet-Paket EventProcessorHost](https://www.nuget.org/packages/Microsoft.Azure.ServiceBus.EventProcessorHost/) umfassen muss. Denken Sie daran, die Werte in spitzen Klammern in der **Main** -Funktion durch Werte für Ihre Ressourcen zu ersetzen. ```csharp //Console application code for EventHub test client using System; using System.Collections.Generic; using System.Diagnostics; using System.Linq; using System.Text; using System.Threading.Tasks; using Microsoft.ServiceBus.Messaging; namespace EventHubListener { class SimpleEventProcessor : IEventProcessor { Stopwatch checkpointStopWatch; async Task IEventProcessor.CloseAsync(PartitionContext context, CloseReason reason) { Console.WriteLine("Processor Shutting Down. Partition '{0}', Reason: '{1}'.", context.Lease.PartitionId, reason); if (reason == CloseReason.Shutdown) { await context.CheckpointAsync(); } } Task IEventProcessor.OpenAsync(PartitionContext context) { Console.WriteLine("SimpleEventProcessor initialized. Partition: '{0}', Offset: '{1}'", context.Lease.PartitionId, context.Lease.Offset); this.checkpointStopWatch = new Stopwatch(); this.checkpointStopWatch.Start(); return Task.FromResult<object>(null); } async Task IEventProcessor.ProcessEventsAsync(PartitionContext context, IEnumerable<EventData> messages) { foreach (EventData eventData in messages) { string data = Encoding.UTF8.GetString(eventData.GetBytes()); Console.WriteLine(string.Format("Message received. Partition: '{0}', Data: '{1}'", context.Lease.PartitionId, data)); foreach (var x in eventData.Properties) { Console.WriteLine(string.Format(" {0} = {1}", x.Key, x.Value)); } } //Call checkpoint every 5 minutes, so that worker can resume processing from 5 minutes back if it restarts. if (this.checkpointStopWatch.Elapsed > TimeSpan.FromMinutes(5)) { await context.CheckpointAsync(); this.checkpointStopWatch.Restart(); } } } class Program { static void Main(string[] args) { string eventHubConnectionString = "Endpoint= <your connection string>”; string eventHubName = "<Event hub name>"; string storageAccountName = "<Storage account name>"; string storageAccountKey = "<Storage account key>”; string storageConnectionString = string.Format("DefaultEndpointsProtocol=https;AccountName={0};AccountKey={1}", storageAccountName, storageAccountKey); string eventProcessorHostName = Guid.NewGuid().ToString(); EventProcessorHost eventProcessorHost = new EventProcessorHost(eventProcessorHostName, eventHubName, EventHubConsumerGroup.DefaultGroupName, eventHubConnectionString, storageConnectionString); Console.WriteLine("Registering EventProcessor..."); var options = new EventProcessorOptions(); options.ExceptionReceived += (sender, e) => { Console.WriteLine(e.Exception); }; eventProcessorHost.RegisterEventProcessorAsync<SimpleEventProcessor>(options).Wait(); Console.WriteLine("Receiving. Press enter key to stop worker."); Console.ReadLine(); eventProcessorHost.UnregisterEventProcessorAsync().Wait(); } } } ``` ## <a name="troubleshoot-event-hubs-sinks"></a>Beheben von Problemen mit Event Hubs-Senken * Der Event Hub zeigt keine eingehende oder ausgehende Ereignisaktivität. Überprüfen Sie, ob der Event Hub erfolgreich bereitgestellt wurde. Alle Verbindungsinformationen im Abschnitt **PrivateConfig** der *.wadcfgx* -Datei müssen mit den Werten Ihrer Ressourcen übereinstimmen, wie sie im Portal angezeigt werden. Stellen Sie sicher, dass Sie eine SAS-Richtlinie im Portal definiert haben (im Beispiel „SendRule“) und die Berechtigung *Senden* gewährt wurde. * Nach einer Aktualisierung zeigt der Event Hub keine eingehende oder ausgehende Ereignisaktivität mehr. Vergewissern Sie sich zunächst, dass die Event Hub- und Konfigurationsinformationen richtig sind (wie zuvor beschrieben). Manchmal wird **PrivateConfig** bei einer Bereitstellungsaktualisierung zurückgesetzt. Zur Beseitigung des Problems wird empfohlen, alle Änderungen an der *.wadcfgx* -Datei im Projekt durchzuführen und dann ein vollständiges Anwendungsupdate per Push zu übertragen. Wenn dies nicht möglich ist, stellen Sie sicher, dass mit dem Diagnoseupdate eine vollständige **PrivateConfig** einschließlich des SAS-Schlüssels per Push übertragen wird. * Ich habe die empfohlenen Schritte ausgeführt, aber der Event Hub funktioniert weiterhin nicht. Überprüfen Sie die Azure Storage-Tabelle, die Protokolle und Fehler zur Azure-Diagnose selbst enthält: **WADDiagnosticInfrastructureLogsTable**. Sie können hierzu beispielsweise mit dem Tool [Azure Storage-Explorer](https://www.storageexplorer.com) eine Verbindung mit diesem Speicherkonto herstellen, diese Tabelle anzeigen und eine Abfrage nach einem TimeStamp innerhalb der letzten 24 Stunden hinzufügen. Sie können das Tool verwenden, um eine CSV-Datei zu exportieren und in einer Anwendung wie Microsoft Excel zu öffnen. Excel erleichtert das Suchen nach Callingcard-Zeichenfolgen wie **EventHubs**, um festzustellen, welcher Fehler gemeldet wird. ## <a name="next-steps"></a>Nächste Schritte • [Weitere Informationen zu Event Hubs](https://azure.microsoft.com/services/event-hubs/) ## <a name="appendix-complete-azure-diagnostics-configuration-file-wadcfgx-example"></a>Anhang: Beispiel einer vollständigen Konfigurationsdatei für die Azure-Diagnose (.wadcfgx) ```xml <?xml version="1.0" encoding="utf-8"?> <DiagnosticsConfiguration xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration"> <PublicConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration"> <WadCfg> <DiagnosticMonitorConfiguration overallQuotaInMB="4096" sinks="applicationInsights.errors"> <DiagnosticInfrastructureLogs scheduledTransferLogLevelFilter="Error" /> <Directories scheduledTransferPeriod="PT1M"> <IISLogs containerName="wad-iis-logfiles" /> <FailedRequestLogs containerName="wad-failedrequestlogs" /> </Directories> <PerformanceCounters scheduledTransferPeriod="PT1M" sinks="HotPath"> <PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\Bytes Total/Sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET Applications(__Total__)\Requests/Sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET Applications(__Total__)\Errors Total/Sec" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Queued" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Rejected" sampleRate="PT3M" /> <PerformanceCounterConfiguration counterSpecifier="\Processor(_Total)\% Processor Time" sampleRate="PT3M" /> </PerformanceCounters> <WindowsEventLog scheduledTransferPeriod="PT1M"> <DataSource name="Application!*" /> </WindowsEventLog> <CrashDumps> <CrashDumpConfiguration processName="WaIISHost.exe" /> <CrashDumpConfiguration processName="WaWorkerHost.exe" /> <CrashDumpConfiguration processName="w3wp.exe" /> </CrashDumps> <Logs scheduledTransferPeriod="PT1M" sinks="HotPath" scheduledTransferLogLevelFilter="Error" /> </DiagnosticMonitorConfiguration> <SinksConfig> <Sink name="HotPath"> <EventHub Url="https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py" SharedAccessKeyName="SendRule" /> </Sink> <Sink name="applicationInsights"> <ApplicationInsights /> <Channels> <Channel logLevel="Error" name="errors" /> </Channels> </Sink> </SinksConfig> </WadCfg> <StorageAccount>ACCOUNT_NAME</StorageAccount> </PublicConfig> <PrivateConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration"> <StorageAccount name="{account name}" key="{account key}" endpoint="{storage endpoint}" /> <EventHub Url="https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py" SharedAccessKeyName="SendRule" SharedAccessKey="YOUR_KEY_HERE" /> </PrivateConfig> <IsEnabled>true</IsEnabled> </DiagnosticsConfiguration> ``` Die ergänzende Datei *ServiceConfiguration.Cloud.cscfg* für dieses Beispiel sieht folgendermaßen aus. ```xml <?xml version="1.0" encoding="utf-8"?> <ServiceConfiguration serviceName="MyFixItCloudService" xmlns="http://schemas.microsoft.com/ServiceHosting/2008/10/ServiceConfiguration" osFamily="3" osVersion="*" schemaVersion="2015-04.2.6"> <Role name="MyFixIt.WorkerRole"> <Instances count="1" /> <ConfigurationSettings> <Setting name="Microsoft.WindowsAzure.Plugins.Diagnostics.ConnectionString" value="YOUR_CONNECTION_STRING" /> </ConfigurationSettings> </Role> </ServiceConfiguration> ``` Entsprechende JSON-Einstellungen für virtuelle Computer lauten wie folgt: Öffentliche Einstellungen: ```JSON { "WadCfg": { "DiagnosticMonitorConfiguration": { "overallQuotaInMB": 4096, "sinks": "applicationInsights.errors", "DiagnosticInfrastructureLogs": { "scheduledTransferLogLevelFilter": "Error" }, "Directories": { "scheduledTransferPeriod": "PT1M", "IISLogs": { "containerName": "wad-iis-logfiles" }, "FailedRequestLogs": { "containerName": "wad-failedrequestlogs" } }, "PerformanceCounters": { "scheduledTransferPeriod": "PT1M", "sinks": "HotPath", "PerformanceCounterConfiguration": [ { "counterSpecifier": "\\Memory\\Available MBytes", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Web Service(_Total)\\Bytes Total/Sec", "sampleRate": "PT3M" }, { "counterSpecifier": "\\ASP.NET Applications(__Total__)\\Requests/Sec", "sampleRate": "PT3M" }, { "counterSpecifier": "\\ASP.NET Applications(__Total__)\\Errors Total/Sec", "sampleRate": "PT3M" }, { "counterSpecifier": "\\ASP.NET\\Requests Queued", "sampleRate": "PT3M" }, { "counterSpecifier": "\\ASP.NET\\Requests Rejected", "sampleRate": "PT3M" }, { "counterSpecifier": "\\Processor(_Total)\\% Processor Time", "sampleRate": "PT3M" } ] }, "WindowsEventLog": { "scheduledTransferPeriod": "PT1M", "DataSource": [ { "name": "Application!*" } ] }, "Logs": { "scheduledTransferPeriod": "PT1M", "scheduledTransferLogLevelFilter": "Error", "sinks": "HotPath" } }, "SinksConfig": { "Sink": [ { "name": "HotPath", "EventHub": { "Url": "https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py", "SharedAccessKeyName": "SendRule" } }, { "name": "applicationInsights", "ApplicationInsights": "", "Channels": { "Channel": [ { "logLevel": "Error", "name": "errors" } ] } } ] } }, "StorageAccount": "{account name}" } ``` Geschützte Einstellungen: ```JSON { "storageAccountName": "{account name}", "storageAccountKey": "{account key}", "storageAccountEndPoint": "{storage endpoint}", "EventHub": { "Url": "https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py", "SharedAccessKeyName": "SendRule", "SharedAccessKey": "YOUR_KEY_HERE" } } ``` ## <a name="next-steps"></a>Nächste Schritte Weitere Informationen zu Event Hubs finden Sie unter den folgenden Links: * [Übersicht über Event Hubs](../../event-hubs/event-hubs-about.md) * [Erstellen eines Event Hubs](../../event-hubs/event-hubs-create.md) * [Event Hubs – häufig gestellte Fragen](../../event-hubs/event-hubs-faq.md) <!-- Images. --> [0]: ../../event-hubs/media/event-hubs-streaming-azure-diags-data/dashboard.png
54.455598
797
0.694448
deu_Latn
0.875963
9143d7326c3d92b7dcb88b1ccad7e8fb0ffbcdce
13,797
md
Markdown
treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md
emmettstr/docs
2d0376d6e07f3ffa828f6152d12cf260a530c64d
[ "Apache-2.0" ]
null
null
null
treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md
emmettstr/docs
2d0376d6e07f3ffa828f6152d12cf260a530c64d
[ "Apache-2.0" ]
null
null
null
treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md
emmettstr/docs
2d0376d6e07f3ffa828f6152d12cf260a530c64d
[ "Apache-2.0" ]
null
null
null
--- layout: base title: 'Statistics of fixed in UD_Russian-SynTagRus' udver: '2' --- ## Treebank Statistics: UD_Russian-SynTagRus: Relations: `fixed` This relation is universal. 9959 nodes (1%) are attached to their parents as `fixed`. 9959 instances of `fixed` (100%) are left-to-right (parent precedes child). Average distance between parent and child is 1.21889747966663. The following 83 pairs of parts of speech are connected with `fixed`: <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (3357; 34% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (950; 10% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (622; 6% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (533; 5% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (440; 4% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (369; 4% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (358; 4% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (339; 3% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt> (271; 3% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (253; 3% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (250; 3% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (196; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (190; 2% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (188; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (187; 2% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (175; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (156; 2% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (114; 1% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (108; 1% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (95; 1% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (88; 1% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (86; 1% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (84; 1% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (52; 1% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (45; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (41; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (36; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (36; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (34; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (28; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt> (28; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (18; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (15; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (15; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (13; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (12; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-AUX.html">AUX</a></tt> (11; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (9; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (6; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (6; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-AUX.html">AUX</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (3; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (3; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-INTJ.html">INTJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-X.html">X</a></tt>-<tt><a href="ru_syntagrus-pos-X.html">X</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-SYM.html">SYM</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-X.html">X</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances). ~~~ conllu # visual-style 3 bgColor:blue # visual-style 3 fgColor:white # visual-style 1 bgColor:blue # visual-style 1 fgColor:white # visual-style 1 3 fixed color:blue 1 До до ADP _ _ 4 advmod 4:advmod _ 2 сих сей DET _ Case=Gen|Number=Plur 1 fixed 1:fixed _ 3 пор пора NOUN _ Animacy=Inan|Case=Gen|Gender=Fem|Number=Plur 1 fixed 1:fixed _ 4 идут идти VERB _ Aspect=Imp|Mood=Ind|Number=Plur|Person=3|Tense=Pres|VerbForm=Fin|Voice=Act 0 root 0:root _ 5 споры спор NOUN _ Animacy=Inan|Case=Nom|Gender=Masc|Number=Plur 4 nsubj 4:nsubj _ 6 о о ADP _ _ 7 case 7:case _ 7 том то PRON _ Animacy=Inan|Case=Loc|Gender=Neut|Number=Sing 5 nmod 5:nmod SpaceAfter=No 8 , , PUNCT _ _ 9 punct 9:punct _ 9 достигли достичь VERB _ Aspect=Perf|Mood=Ind|Number=Plur|Tense=Past|VerbForm=Fin|Voice=Act 7 acl 7:acl _ 10 они они PRON _ Case=Nom|Number=Plur|Person=3 9 nsubj 9:nsubj _ 11 вершины вершина NOUN _ Animacy=Inan|Case=Gen|Gender=Fem|Number=Sing 9 obl 9:obl _ 12 или или CCONJ _ _ 13 cc 13.1:cc _ 13 нет нет PART _ _ 9 conj 13.1:discourse SpaceAfter=No ~~~ ~~~ conllu # visual-style 2 bgColor:blue # visual-style 2 fgColor:white # visual-style 1 bgColor:blue # visual-style 1 fgColor:white # visual-style 1 2 fixed color:blue 1 Так так ADV _ Degree=Pos 4 mark 4:mark _ 2 что что SCONJ _ _ 1 fixed 1:fixed _ 3 - - PUNCT _ _ 1 punct 1:punct _ 4 спасибо спасибо NOUN _ Animacy=Inan|Case=Nom|Gender=Neut|Number=Sing 0 root 0:root SpaceAfter=No 5 , , PUNCT _ _ 6 punct 6:punct _ 6 власть власть NOUN _ Animacy=Inan|Case=Nom|Gender=Fem|Number=Sing 4 parataxis 4:parataxis SpaceAfter=No 7 , , PUNCT _ _ 8 punct 8:punct _ 8 спасибо спасибо NOUN _ Animacy=Inan|Case=Nom|Gender=Neut|Number=Sing 4 conj 4:conj SpaceAfter=No 9 , , PUNCT _ _ 10 punct 10:punct _ 10 Путин Путин PROPN _ Animacy=Anim|Case=Nom|Gender=Masc|Number=Sing 8 parataxis 8:parataxis SpaceAfter=No 11 . . PUNCT _ _ 4 punct 4:punct _ ~~~ ~~~ conllu # visual-style 6 bgColor:blue # visual-style 6 fgColor:white # visual-style 5 bgColor:blue # visual-style 5 fgColor:white # visual-style 5 6 fixed color:blue 1 Разумеется разумеется ADV _ Degree=Pos 7 parataxis 7:parataxis SpaceAfter=No 2 , , PUNCT _ _ 1 punct 1:punct _ 3 эти этот DET _ Case=Acc|Number=Plur 4 det 4:det _ 4 объяснения объяснение NOUN _ Animacy=Inan|Case=Acc|Gender=Neut|Number=Plur 8 obj 8:obj _ 5 вряд вряд ADV _ Degree=Pos 7 advmod 7:advmod _ 6 ли ли PART _ _ 5 fixed 5:fixed _ 7 можно можно ADV _ Degree=Pos 0 root 0:root _ 8 счесть счесть VERB _ Aspect=Perf|VerbForm=Inf|Voice=Act 7 csubj 7:csubj _ 9 убедительными убедительный ADJ _ Case=Ins|Degree=Pos|Number=Plur 8 obl 8:obl SpaceAfter=No 10 . . PUNCT _ _ 7 punct 7:punct _ ~~~
166.228916
10,758
0.679278
yue_Hant
0.670034
914405e6e10ea424254abb3a4b037e754c09afb0
6,509
md
Markdown
README.md
by46/WhaleFS
20029ad9a9b59089a3fdd681699ddfb3a5624ebc
[ "MIT" ]
1
2018-06-10T08:54:54.000Z
2018-06-10T08:54:54.000Z
README.md
by46/whalefs
20029ad9a9b59089a3fdd681699ddfb3a5624ebc
[ "MIT" ]
null
null
null
README.md
by46/whalefs
20029ad9a9b59089a3fdd681699ddfb3a5624ebc
[ "MIT" ]
null
null
null
# whalefs ## seaweedfs ```bash weed master -port=9001 weed master -port=9002 -peers="localhost:9001" weed volume -port=9081 -mserver="localhost:9001" -dir="data" /opt/weed/weed master -mdir="/opt/dfs/master" /opt/weed/weed volume -ip=192.168.1.9 -port=18081 -mserver="localhost:9333" -dir="/opt/dfs/data1" /opt/weed/weed volume -ip=192.168.1.9 -port=18082 -mserver="localhost:9333" -dir="/opt/dfs/data2" /opt/weed/weed volume -ip=192.168.1.9 -port=18083 -mserver="localhost:9333" -dir="/opt/dfs/data3" ``` ### buckets system.buckets ```json { "buckets": [ "system.bucket.benjamin" ] } ``` ### bucket system.bucket.benjamin ```json { "name": "benjamin", "memo":"mo bucket", "basis": { "alias": "pdt", "collection": "", "replication": "100", "expires": 20, "prepare_thumbnail_min_width": 1024, "prepare_thumbnail": "" }, "expires": 20, "extends":[{"key":"keepdate", "value":"21"}], "limit": { "min_size": null, "max_size": 102400, "width": null, "height": null, "mime_types": ["image/png", "image/jpeg", "image/png"] }, "last_edit_date": 123143, "last_edit_user":"by46", "overlays": [ {"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8} ], "sizes": [ {"name": "p200", "width":200, "height": 150, "mode": "stretch"}, {"name": "p60", "width":60, "height": 45, "mode": "fit"}, {"name": "p160", "width":160, "height": 120, "mode": "thumbnail"} ] } ``` system.bucket.package ```json { "name": "package", "memo":"mo bucket", "basis": { "collection": "", "replication": "100", "expires": 20, "prepare_thumbnail_min_width": 1024, "prepare_thumbnail": "" }, "expires": 20, "extends":[{"key":"keepdate", "value":"21"}], "limit": { "min_size": null, "max_size": 102400, "width": null, "height": null, "mime_types": ["image/png", "image/jpeg", "image/png"] }, "last_edit_date": 123143, "last_edit_user":"by46", "overlays": [ {"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8} ], "sizes": [ {"name": "p200", "width":200, "height": 150, "mode": "stretch"}, {"name": "p60", "width":60, "height": 45, "mode": "fit"}, {"name": "p160", "width":160, "height": 120, "mode": "thumbnail"} ] } ``` system.bucket.product ```json { "name": "product", "memo":"mo bucket", "basis": { "alias": "pdt", "collection": "", "replication": "100", "expires": 20, "prepare_thumbnail_min_width": 1024, "prepare_thumbnail": "" }, "expires": 20, "extends":[{"key":"keepdate", "value":"21"}], "limit": { "min_size": null, "max_size": 102400, "width": null, "height": null, "mime_types": ["image/png", "image/jpeg", "image/png"] }, "last_edit_date": 123143, "last_edit_user":"by46", "overlays": [ {"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8}, {"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8} ], "sizes": [ {"name": "p200", "width":200, "height": 150, "mode": "stretch"}, {"name": "p60", "width":60, "height": 45, "mode": "fit"}, {"name": "p160", "width":160, "height": 120, "mode": "thumbnail"} ] } ``` ``` package: github.com/by46/whalefs homepage: https://github.com/by46/whalefs license: MIT owners: - name: benjamin.c.yan email: ycs_ctbu_2010@126.com import: - package: github.com/spf13/viper version: ^1.0.2 - package: github.com/sirupsen/logrus version: ^1.0.5 - package: github.com/couchbase/go-couchbase - package: github.com/couchbase/gomemcached - package: github.com/couchbase/goutils - package: github.com/mholt/binding version: ^0.3.0 - package: github.com/spf13/cobra version: 0.0.2 - package: github.com/pkg/errors version: ^0.8.0 ``` ### 普罗米修斯 http://172.16.0.158:9090/graph ### golang 编程规范 ### 接口 大文件上传 #### 初始化上传 POST /benjamin/demo/hello.jpg?uploads { "upload_id":"uuid1" } #### 上传chunk PUT /benjamin/demo/hello.jpg?uploadId=uuid1&partNumber=partNumber1 <multipart-form> </multipart-form> #### 完成上传 POST /benjamin/demo/hello.jpg?uploadId=uuid1 [{ "part_number": "part number 1", "etag": "etag1" }] whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="banner,banner_01" whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="bond,bond_01,contract,contract_01,eInvoice,eInvoice_01" whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="ibuild,ibuild_01" 192.168.110.81:5672,192.168.110.84:5672 spring.rabbitmq.virtual-host = /blockchain spring.rabbitmq.username = ifs_admin spring.rabbitmq.password = ifs@123 ```bash rsync -av -e ssh /Users/benjamin.c.yan/golang/pkg/mod user@remotehost:~/mod ```
29.586364
156
0.583039
yue_Hant
0.093307
91442347ac9791c6716df4e68ad870c0da73ab48
1,906
md
Markdown
docs/database/tometek/README.md
friendly-router/friendly-router
5020871d222b912d280eb90c1f9fc1bde71fad51
[ "0BSD" ]
3
2021-04-01T12:41:55.000Z
2021-04-05T13:07:08.000Z
docs/database/tometek/README.md
friendly-router/friendly-router
5020871d222b912d280eb90c1f9fc1bde71fad51
[ "0BSD" ]
null
null
null
docs/database/tometek/README.md
friendly-router/friendly-router
5020871d222b912d280eb90c1f9fc1bde71fad51
[ "0BSD" ]
null
null
null
--- lang: en-US title: Tometek sidebar: auto draft: true prev: ../ meta: - name: "twitter:card" value: "Friendly Router Project" - name: "twitter:site" value: "https://friendly-router.org/database/tometek" - name: "twitter:title" value: "Database | Tometek" - name: "description" value: | Tometek manufacturer devices and hardware. - name: "twitter:creator" value: "@FriendlyRouter" - name: "twitter:image" value: "https://friendly-router.org/images/switchy-200x200.png" - property: "og:site_name" content: "Friendly Router Project" - property: "og:title" content: "Database | Tometek" - property: "og:type" content: "website" - property: "og:url" content: "https://friendly-router.org/database/tometek" - property: "og:image" content: "https://friendly-router.org/images/switchy-200x200.png" - property: "og:description" content: | Tometek manufacturer devices and hardware. --- > TOM was founded in 2004 by three co-founders who have working > experience in world-class enterprise in Low-voltage Electric and > New energy & Computing application & Telecommunication field while > combined 10+ industry experience in China. In October 2013 TOM and > his partner provided power facilities and IT solution to power the > 27th SEA Games in 4 cities in Myanmar. > > -- [Official Website](http://www.tometek.com/about.asp?id=207) ## Products | Name | Doc | Test | Bench | Hack | Note | |---------------------------|-----|------|-------|------|------| | [max-sppc](max-sppc/) | [max-tsppc](max-tsppc/) | [max-tsssc](max-tsssc/) | [max-tts](max-tts/) | [pro-sccc](pro-sccc/) | [pro-sscc](pro-sscc/) | [pro-sssc](pro-sssc/) | [pro-sc](pro-sc/) | [lite-cc](lite-cc/) | [lite-cccc](lite-cccc/) ## Archives ## References and Resources * [http://www.tometek.com](http://www.tometek.com)
27.623188
69
0.641658
eng_Latn
0.584988
9144861c3ea492cc1437a3a22b1e8c3a60e62d72
1,500
md
Markdown
README.md
Szlavicsek/Quiz
7e70d0e32db0a1da5c1cbee36dd60a435c49a522
[ "MIT" ]
null
null
null
README.md
Szlavicsek/Quiz
7e70d0e32db0a1da5c1cbee36dd60a435c49a522
[ "MIT" ]
null
null
null
README.md
Szlavicsek/Quiz
7e70d0e32db0a1da5c1cbee36dd60a435c49a522
[ "MIT" ]
null
null
null
# Quizzit app A quiz game project using the [Open Trivia API](https://opentdb.com/api_config.php) ## Getting Started To get started, clone the repo to your local machine and install the dependencies as listed below. Please note that for the Google login api to work properly, you will also need: * to configure the game as a new project on Google (you can do so [here](https://developers.google.com/identity/sign-in/web/sign-in)) * a development url to register within the Google dev console for your project (which is other than your local host's url, else it won't work either) After registering, replace the google client id in the "google-signin-client_id" meta tag with yours. For testing purposes, registering a new project is not mandatory, skipping the process above will only affect the corresponding functionality of the app. ### Installing prerequisites The project files are bundled with webpack. To install it, run the following code in your terminal: ``` npm install --save-dev webpack-cli ``` To manage audio files properly on all devices, you will also need the [howlerjs](https://github.com/goldfire/howler.js) library: ``` npm install howler ``` Once the dependencies are installed, you can open the project with live server and run `npm run dev` which will watch and bundle the files automatically during development. To build the production version, you can run `npm run build`. ## Authors Pataki Péter ## License This project is licensed under the MIT License.
35.714286
178
0.772
eng_Latn
0.99651
9144e74aef19e60a189a4f3a507b250bbec82a04
39
md
Markdown
readme.md
eladzlot/minno-sequencer
0c34902ce359b5e4850089c5ad341f0cd4f58d4e
[ "Apache-2.0" ]
1
2021-04-13T05:03:38.000Z
2021-04-13T05:03:38.000Z
readme.md
eladzlot/minno-sequencer
0c34902ce359b5e4850089c5ad341f0cd4f58d4e
[ "Apache-2.0" ]
5
2019-11-20T17:12:55.000Z
2022-03-02T05:05:16.000Z
readme.md
eladzlot/minno-sequencer
0c34902ce359b5e4850089c5ad341f0cd4f58d4e
[ "Apache-2.0" ]
null
null
null
# Minno sequencer The minno sequencer
9.75
19
0.794872
eng_Latn
0.951134
9145812f2f6f975b3760b2daec93decd6b6e84aa
2,025
md
Markdown
docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md
turibbio/docs.it-it
2212390575baa937d6ecea44d8a02e045bd9427c
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md
turibbio/docs.it-it
2212390575baa937d6ecea44d8a02e045bd9427c
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md
turibbio/docs.it-it
2212390575baa937d6ecea44d8a02e045bd9427c
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Modalità di selezione di un controllo Button ms.date: 03/30/2017 helpviewer_keywords: - Button control [Windows Forms], selecting ms.assetid: fe2fc058-5118-4f70-b264-6147d64a7a8d ms.openlocfilehash: 145166d182f1ec51068ab3e0c23c12b471b69231 ms.sourcegitcommit: de17a7a0a37042f0d4406f5ae5393531caeb25ba ms.translationtype: MT ms.contentlocale: it-IT ms.lasthandoff: 01/24/2020 ms.locfileid: "76740011" --- # <a name="ways-to-select-a-windows-forms-button-control"></a>Modalità di selezione di un controllo Button Windows Form È possibile selezionare un pulsante Windows Forms nei modi seguenti: - Utilizzare un mouse per fare clic sul pulsante. - Richiama l'evento <xref:System.Windows.Forms.Control.Click> del pulsante nel codice. - Spostare lo stato attivo sul pulsante premendo il tasto TAB, quindi scegliere il pulsante premendo la barra SPAZIAtrice o invio. - Premere il tasto di accesso (ALT + la lettera sottolineata) per il pulsante. Per altre informazioni sulle chiavi di accesso, vedere [procedura: creare chiavi di accesso per controlli Windows Forms](how-to-create-access-keys-for-windows-forms-controls.md). - Se il pulsante è il pulsante "Accetto" del modulo, premendo invio viene scelto il pulsante, anche se un altro controllo ha lo stato attivo, ad eccezione del fatto che un altro controllo è un altro pulsante, una casella di testo a più righe o un controllo personalizzato che intrappola il tasto INVIO. - Se il pulsante è il pulsante "Annulla" del modulo, premendo ESC viene scelto il pulsante anche se un altro controllo ha lo stato attivo. - Chiamare il metodo <xref:System.Windows.Forms.Button.PerformClick%2A> per selezionare il pulsante a livello di codice. ## <a name="see-also"></a>Vedere anche - [Panoramica sul controllo Button](button-control-overview-windows-forms.md) - [Procedura: Rispondere alla selezione dei pulsanti di Windows Form](how-to-respond-to-windows-forms-button-clicks.md) - [Controllo Button](button-control-windows-forms.md)
56.25
304
0.784691
ita_Latn
0.994376
9145d227d380c093e845def87ac0a8db15ec55d1
6,400
md
Markdown
_posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
2
2019-02-28T03:47:33.000Z
2020-04-06T07:49:53.000Z
_posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
null
null
null
_posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
null
null
null
--- layout: post comments: true categories: Other --- ## Download Miss rosie apos s spice of life quilts leisure arts book "I can't imagine whole cities burning. could see the silver drops pooling on his tongue before he swallowed. He wouldn't need the bottle any more, and he quickly slipped inside. D and Micky at the kitchen gardening supplies? " as possible to miss rosie apos s spice of life quilts leisure arts guests, leaving her employed she sits on miss rosie apos s spice of life quilts leisure arts part of the skin and stretches out the other The pending storm had broken, he found in Zedd the welcome reassurance that it was all right to be you had to do with a good hard-packed clay miss rosie apos s spice of life quilts leisure arts was sweep it and now and then sprinkle it to keep couldn't see through you. I'm not sure that God, it's mostly lies, O my masters, Sarytschev says that at St, he had taken a preventive dose of paregoric, Sarytschev says that at St. He is unable to work up "Not. Nevertheless the resemblance is so strong that he must be a how he might ever again trust anyone sufficiently to take the wedding Without breaking stride, but those were daylight. How strange. When Micky heard miss rosie apos s spice of life quilts leisure arts pet name, so acidic that his gums burned from contact with tortured beyond endurance, but a would-be stage magician eventually needed a mentor to campsites with power-and-water hookups to motor homes and travel trailers. with someone headed for a more populous area that will provide even better Throughout the morning, when he was out from under the tower. "When you're a star," she said once, macaroni and cheese, Geneva gone. tip?" "Like most of us these days, she'd she must have noticed it. A less self-controlled man might have seized a nearby bronze vase-fashioned to resemble miss rosie apos s spice of life quilts leisure arts stool-and stuffed her into it or vice versa? They broke a law that must not be broken! Nearing the house, _although the safety notes. " had eons ago circled too close to the treacherous bogs that swallowed them by no louder than before, an obituary should be published one day or the At the sofabed again. Anselmo's Orphanage, more important, wise men, my caseload being the new Adam or with a yearning for fatherhood, more efficient and more thoroughly reasoned, more relaxed, for those who hate because they have no recognition of the Leilani? A shudder, when mourners gone. A police officer. I can now inform you that, the king summoned his vizier and bade him tell the story of the king who lost kingdom and wife and wealth, but in a way that slipped subliminally into the mind of the beholder and gripped it by its deepest roots, There was some strength in him or with him. Barty came out of the house with the library copy of Podkayne Of Mary, i. " might enable the magic to repeat. " shapes, I'd crawl in there again, regardless of what if she miss rosie apos s spice of life quilts leisure arts been merely frightened. Micky clawed in miss rosie apos s spice of life quilts leisure arts, Hal. "You behave yourself tonight," Celestina said. Where and when?" Hanlon said. Now he had no hope of resisting Gelluk in any way. He'd read it in a novel, as if she were a high-school girl and he were her with utmost consideration, little sticky spots, but at last tore it off, which would explain the stonecarver's accelerated service, clatters as he climbs across it. He himself is not a hunter, was there tap water?" of the crystal rended reflections of the flames into red-orange-yellow-green- with morning drinking or perhaps with drinking binges at any hour, and Without the advantage of surprise, and he believes that he could head of a monster, to cover her discomfort at how close Micky had come to the truth, Curtis watches the receding figure until it's clear the man won't attempt to sneak back, ii. Nolan had shrugged, opening herself to the anger and hatred with which she had regarded this child in the operating room, but not by the Circle of Friends. This always proved to be a had of bold strategy, the mind had a thermostat of its own, called, sprints on, but I'm pissed. social organisation. There were people of the Hand there whom he trusted. TETGALES, 1741. Gelluk was powerful, they used them on pickles on the side, she had reached him even though he didn't want to become involved, he backed out. Or the kid might have been placed for adoption through some baby brokerage in it strictly for the money. Just forget the busload of nuns smashed on the tracks, you were depressed and feeling hopeless. You don't have to worry about me. What was Simon's fee?" They're illegal but not hard to find. " She looked at me through her lashes. The sail struck the water, with the old wizard, Chapter 71 anyone but herself, to cover her discomfort at how close Micky had come to the truth. Along with the bird now described there occur, the, I take grasp. "Do I have a cloud inside me now?" breast of darkness. thought he had no room to plant the seeds for any more? Shaw, 'Whoso stumbleth on a stone and returneth thereto, Curtis retreats from the dog and from the motor home, i. They're most likely fast approaching from the other side of the vehicle. of cold was the same; that is to say, he saw them spread out the gold before them and play with it and heard one of them say, he looked at her as he spoke. Working with patching kits and lasers to cut the tough material, and I am very much at home. In its "More vanilla Coke, taking with them a few bewildered secretaries and office workers that they had bumped into on the way. "She was a good cow, of course? It was clear by this time that the Organizer had no intention of settling for a smaller package, which had all been partially crushed by that pewter candlestick. " He shrugged. " struck numb and mute by the conflict between yearning and inexperience. in a _baydar_ and kill them all but two. together as if with fine-draw stitches. the beginning of the seventeenth century reproduced below, Mrs. "What I hope I found there was direction, by rolling onto my back. were buying Grandma's locally famous black bean-and-corn salsa. 9 -1. "Good thing we weren't playing Russian roulette," Leilani said. It was not until afterwards that we of aspirin! I'm writing to you about an In addition to the firearm, one Maria.
711.111111
6,275
0.78625
eng_Latn
0.999962
91460c75c66d3563f4add5912fbe94e1f77ccf30
15,441
md
Markdown
articles/virtual-machines/windows/tutorial-create-vmss.md
changeworld/azure-docs.it-
34f70ff6964ec4f6f1a08527526e214fdefbe12a
[ "CC-BY-4.0", "MIT" ]
1
2017-06-06T22:50:05.000Z
2017-06-06T22:50:05.000Z
articles/virtual-machines/windows/tutorial-create-vmss.md
changeworld/azure-docs.it-
34f70ff6964ec4f6f1a08527526e214fdefbe12a
[ "CC-BY-4.0", "MIT" ]
41
2016-11-21T14:37:50.000Z
2017-06-14T20:46:01.000Z
articles/virtual-machines/windows/tutorial-create-vmss.md
changeworld/azure-docs.it-
34f70ff6964ec4f6f1a08527526e214fdefbe12a
[ "CC-BY-4.0", "MIT" ]
7
2016-11-16T18:13:16.000Z
2017-06-26T10:37:55.000Z
--- title: 'Esercitazione: Creare un set di scalabilità di macchine virtuali Windows' description: Informazioni su come usare Azure PowerShell per creare e distribuire un'applicazione a disponibilità elevata nelle VM Windows usando un set di scalabilità di macchine virtuali author: ju-shim ms.author: jushiman ms.topic: tutorial ms.service: virtual-machine-scale-sets ms.subservice: windows ms.date: 11/30/2018 ms.reviewer: mimckitt ms.custom: mimckitt, devx-track-azurepowershell ms.openlocfilehash: b3853ddc71d1a9be26b2492764a9b341446e0eeb ms.sourcegitcommit: f28ebb95ae9aaaff3f87d8388a09b41e0b3445b5 ms.translationtype: MT ms.contentlocale: it-IT ms.lasthandoff: 03/29/2021 ms.locfileid: "89078742" --- # <a name="tutorial-create-a-virtual-machine-scale-set-and-deploy-a-highly-available-app-on-windows-with-azure-powershell"></a>Esercitazione: Creare un set di scalabilità di macchine virtuali e distribuire un'app a disponibilità elevata in Windows con Azure PowerShell Un set di scalabilità di macchine virtuali consente di distribuire e gestire un set di macchine virtuali identiche con scalabilità automatica. È possibile adattare manualmente il numero di macchine virtuali nel set di scalabilità, nonché definire regole di scalabilità automatica in base all'utilizzo della CPU, alla richiesta di memoria o al traffico di rete. In questa esercitazione viene distribuito un set di scalabilità di macchine virtuali in Azure e viene spiegato come: > [!div class="checklist"] > * Usare l'estensione dello script personalizzata per definire un sito IIS da ridimensionare > * Creare un bilanciamento del carico per il set di scalabilità > * Creare un set di scalabilità di macchine virtuali > * Aumentare o diminuire il numero di istanze in un set di scalabilità > * Creare regole di scalabilità automatica ## <a name="launch-azure-cloud-shell"></a>Avviare Azure Cloud Shell Azure Cloud Shell è una shell interattiva gratuita che può essere usata per eseguire la procedura di questo articolo. Include strumenti comuni di Azure preinstallati e configurati per l'uso con l'account. Per aprire Cloud Shell, basta selezionare **Prova** nell'angolo superiore destro di un blocco di codice. È anche possibile avviare Cloud Shell in una scheda separata del browser visitando [https://shell.azure.com/powershell](https://shell.azure.com/powershell). Selezionare **Copia** per copiare i blocchi di codice, incollarli in Cloud Shell e premere INVIO per eseguirli. ## <a name="scale-set-overview"></a>Informazioni generali sui set di scalabilità Un set di scalabilità di macchine virtuali consente di distribuire e gestire un set di macchine virtuali identiche con scalabilità automatica. Le macchine virtuali di un set di scalabilità vengono distribuite in domini logici di errore e di aggiornamento in uno o più *gruppi di posizionamento*. I gruppi di posizionamento sono gruppi di macchine virtuali configurate in modo simile, analoghi ai [set di disponibilità](tutorial-availability-sets.md). Le VM vengono create in base alle esigenze in un set di scalabilità. È possibile definire regole di scalabilità automatica per controllare le modalità e i tempi di aggiunta e rimozione delle VM dal set di scalabilità. Queste regole possono essere attivate in base a determinate metriche, ad esempio il carico della CPU, l'utilizzo della memoria o il traffico di rete. I set di scalabilità supportano fino a 1000 macchine virtuali quando si usa un'immagine della piattaforma Azure. Per i carichi di lavoro con requisiti significativi di installazione o personalizzazione di VM, si consiglia di [creare un'immagine di VM personalizzata](tutorial-custom-images.md). È possibile creare fino a 600 macchine virtuali in un set di scalabilità quando si usa un'immagine personalizzata. ## <a name="create-a-scale-set"></a>Creare un set di scalabilità Creare un set di scalabilità di macchine virtuali con [New-AzVmss](/powershell/module/az.compute/new-azvmss). L'esempio seguente crea un set di scalabilità denominato *myScaleSet* che usa l'immagine della piattaforma *Windows Server 2016 Datacenter*. Vengono create automaticamente le risorse di rete di Azure per la rete virtuale, l'indirizzo IP pubblico e il bilanciamento del carico. Quando richiesto, è possibile impostare le credenziali amministrative per le istanze di macchina virtuale nel set di scalabilità: ```azurepowershell-interactive New-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -Location "EastUS" ` -VMScaleSetName "myScaleSet" ` -VirtualNetworkName "myVnet" ` -SubnetName "mySubnet" ` -PublicIpAddressName "myPublicIPAddress" ` -LoadBalancerName "myLoadBalancer" ` -UpgradePolicyMode "Automatic" ``` La creazione e la configurazione di tutte le macchine virtuali e risorse del set di scalabilità richiedono alcuni minuti. ## <a name="deploy-sample-application"></a>Distribuire l'applicazione di esempio Per testare il set di scalabilità, installare un'applicazione Web di base. Usare l'estensione script personalizzati di Azure per scaricare ed eseguire uno script che installa IIS nelle istanze di macchina virtuale. Questa estensione è utile per la configurazione post-distribuzione, l'installazione di software o qualsiasi altra attività di configurazione o gestione. Per altre informazioni, vedere [Panoramica dell'estensione script personalizzata](../extensions/custom-script-windows.md). Usare l'estensione script personalizzati per installare un server Web IIS di base. Applicare l'estensione di script personalizzati che installa IIS nel modo seguente: ```azurepowershell-interactive # Define the script for your Custom Script Extension to run $publicSettings = @{ "fileUris" = (,"https://raw.githubusercontent.com/Azure-Samples/compute-automation-configurations/master/automate-iis.ps1"); "commandToExecute" = "powershell -ExecutionPolicy Unrestricted -File automate-iis.ps1" } # Get information about the scale set $vmss = Get-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" # Use Custom Script Extension to install IIS and configure basic website Add-AzVmssExtension -VirtualMachineScaleSet $vmss ` -Name "customScript" ` -Publisher "Microsoft.Compute" ` -Type "CustomScriptExtension" ` -TypeHandlerVersion 1.8 ` -Setting $publicSettings # Update the scale set and apply the Custom Script Extension to the VM instances Update-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -Name "myScaleSet" ` -VirtualMachineScaleSet $vmss ``` ## <a name="allow-traffic-to-application"></a>Consentire il traffico verso l'applicazione Per consentire l'accesso all'applicazione Web di base, creare un gruppo di sicurezza di rete con [New-AzNetworkSecurityRuleConfig](/powershell/module/az.network/new-aznetworksecurityruleconfig) e [New-AzNetworkSecurityGroup](/powershell/module/az.network/new-aznetworksecuritygroup). Per altre informazioni, vedere [Rete per i set di scalabilità di macchine virtuali](../../virtual-machine-scale-sets/virtual-machine-scale-sets-networking.md). ```azurepowershell-interactive # Get information about the scale set $vmss = Get-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" #Create a rule to allow traffic over port 80 $nsgFrontendRule = New-AzNetworkSecurityRuleConfig ` -Name myFrontendNSGRule ` -Protocol Tcp ` -Direction Inbound ` -Priority 200 ` -SourceAddressPrefix * ` -SourcePortRange * ` -DestinationAddressPrefix * ` -DestinationPortRange 80 ` -Access Allow #Create a network security group and associate it with the rule $nsgFrontend = New-AzNetworkSecurityGroup ` -ResourceGroupName "myResourceGroupScaleSet" ` -Location EastUS ` -Name myFrontendNSG ` -SecurityRules $nsgFrontendRule $vnet = Get-AzVirtualNetwork ` -ResourceGroupName "myResourceGroupScaleSet" ` -Name myVnet $frontendSubnet = $vnet.Subnets[0] $frontendSubnetConfig = Set-AzVirtualNetworkSubnetConfig ` -VirtualNetwork $vnet ` -Name mySubnet ` -AddressPrefix $frontendSubnet.AddressPrefix ` -NetworkSecurityGroup $nsgFrontend Set-AzVirtualNetwork -VirtualNetwork $vnet # Update the scale set and apply the Custom Script Extension to the VM instances Update-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -Name "myScaleSet" ` -VirtualMachineScaleSet $vmss ``` ## <a name="test-your-scale-set"></a>Testare il set di scalabilità Per verificare il funzionamento del set di scalabilità, ottenere l'indirizzo IP pubblico del servizio di bilanciamento del carico con [Get-AzPublicIPAddress](/powershell/module/az.network/get-azpublicipaddress). L'esempio seguente mostra l'indirizzo IP per *myPublicIP* creato come parte del set di scalabilità: ```azurepowershell-interactive Get-AzPublicIPAddress ` -ResourceGroupName "myResourceGroupScaleSet" ` -Name "myPublicIPAddress" | select IpAddress ``` Immettere l'indirizzo IP pubblico in un Web browser. Verrà visualizzata l'app Web, con il nome host della macchina virtuale a cui il servizio di bilanciamento del carico ha distribuito il traffico: ![Esecuzione del sito IIS](./media/tutorial-create-vmss/running-iis-site.png) Per verificare il funzionamento del set di scalabilità, è possibile imporre l'aggiornamento del Web browser per visualizzare la distribuzione del traffico da parte del bilanciamento del carico tra tutte le macchine virtuali che eseguono l'app. ## <a name="management-tasks"></a>Attività di gestione Nel ciclo di vita del set di scalabilità, potrebbe essere necessario eseguire una o più attività di gestione. Si potrebbe anche voler creare script per automatizzare le attività di ciclo di vita. Azure PowerShell offre un modo rapido per eseguire queste operazioni. Di seguito vengono illustrate alcune attività comuni. ### <a name="view-vms-in-a-scale-set"></a>Visualizzare le macchine virtuali in un set di scalabilità Per visualizzare un elenco delle istanze di VM in un set di scalabilità, usare [Get-AzVmssVM](/powershell/module/az.compute/get-azvmssvm) come illustrato di seguito: ```azurepowershell-interactive Get-AzVmssVM ` -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" ``` L'output di esempio seguente mostra due istanze di VM nel set di scalabilità: ```powershell ResourceGroupName Name Location Sku InstanceID ProvisioningState ----------------- ---- -------- --- ---------- ----------------- MYRESOURCEGROUPSCALESET myScaleSet_0 eastus Standard_DS1_v2 0 Succeeded MYRESOURCEGROUPSCALESET myScaleSet_1 eastus Standard_DS1_v2 1 Succeeded ``` Per visualizzare altre informazioni su un'istanza di VM specifica, aggiungere il parametro `-InstanceId` a [Get-AzVmssVM](/powershell/module/az.compute/get-azvmssvm). L'esempio seguente consente di visualizzare informazioni sull'istanza di macchina virtuale *1*: ```azurepowershell-interactive Get-AzVmssVM ` -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" ` -InstanceId "1" ``` ### <a name="increase-or-decrease-vm-instances"></a>Aumentare o diminuire le istanze delle macchine virtuali Per visualizzare il numero di istanze attualmente presente in un set di scalabilità, usare [Get-AzVmss](/powershell/module/az.compute/get-azvmss) ed eseguire una query su *sku.capacity*: ```azurepowershell-interactive Get-AzVmss -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" | ` Select -ExpandProperty Sku ``` È quindi possibile aumentare o ridurre manualmente il numero di macchine virtuali nel set di scalabilità con [Update-AzVmss](/powershell/module/az.compute/update-azvmss). L'esempio seguente imposta il numero di macchine virtuali del set di scalabilità su *3*: ```azurepowershell-interactive # Get current scale set $scaleset = Get-AzVmss ` -ResourceGroupName "myResourceGroupScaleSet" ` -VMScaleSetName "myScaleSet" # Set and update the capacity of your scale set $scaleset.sku.capacity = 3 Update-AzVmss -ResourceGroupName "myResourceGroupScaleSet" ` -Name "myScaleSet" ` -VirtualMachineScaleSet $scaleset ``` Sono necessari alcuni minuti per aggiornare il numero specificato di istanze del set di scalabilità. ### <a name="configure-autoscale-rules"></a>Configurare le regole di scalabilità automatica Anziché scalare manualmente il numero di istanze del set di scalabilità, si definiscono regole di scalabilità automatica. Queste regole monitorano le istanze nel set di scalabilità e rispondono di conseguenza in base alle metriche e alle soglie definite. L'esempio seguente aumenta il numero di istanze di uno quando il carico della CPU medio è maggiore del 60% per un periodo di 5 minuti. Se il carico della CPU medio scende poi al di sotto del 30% per un periodo di 5 minuti, le istanze vengono ridotte di una istanza: ```azurepowershell-interactive # Define your scale set information $mySubscriptionId = (Get-AzSubscription)[0].Id $myResourceGroup = "myResourceGroupScaleSet" $myScaleSet = "myScaleSet" $myLocation = "East US" $myScaleSetId = (Get-AzVmss -ResourceGroupName $myResourceGroup -VMScaleSetName $myScaleSet).Id # Create a scale up rule to increase the number instances after 60% average CPU usage exceeded for a 5-minute period $myRuleScaleUp = New-AzAutoscaleRule ` -MetricName "Percentage CPU" ` -MetricResourceId $myScaleSetId ` -Operator GreaterThan ` -MetricStatistic Average ` -Threshold 60 ` -TimeGrain 00:01:00 ` -TimeWindow 00:05:00 ` -ScaleActionCooldown 00:05:00 ` -ScaleActionDirection Increase ` -ScaleActionValue 1 # Create a scale down rule to decrease the number of instances after 30% average CPU usage over a 5-minute period $myRuleScaleDown = New-AzAutoscaleRule ` -MetricName "Percentage CPU" ` -MetricResourceId $myScaleSetId ` -Operator LessThan ` -MetricStatistic Average ` -Threshold 30 ` -TimeGrain 00:01:00 ` -TimeWindow 00:05:00 ` -ScaleActionCooldown 00:05:00 ` -ScaleActionDirection Decrease ` -ScaleActionValue 1 # Create a scale profile with your scale up and scale down rules $myScaleProfile = New-AzAutoscaleProfile ` -DefaultCapacity 2 ` -MaximumCapacity 10 ` -MinimumCapacity 2 ` -Rule $myRuleScaleUp,$myRuleScaleDown ` -Name "autoprofile" # Apply the autoscale rules Add-AzAutoscaleSetting ` -Location $myLocation ` -Name "autosetting" ` -ResourceGroup $myResourceGroup ` -TargetResourceId $myScaleSetId ` -AutoscaleProfile $myScaleProfile ``` Per altre informazioni di progettazione sull'uso della scalabilità automatica, vedere [Procedure consigliate per la scalabilità automatica](/azure/architecture/best-practices/auto-scaling). ## <a name="next-steps"></a>Passaggi successivi In questa esercitazione è stato creato un set di scalabilità di macchine virtuali. Si è appreso come: > [!div class="checklist"] > * Usare l'estensione dello script personalizzata per definire un sito IIS da ridimensionare > * Creare un bilanciamento del carico per il set di scalabilità > * Creare un set di scalabilità di macchine virtuali > * Aumentare o diminuire il numero di istanze in un set di scalabilità > * Creare regole di scalabilità automatica Passare all'esercitazione successiva per maggiori informazioni sui concetti di bilanciamento del carico per le macchine virtuali. > [!div class="nextstepaction"] > [Bilanciare il carico di macchine virtuali](tutorial-load-balancer.md)
53.614583
520
0.786348
ita_Latn
0.968289
9146fad553db1c08e8ff90baad6c2a72579e035f
11,600
md
Markdown
src/pages/posts/2021-05-30T12:00:04-post.md
evanmacbride/reddit-digest
47659c6b52d9b7d74025c517931e107cb2f6be94
[ "MIT" ]
1
2020-02-03T02:35:55.000Z
2020-02-03T02:35:55.000Z
src/pages/posts/2021-05-30T12:00:04-post.md
evanmacbride/reddit-snapshots
00dcad012a949243e7399a45dd9a37720cfe6576
[ "MIT" ]
null
null
null
src/pages/posts/2021-05-30T12:00:04-post.md
evanmacbride/reddit-snapshots
00dcad012a949243e7399a45dd9a37720cfe6576
[ "MIT" ]
null
null
null
--- title: '05/30/21 12:00PM UTC Snapshot' date: '2021-05-30T12:00:04' --- <ul> <h2>Sci/Tech</h2> <li><a href='https://i.redd.it/41rq1hape5271.jpg'><img src='https://b.thumbs.redditmedia.com/UeqXq9JRz54nw-Zlwbk3VeD9mpMV79ZVO3_eXACaRCs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/41rq1hape5271.jpg'>After 4 nights of clouds, I finally captured an incredibly rare atmospheric phenomenon in my astrophoto (OC)</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/MDieterich'>MDieterich</a> in <a href='https://www.reddit.com/r/space'>space</a> 40917 points & 639 <a href='https://www.reddit.com/r/space/comments/nnzc96/after_4_nights_of_clouds_i_finally_captured_an/'>comments</a></div></li> <li><a href='https://www.businessinsider.com/cybersecurity-colonial-pipeline-companies-resemble-preschoolers-running-with-pointy-scissors-2021-5'><img src='https://a.thumbs.redditmedia.com/mKOBTn2htsI14JECVfdBSWrN8ORk6fFwdmzbcmOBMc8.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.businessinsider.com/cybersecurity-colonial-pipeline-companies-resemble-preschoolers-running-with-pointy-scissors-2021-5'>American companies resemble a bunch of preschoolers running with pointy scissors when it comes to cybersecurity</a></div>(businessinsider.com) posted by <a href='https://www.reddit.com/user/wewewawa'>wewewawa</a> in <a href='https://www.reddit.com/r/technology'>technology</a> 8775 points & 363 <a href='https://www.reddit.com/r/technology/comments/nnoofg/american_companies_resemble_a_bunch_of/'>comments</a></div></li> <li><a href='https://www.businessinsider.com/unredacted-google-lawsuit-docs-detail-efforts-to-collect-user-location-2021-5'><svg version='1.1' viewBox='-34 -14 104 64' preserveAspectRatio='xMidYMid meet' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'> <title>link thumbnail</title> <path d='M32,4H4A2,2,0,0,0,2,6V30a2,2,0,0,0,2,2H32a2,2,0,0,0,2-2V6A2,2,0,0,0,32,4ZM4,30V6H32V30Z'></path> <path d='M8.92,14a3,3,0,1,0-3-3A3,3,0,0,0,8.92,14Zm0-4.6A1.6,1.6,0,1,1,7.33,11,1.6,1.6,0,0,1,8.92,9.41Z'></path> <path d='M22.78,15.37l-5.4,5.4-4-4a1,1,0,0,0-1.41,0L5.92,22.9v2.83l6.79-6.79L16,22.18l-3.75,3.75H15l8.45-8.45L30,24V21.18l-5.81-5.81A1,1,0,0,0,22.78,15.37Z'></path> </svg></a><div><div class='linkTitle'><a href='https://www.businessinsider.com/unredacted-google-lawsuit-docs-detail-efforts-to-collect-user-location-2021-5'>'Apple is eating our lunch': Google employees admit in lawsuit that the company made it nearly impossible for users to keep their location private</a></div>(businessinsider.com) posted by <a href='https://www.reddit.com/user/giantyetifeet'>giantyetifeet</a> in <a href='https://www.reddit.com/r/tech'>tech</a> 2863 points & 185 <a href='https://www.reddit.com/r/tech/comments/nnvejg/apple_is_eating_our_lunch_google_employees_admit/'>comments</a></div></li> <li><a href='https://www.reddit.com/gallery/nnomas'><img src='https://a.thumbs.redditmedia.com/gFC7bCfyKhf3yL2gZD7Ke28JAR0hE5lJRSuDx59hcf0.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nnomas'>No longer thought to be the smallest-known dinosaur, this bird-like lizard skull, trapped in Burmese amber nearly 100 million years ago, has been subject to a great deal of controversy because of its unusual combination of features, and it continues to be a very interesting specimen to science.</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/KimCureAll'>KimCureAll</a> in <a href='https://www.reddit.com/r/Naturewasmetal'>Naturewasmetal</a> 2769 points & 63 <a href='https://www.reddit.com/r/Naturewasmetal/comments/nnomas/no_longer_thought_to_be_the_smallestknown/'>comments</a></div></li> <li><a href='https://i.redd.it/a5it7v4mc2271.jpg'><img src='https://b.thumbs.redditmedia.com/S0eEZ4cvYHUVwmCGrfPdh6kLcEU84Ujc4-u5H0Xi_HI.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/a5it7v4mc2271.jpg'>Whose THICC legs are these?</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/Onnisciente'>Onnisciente</a> in <a href='https://www.reddit.com/r/whatsthisbug'>whatsthisbug</a> 2119 points & 53 <a href='https://www.reddit.com/r/whatsthisbug/comments/nnnqjj/whose_thicc_legs_are_these/'>comments</a></div></li> <li><a href='https://www.theverge.com/2021/5/29/22460070/google-difficult-android-privacy-settings-arizona'><img src='https://b.thumbs.redditmedia.com/z9bKr-WSlEEqJ1kkjh7u47Df4HcaX4Mh0zXKWqdqqKo.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.theverge.com/2021/5/29/22460070/google-difficult-android-privacy-settings-arizona'>Google reportedly made it difficult for smartphone users to find privacy settings</a></div>(theverge.com) posted by <a href='https://www.reddit.com/user/kry_some_more'>kry_some_more</a> in <a href='https://www.reddit.com/r/technology'>technology</a> 5863 points & 233 <a href='https://www.reddit.com/r/technology/comments/no0ruf/google_reportedly_made_it_difficult_for/'>comments</a></div></li> <h2>Maker</h2> <li><a href='https://www.reddit.com/gallery/nns4st'><img src='https://b.thumbs.redditmedia.com/Y_OotEvmZ7ko0zvxay56aUaY9YGKgKN64miD1FsfhQs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nns4st'>No it’s wasn’t photoshopped and yes it’s a Bust! Happy now?</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/evilmaul'>evilmaul</a> in <a href='https://www.reddit.com/r/3Dprinting'>3Dprinting</a> 5483 points & 317 <a href='https://www.reddit.com/r/3Dprinting/comments/nns4st/no_its_wasnt_photoshopped_and_yes_its_a_bust/'>comments</a></div></li> <li><a href='https://i.imgur.com/Uyr5bHL.jpg'><img src='https://b.thumbs.redditmedia.com/_WBNm4FJVs_x6b7ZaKqA7IttEEKgwm5KO74Pv8VMkIs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.imgur.com/Uyr5bHL.jpg'>Array methods in JavaScript. Original author unknown.</a></div>(i.imgur.com) posted by <a href='https://www.reddit.com/user/imAmarok'>imAmarok</a> in <a href='https://www.reddit.com/r/webdev'>webdev</a> 4170 points & 115 <a href='https://www.reddit.com/r/webdev/comments/nnwtf6/array_methods_in_javascript_original_author/'>comments</a></div></li> <li><a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'><svg version='1.1' viewBox='-34 -12 104 64' preserveAspectRatio='xMidYMid slice' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'> <title>text link thumbnail</title> <path d='M12.19,8.84a1.45,1.45,0,0,0-1.4-1h-.12a1.46,1.46,0,0,0-1.42,1L1.14,26.56a1.29,1.29,0,0,0-.14.59,1,1,0,0,0,1,1,1.12,1.12,0,0,0,1.08-.77l2.08-4.65h11l2.08,4.59a1.24,1.24,0,0,0,1.12.83,1.08,1.08,0,0,0,1.08-1.08,1.64,1.64,0,0,0-.14-.57ZM6.08,20.71l4.59-10.22,4.6,10.22Z'> </path> <path d='M32.24,14.78A6.35,6.35,0,0,0,27.6,13.2a11.36,11.36,0,0,0-4.7,1,1,1,0,0,0-.58.89,1,1,0,0,0,.94.92,1.23,1.23,0,0,0,.39-.08,8.87,8.87,0,0,1,3.72-.81c2.7,0,4.28,1.33,4.28,3.92v.5a15.29,15.29,0,0,0-4.42-.61c-3.64,0-6.14,1.61-6.14,4.64v.05c0,2.95,2.7,4.48,5.37,4.48a6.29,6.29,0,0,0,5.19-2.48V26.9a1,1,0,0,0,1,1,1,1,0,0,0,1-1.06V19A5.71,5.71,0,0,0,32.24,14.78Zm-.56,7.7c0,2.28-2.17,3.89-4.81,3.89-1.94,0-3.61-1.06-3.61-2.86v-.06c0-1.8,1.5-3,4.2-3a15.2,15.2,0,0,1,4.22.61Z'> </path> </svg></a><div><div class='linkTitle'><a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'>RTX 3090 but still I have micro stutters</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/furytayx'>furytayx</a> in <a href='https://www.reddit.com/r/buildapc'>buildapc</a> 2176 points & 332 <a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'>comments</a></div></li> <li><a href='https://learnxinyminutes.com/'><svg version='1.1' viewBox='-34 -14 104 64' preserveAspectRatio='xMidYMid meet' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'> <title>link thumbnail</title> <path d='M32,4H4A2,2,0,0,0,2,6V30a2,2,0,0,0,2,2H32a2,2,0,0,0,2-2V6A2,2,0,0,0,32,4ZM4,30V6H32V30Z'></path> <path d='M8.92,14a3,3,0,1,0-3-3A3,3,0,0,0,8.92,14Zm0-4.6A1.6,1.6,0,1,1,7.33,11,1.6,1.6,0,0,1,8.92,9.41Z'></path> <path d='M22.78,15.37l-5.4,5.4-4-4a1,1,0,0,0-1.41,0L5.92,22.9v2.83l6.79-6.79L16,22.18l-3.75,3.75H15l8.45-8.45L30,24V21.18l-5.81-5.81A1,1,0,0,0,22.78,15.37Z'></path> </svg></a><div><div class='linkTitle'><a href='https://learnxinyminutes.com/'>Programming in many languages, and switching back and forth, I was looking for something with syntax and basics to make a working snippet like example file for myself. Found this gem.</a></div>(learnxinyminutes.com) posted by <a href='https://www.reddit.com/user/megaman5'>megaman5</a> in <a href='https://www.reddit.com/r/programming'>programming</a> 471 points & 21 <a href='https://www.reddit.com/r/programming/comments/nnooum/programming_in_many_languages_and_switching_back/'>comments</a></div></li> <h2>Etcetera</h2> <li><a href='https://i.redd.it/46hb188im2271.gif'><img src='https://b.thumbs.redditmedia.com/3_SUfqfXMu4XK5etgZprQao98f3UADXDXDDlCrKrf9U.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/46hb188im2271.gif'>A new bar for the game!</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/Ihor_Hontarenko'>Ihor_Hontarenko</a> in <a href='https://www.reddit.com/r/PixelArt'>PixelArt</a> 1845 points & 53 <a href='https://www.reddit.com/r/PixelArt/comments/nnoso4/a_new_bar_for_the_game/'>comments</a></div></li> <li><a href='https://i.redd.it/s15lquq5o2271.jpg'><img src='https://b.thumbs.redditmedia.com/sAhLwplykSb6ubXDqnhOk38gc4BQAorUmgY09LFK4jk.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/s15lquq5o2271.jpg'>Sci-fi diagnostic equipment? Nah, barcode scanner (Star Trek: Into Darkness, 2013)</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/HalDimond'>HalDimond</a> in <a href='https://www.reddit.com/r/Thatsabooklight'>Thatsabooklight</a> 1200 points & 26 <a href='https://www.reddit.com/r/Thatsabooklight/comments/nnoym2/scifi_diagnostic_equipment_nah_barcode_scanner/'>comments</a></div></li> <li><a href='https://www.reddit.com/gallery/nnsq8v'><img src='https://b.thumbs.redditmedia.com/k4CmxLH_rTH3wHkdspJib-8tyMYST1md9979hvufPiA.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nnsq8v'>A Toyota Coaster based Commuter Draisine Used in Northeastern China by a mining company to transport workers on and off work.</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/Mr_Camhed'>Mr_Camhed</a> in <a href='https://www.reddit.com/r/WeirdWheels'>WeirdWheels</a> 1082 points & 23 <a href='https://www.reddit.com/r/WeirdWheels/comments/nnsq8v/a_toyota_coaster_based_commuter_draisine_used_in/'>comments</a></div></li> <li><a href='https://i.redd.it/36jer9tav7271.jpg'><img src='https://a.thumbs.redditmedia.com/-8bC-x4dNEdGpNeKeOf514B1IotYl3WRPZg_EBZLrT8.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/36jer9tav7271.jpg'>Found on interestingasfuck</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/EntraByte'>EntraByte</a> in <a href='https://www.reddit.com/r/RetroFuturism'>RetroFuturism</a> 382 points & 12 <a href='https://www.reddit.com/r/RetroFuturism/comments/no6l0y/found_on_interestingasfuck/'>comments</a></div></li> </ul>
203.508772
858
0.736552
yue_Hant
0.303362
914716817163ac3062f9d730be9d770c99d9ac83
2,033
md
Markdown
_posts/2011-7-16-UVA12983 The Battle of Chibi.md
FutaRimeWoawaSete/FutaRimeWoawaSete.github.io
714d0ae43929dc5a4672f82e4c1666fa798d3e38
[ "MIT" ]
null
null
null
_posts/2011-7-16-UVA12983 The Battle of Chibi.md
FutaRimeWoawaSete/FutaRimeWoawaSete.github.io
714d0ae43929dc5a4672f82e4c1666fa798d3e38
[ "MIT" ]
null
null
null
_posts/2011-7-16-UVA12983 The Battle of Chibi.md
FutaRimeWoawaSete/FutaRimeWoawaSete.github.io
714d0ae43929dc5a4672f82e4c1666fa798d3e38
[ "MIT" ]
null
null
null
表面看起来是一道计数题,实际是一道 DP 题。 我们首先设 $dp_{i,j}$ 表示长度为 $i$ 并且以 $j$ 结尾的严格上升子序列的个数, 我们经过一定的推导后可以得到一个 DP 转移式: - $ dp_{i,j} = \sum_{k = 1}^{j - 1}dp_{i-1,k}$ 其中 $a_k < a_j$ - 照着这个 $DP$ 转移式写上去我们发现这是一个 $O(n ^ 2m)$ 的 DP 很明显我们过不掉这道题。 这时我们考虑如何优化这个 DP 转移式,毕竟 DP 式推出来了但是时间复杂度过不掉的话基本都是需要优化的,我们发现我们可以先离散所有 $a_i$ 然后用线段树维护前面的 $dp_{i,j - 1}$ 。 不过这里需要注意的是,我们由于只能取 $\sum_{k = 1}^{j - 1}$,所以我们必须一个个加。 也就是这样: ```cpp for(int i = 2 ; i <= m ; i ++) { for(int j = 1 ; j <= n ; j ++) { add(a[j] , dp[i - 1][j]); if(a[j] != 1) dp[i][j] = query(a[j] - 1); if(i == m) Ans += dp[i][j] , Ans %= mod; } } ``` ~~不过由于线段树还是大常数~~,在本人亲测线段树会被卡常后就换成了树状数组来卡这道题,不得不说这道题还是有点卡…… ```cpp #include<iostream> #include<cstdio> #include<cstring> #include<cmath> #include<algorithm> #include<cmath> using namespace std; const int Len = 1e3 + 5; const long long mod = 1e9 + 7; int cnt,n,m,cnts; long long c[Len],a[Len],Ans,dp[Len][Len],lsh[Len]; int lowbit(int x){return x & (-x);} void Clear(){for(int i = 1 ; i <= cnts ; i ++) c[i] = 0;} void add(int x,long long d){for( ; x <= cnts ; x += lowbit(x)) c[x] += d , c[x] %= mod;} long long query(int x){long long res = 0;for( ; x ; x -= lowbit(x)) res += c[x] , res %= mod;return res;} int main() { int T; scanf("%d",&T); while(T --) { cnt ++; Ans = 0; scanf("%d %d",&n,&m); for(int i = 1 ; i <= n ; i ++) { scanf("%lld",&a[i]); lsh[i] = a[i]; dp[1][i] = 1; } sort(lsh + 1 , lsh + 1 + n); cnts = unique(lsh + 1 , lsh + 1 + n) - lsh - 1; for(int i = 1 ; i <= n ; i ++) a[i] = lower_bound(lsh + 1 , lsh + 1 + cnts , a[i]) - lsh; if(m == 1) Ans = n; else { Clear(); for(int i = 2 ; i <= m ; i ++) { for(int j = 1 ; j <= n ; j ++) { add(a[j] , dp[i - 1][j]); if(a[j] != 1) dp[i][j] = query(a[j] - 1); if(i == m) Ans += dp[i][j] , Ans %= mod; } Clear(); } } for(int i = 2 ; i <= m ; i ++) for(int j = 1 ; j <= n ; j ++) dp[i][j] = 0; Clear(); printf("Case #%d: %lld\n",cnt,Ans); } return 0; } ```
23.367816
105
0.495819
eng_Latn
0.108548
9147e13504f9d3d07455230f455e2f8e8e651534
1,455
md
Markdown
README.md
david-mcgillicuddy-moixa/tokio-proto
0b72a1978064e0e087fd8682ccd1aba6064ad6c0
[ "Apache-2.0", "MIT" ]
334
2016-08-27T01:08:35.000Z
2022-03-16T23:29:19.000Z
README.md
david-mcgillicuddy-moixa/tokio-proto
0b72a1978064e0e087fd8682ccd1aba6064ad6c0
[ "Apache-2.0", "MIT" ]
174
2016-08-27T08:57:26.000Z
2018-08-01T19:09:59.000Z
README.md
david-mcgillicuddy-moixa/tokio-proto
0b72a1978064e0e087fd8682ccd1aba6064ad6c0
[ "Apache-2.0", "MIT" ]
100
2016-08-27T00:46:49.000Z
2021-05-14T07:00:32.000Z
# This crate is deprecated! This crate is deprecated without an immediate replacement. Discussion about a successor can be found in [tokio-rs/tokio#118](https://github.com/tokio-rs/tokio/issues/118). # tokio-proto `tokio-proto` makes it easy to implement clients and servers for **request / response** oriented protocols. It takes a [transport] and provides the request / response API. It is a part of the [Tokio] platform. [![Build Status](https://travis-ci.org/tokio-rs/tokio-proto.svg?branch=master)](https://travis-ci.org/tokio-rs/tokio-proto) [Documentation](https://docs.rs/tokio-proto) | [Gitter](https://gitter.im/tokio-rs/tokio) | [Tutorial](https://tokio.rs) [transport]: https://tokio.rs/docs/going-deeper-tokio/transports/ [Tokio]: https://tokio.rs ## Usage First, add this to your `Cargo.toml`: ```toml [dependencies] tokio-proto = { git = "https://github.com/tokio-rs/tokio-proto" } ``` Next, add this to your crate: ```rust extern crate tokio_proto; ``` You can find extensive examples and tutorials at [https://tokio.rs](https://tokio.rs). ## Getting Help If you have questions or need further help getting started, consider joining the chat in our [Gitter Channel](http://gitter.im/tokio-rs/tokio). ## License Tokio is primarily distributed under the terms of both the MIT license and the Apache License (Version 2.0), with portions covered by various BSD-like licenses. See LICENSE-APACHE, and LICENSE-MIT for details.
29.1
172
0.740893
eng_Latn
0.89593
91495e76ceaa936675d9d31fa3399a03e5ed91cc
6,769
md
Markdown
documents/DesignNotes/appearance.md
hangle/Notecard
fdbed0ce0d15e0288794e18680da7360a0daeed7
[ "Apache-2.0" ]
null
null
null
documents/DesignNotes/appearance.md
hangle/Notecard
fdbed0ce0d15e0288794e18680da7360a0daeed7
[ "Apache-2.0" ]
null
null
null
documents/DesignNotes/appearance.md
hangle/Notecard
fdbed0ce0d15e0288794e18680da7360a0daeed7
[ "Apache-2.0" ]
null
null
null
<h1>Appearance Features </h1> <p>Appearance features cover the size, style, color, and font style <br /> of text. It also includes the size and length of the input field, <br /> as well as the appearance features of input characters. The <br /> number of input characters can be limited. The window height <br /> and window width are Appearance features. Activation and <br /> deactivation of the '*' and 'PRIOR' buttons are included <br /> as Appearance features. </p> <p>The Appearance features are established by the default values <br /> of the program as listed below: </p> <pre> key value xlocate 0 //window x screen location ylocate 0 //window y screen location height 300 //window size argument width 400 //window size argument name TimesRoman // name of Font size 14 // pixel size of lettering color black // color of lettering style 1 // 1=normal, 2=bold, 3=italics length 10 // input field length limit 99 // limits the number of input characters asteriskButton on // "on" allows '* button' to be armed (active) priorButton on // "on" allow 'PRIOR button' to be armed (active) </pre> <p>The script command file:</p> <pre> c d now is the time for all good men * end </pre> <p>The text of the 'd' command is displayed in a window <br /> dimensioned as 300 x 400. The text color is black. Its size <br /> is 14, and its font style is TimesRoman. </p> <h3>Ways to override default value.</h3> <p>Asterisk Commands. The '*' command with an appearance <br /> key/value pair, supercedes the corresponding program <br /> default value. </p> <p>The script command file: (ParserValidator.AsteriskCommand)</p> <pre> * width 500 * height 600 * size 22 c d now is the time for all good men * end </pre> <p>The window size changes to 500 x 600, and the text size becomes <br /> '22'. The text color is still black and the font size is still <br /> TimesRoman. However, the introduction of the following '*' <br /> commands can change this. </p> <pre> * color green * name Arial </pre> <p>The '*' commands of the script file remains in effect until <br /> the next script file is executed. </p> <h3>The appearance.ini file.</h3> <p>The 'appearance.ini' file provides values to the script program. <br /> It does not directly involve the Notecard program. The script <br /> program uses the 'appearance.ini' to create the '.struct' files <br /> that are employed by the Notecard program. </p> <p>The 'appearance.ini' file, consisting of key/value pairs. <br /> These key/value pairs change the default values for all <br /> script files in a session. The following is an example of an <br /> 'appearance.ini' file. </p> <pre> appearance.ini ____________________________ | height 320 | width 260 | size 16 | color green </pre> <p>The file changes four default, appearance values. These <br /> values become new default values overriding the program <br /> values. They remain in effect for the session's script <br /> files ( the '* end' command terminates a session). </p> <p>The directory in which the Notecard program is executed is <br /> termed the 'home' directory. The 'appearance.ini' file in <br /> the 'home' directory is applied to the script files in 'home' <br /> directory and to all script files in its subdirectories. <br /> An 'appearance.ini' directory within a 'home' subdirectory <br /> is applied to just the script files in that directory. </p> <p>In the event that 'appearance.ini' files are in both the <br /> 'home' directory and in a 'home' subdirectory, then the <br /> subdirectory 'ini' file supercedes the 'home' 'ini' file, <br /> but only for script files in the subdirectory. </p> <p>The Appearance feature of an '*' command overrides a <br /> corresponding feature of the 'appearance.ini' file. <br /> For example, the following 'appearance.ini' file in the <br /> 'home' directory is: </p> <pre> appearance.ini __________________ | size 12 </pre> <p>The initial default value of 14 is becomes 12 for all <br /> script files in the 'home' directory and its subdirectories. </p> <p>Next, an 'appeaance.ini' file in a subdirectory is:</p> <pre> appearance.ini ____________________ | size 10 </pre> <p>The previous defaults letter size of 14 and 12 become 10 for <br /> just the script files in the subdirectory. In the 'home' <br /> directory and its other subdirectories, the default value <br /> is 12. </p> <p>Finally the following '*' command is added to a script file <br /> in the subdirectory whose 'appearance.ini' file specifies a <br /> size is 10 : </p> <pre> * size 22 </pre> <p>The default letter size is 22 for text in the script file <br /> having the '*' size 22 command. The other script files <br /> have a default letter size of 10. </p> <h3>Two Types of '*' commands.</h3> <p>The '<em>' Appearance commands, such as, '</em> size 22', are <br /> not executed by the Notecard program. The Notecard program <br /> executes the following '*' commands: </p> <pre> * end * continue * save * manage <filename> </pre> <p>The '*' Appearance commands are consumed in the Script <br /> program and are removed. </p> <h3>Script Program</h3> <p>The default values in AsteriskCommand.defaultSetting:Map are copied to 'overrideSetting'. </p> <pre><code>val defaultSetting=Map ( "height"-&gt; "300", //window size argument "width"-&gt; "400", //window size argument "name"-&gt; "TimesRoman",// name of Font "size"-&gt; "14", // pixel size of lettering "color"-&gt; "black", // color of lettering "style"-&gt; "0", // 0=normal, 1=bold, 2=italics "length"-&gt; "10", // used in Display cmd for BoxField "limit"-&gt; "99", // used in Display cmd for BoxField "column"-&gt; "0", // not operational "manage"-&gt; "task" , // to create FramerTask "asteriskButton"-&gt; "on", // "on" allows '* button' to be armed (active) "save"-&gt; "task" , // n to create FramerTask-- save symbolTable data "priorButton"-&gt; "on" , // "on" allow 'PRIOR button' to be armed (active) "end"-&gt; "task", // to create FramerTask-- terminate session "status"-&gt;"task", // display msg in status field "continue"-&gt; "task" // to create CardSetTask-- ) </code></pre>
34.712821
80
0.637465
eng_Latn
0.992658
914a52608291f40eb2adef88dd684802888b5b29
126
md
Markdown
CHANGELOG.md
cxfans/ftplib
9b5bf9fa0f314c294d180380f6a891b581378342
[ "MIT" ]
1
2020-04-13T19:20:08.000Z
2020-04-13T19:20:08.000Z
CHANGELOG.md
cxfans/ftplib
9b5bf9fa0f314c294d180380f6a891b581378342
[ "MIT" ]
null
null
null
CHANGELOG.md
cxfans/ftplib
9b5bf9fa0f314c294d180380f6a891b581378342
[ "MIT" ]
1
2020-07-11T08:53:28.000Z
2020-07-11T08:53:28.000Z
# Change Log of ftplib Library ## [0.1.0] - 2019-11-8 ### Release - Implement basic function for File Transfer Protocol (FTP)
25.2
59
0.706349
kor_Hang
0.431979
914aa84b6b10631b7936d349a26ff6ab8741f004
596
md
Markdown
qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md
DanTupi/personal_setup
911b4951e4d8b78d6ea8ca335229e2e970fda871
[ "MIT" ]
null
null
null
qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md
DanTupi/personal_setup
911b4951e4d8b78d6ea8ca335229e2e970fda871
[ "MIT" ]
null
null
null
qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md
DanTupi/personal_setup
911b4951e4d8b78d6ea8ca335229e2e970fda871
[ "MIT" ]
null
null
null
# bcat's Quefrency 65% layout This is a standard 65% keyboard layout, with a split spacebar, an HHKB-style (split) backspace, media controls in the function layer (centered around the ESDF cluster), and RGB controls in the function layer (on the arrow/nav keys). ## Default layer ![Default layer layout](https://i.imgur.com/gfVTuPO.png) ([KLE](http://www.keyboard-layout-editor.com/#/gists/60a262432bb340b37d364a4424f3037b)) ## Function layer ![Function layer layout](https://i.imgur.com/Fzshl0F.png) ([KLE](http://www.keyboard-layout-editor.com/#/gists/59636898946da51f91fb290f8e078b4d))
33.111111
87
0.761745
eng_Latn
0.549269
914b1662630f79b0c8d5c356f1f4dc2ab8473602
2,082
md
Markdown
README.md
facorazza/Loot_Quotes_Bot
202e3e6b163b92d67af578203501a4857596a64c
[ "Apache-2.0" ]
null
null
null
README.md
facorazza/Loot_Quotes_Bot
202e3e6b163b92d67af578203501a4857596a64c
[ "Apache-2.0" ]
null
null
null
README.md
facorazza/Loot_Quotes_Bot
202e3e6b163b92d67af578203501a4857596a64c
[ "Apache-2.0" ]
null
null
null
# Loot_Quotes_Bot ### Media dei prezzi L'approssimazione più sostanziale di tutto il processo riguarda il calcolo della media dei prezzi degli oggetti. Essendo i prezzi una misura dipendente dal tempo e dai valori passati degli stessi, l'insieme considerato non può essere considerato un **campione**. Per ovviare a ciò utilizzeremo i valori contenuti nell'intervallo di tempo più ristretto possibile, cercando comunque di mantenere una numerosità sostanziale per ogni singolo oggetto, in modo tale da eliminare, o almeno da ridurre drasticamente, la dipendenza tra le misure passate e da rendere la distribuzione il più uniforme possibile per ogni singolo oggetto. ### Incertezza L'incertezza delle quotazioni viene calcolata attraverso la seguente formula: *Formula* Il valore trovato deve essere aggiunto e sottratto al valore della quotazione, trovando, in questo modo, un intervallo di confidenza entro cui il prezzo reale può cadere. *Intervallo di confidenza:* ksigma L'incertezza mostrata dal bot utilizza k=1 che indica una confidenza di circa (simbolo)il 68% che il prezzo cada nell'intervallo. Per esempio: Raddoppiando il valore di k la confidenza dell'intervallo sale a circa il (simbolo)95%. Quindi, riprendendo l'esempio precedente: È possibile aumentare k ulteriormente ma ovviamente ne consegue un aumento dell'intervallo. ### Mediana e quantili Il bot riordina i dati per prezzo in modo crescente e ne calcola la mediana. La mediana è il valore che si trova in mezzo e che divide i prezzi in due insiemi equipartiti. Analogamente il primo e il terzo quartile dividono ulteriormente i due sottoinsiemi appena creati. Questa suddivisione permette di avere un'idea della distribuzione dei dati. Infatti è possibile constatare che il 50% dei prezzi è contenuto nell'intervallo tra il *primo* e il *terzo quartile*. Conoscendo inoltre il prezzo *minimo* si deduce che l'intervallo più piccolo contenete il 25% dei dati si trova tra il minimo e il primo quartile. Analogamente tutto ciò vale anche per l'intervallo superiore, se si conosce il *massimo* dell'insieme.
115.666667
715
0.8122
ita_Latn
0.999969
914b93198659312b955bb474aa94a5dce7583f1e
7,971
md
Markdown
docs/usage/validation_decorator.md
jasujm/pydantic
cc1cb4826c74ac5b651ef2d80c3478428a9950ca
[ "MIT" ]
6
2021-08-11T11:37:59.000Z
2021-11-12T01:33:11.000Z
docs/usage/validation_decorator.md
jasujm/pydantic
cc1cb4826c74ac5b651ef2d80c3478428a9950ca
[ "MIT" ]
189
2020-07-12T08:13:29.000Z
2022-03-28T01:16:29.000Z
docs/usage/validation_decorator.md
jasujm/pydantic
cc1cb4826c74ac5b651ef2d80c3478428a9950ca
[ "MIT" ]
2
2021-11-23T16:28:21.000Z
2021-11-23T16:28:33.000Z
The `validate_arguments` decorator allows the arguments passed to a function to be parsed and validated using the function's annotations before the function is called. While under the hood this uses the same approach of model creation and initialisation; it provides an extremely easy way to apply validation to your code with minimal boilerplate. !!! info "In Beta" The `validate_arguments` decorator is in **beta**, it has been added to *pydantic* in **v1.5** on a **provisional basis**. It may change significantly in future releases and its interface will not be concrete until **v2**. Feedback from the community while it's still provisional would be extremely useful; either comment on [#1205](https://github.com/samuelcolvin/pydantic/issues/1205) or create a new issue. Example of usage: ```py {!.tmp_examples/validation_decorator_main.py!} ``` _(This script is complete, it should run "as is")_ ## Argument Types Argument types are inferred from type annotations on the function, arguments without a type decorator are considered as `Any`. Since `validate_arguments` internally uses a standard `BaseModel`, all types listed in [types](types.md) can be validated, including *pydantic* models and [custom types](types.md#custom-data-types). As with the rest of *pydantic*, types can be coerced by the decorator before they're passed to the actual function: ```py {!.tmp_examples/validation_decorator_types.py!} ``` _(This script is complete, it should run "as is")_ A few notes: - though they're passed as strings, `path` and `regex` are converted to a `Path` object and regex respectively by the decorator - `max` has no type annotation, so will be considered as `Any` by the decorator Type coercion like this can be extremely helpful but also confusing or not desired, see [below](#coercion-and-strictness) for a discussion of `validate_arguments`'s limitations in this regard. ## Function Signatures The decorator is designed to work with functions using all possible parameter configurations and all possible combinations of these: * positional or keyword arguments with or without defaults * variable positional arguments defined via `*` (often `*args`) * variable keyword arguments defined via `**` (often `**kwargs`) * keyword only arguments - arguments after `*,` * positional only arguments - arguments before `, /` (new in python 3.8) To demonstrate all the above parameter types: ```py {!.tmp_examples/validation_decorator_parameter_types.py!} ``` _(This script is complete, it should run "as is")_ ## Using Field to describe function arguments [Field](schema.md#field-customisation) can also be used with `validate_arguments` to provide extra information about the field and validations. In general it should be used in a type hint with [Annotated](schema.md#typingannotated-fields), unless `default_factory` is specified, in which case it should be used as the default value of the field: ```py {!.tmp_examples/validation_decorator_field.py!} ``` _(This script is complete, it should run "as is")_ ## Usage with mypy The `validate_arguments` decorator should work "out of the box" with [mypy](http://mypy-lang.org/) since it's defined to return a function with the same signature as the function it decorates. The only limitation is that since we trick mypy into thinking the function returned by the decorator is the same as the function being decorated; access to the [raw function](#raw-function) or other attributes will require `type: ignore`. ## Validate without calling the function By default, arguments validation is done by directly calling the decorated function with parameters. But what if you wanted to validate them without *actually* calling the function? To do that you can call the `validate` method bound to the decorated function. ```py {!.tmp_examples/validation_decorator_validate.py!} ``` _(This script is complete, it should run "as is")_ ## Raw function The raw function which was decorated is accessible, this is useful if in some scenarios you trust your input arguments and want to call the function in the most performant way (see [notes on performance](#performance) below): ```py {!.tmp_examples/validation_decorator_raw_function.py!} ``` _(This script is complete, it should run "as is")_ ## Async Functions `validate_arguments` can also be used on async functions: ```py {!.tmp_examples/validation_decorator_async.py!} ``` ## Custom Config The model behind `validate_arguments` can be customised using a config setting which is equivalent to setting the `Config` sub-class in normal models. !!! warning The `fields` and `alias_generator` properties of `Config` which allow aliases to be configured are not supported yet with `@validate_arguments`, using them will raise an error. Configuration is set using the `config` keyword argument to the decorator, it may be either a config class or a dict of properties which are converted to a class later. ```py {!.tmp_examples/validation_decorator_config.py!} ``` _(This script is complete, it should run "as is")_ ## Limitations `validate_arguments` has been released on a provisional basis without all the bells and whistles, which may be added later, see [#1205](https://github.com/samuelcolvin/pydantic/issues/1205) for some more discussion of this. In particular: ### Validation Exception Currently upon validation failure, a standard *pydantic* `ValidationError` is raised, see [model error handling](models.md#error-handling). This is helpful since it's `str()` method provides useful details of the error which occurred and methods like `.errors()` and `.json()` can be useful when exposing the errors to end users, however `ValidationError` inherits from `ValueError` **not** `TypeError` which may be unexpected since python would raise a `TypeError` upon invalid or missing arguments. This may be addressed in future by either allow a custom error or raising a different exception by default, or both. ### Coercion and Strictness *pydantic* currently leans on the side of trying to coerce types rather than raise an error if a type is wrong, see [model data conversion](models.md#data-conversion) and `validate_arguments` is no different. See [#1098](https://github.com/samuelcolvin/pydantic/issues/1098) and other issues with the "strictness" label for a discussion of this. If *pydantic* gets a "strict" mode in future, `validate_arguments` will have an option to use this, it may even become the default for the decorator. ### Performance We've made a big effort to make *pydantic* as performant as possible (see [the benchmarks](../benchmarks.md)) and argument inspect and model creation is only performed once when the function is defined, however there will still be a performance impact to using the `validate_arguments` decorator compared to calling the raw function. In many situations this will have little or no noticeable effect, however be aware that `validate_arguments` is not an equivalent or alternative to function definitions in strongly typed languages; it never will be. ### Return Value The return value of the function is not validated against its return type annotation, this may be added as an option in future. ### Config and Validators `fields` and `alias_generator` on custom [`Config`](model_config.md) are not supported, see [above](#custom-config). Neither are [validators](validators.md). ### Model fields and reserved arguments The following names may not be used by arguments since they can be used internally to store information about the function's signature: * `v__args` * `v__kwargs` * `v__positional_only` These names (together with `"args"` and `"kwargs"`) may or may not (depending on the function's signature) appear as fields on the internal *pydantic* model accessible via `.model` thus this model isn't especially useful (e.g. for generating a schema) at the moment. This should be fixable in future as the way error are raised is changed.
42.854839
117
0.774056
eng_Latn
0.998529
914b988cd43f3764aede371c8d9a52b5585c1d56
3,516
md
Markdown
api/Publisher.Shapes.AddTextEffect.md
RichardCory/VBA-Docs
1240462311fb77ee051d4e8b7d7a434d7d020dd3
[ "CC-BY-4.0", "MIT" ]
2
2020-03-09T13:24:12.000Z
2020-03-09T16:19:11.000Z
api/Publisher.Shapes.AddTextEffect.md
MarkFern/VBA-Docs
b84627cc8e24acfd336d1e9761a9ddd58f19d352
[ "CC-BY-4.0", "MIT" ]
null
null
null
api/Publisher.Shapes.AddTextEffect.md
MarkFern/VBA-Docs
b84627cc8e24acfd336d1e9761a9ddd58f19d352
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Shapes.AddTextEffect method (Publisher) keywords: vbapb10.chm2162721 f1_keywords: - vbapb10.chm2162721 ms.prod: publisher api_name: - Publisher.Shapes.AddTextEffect ms.assetid: 21af82f1-d507-3c16-72df-bde1b5e00717 ms.date: 06/08/2017 localization_priority: Normal --- # Shapes.AddTextEffect method (Publisher) Adds a new **Shape** object representing a WordArt object to the specified **Shapes** collection. ## Syntax _expression_.**AddTextEffect**(**_PresetTextEffect_**, **_Text_**, **_FontName_**, **_FontSize_**, **_FontBold_**, **_FontItalic_**, **_Left_**, **_Top_**) _expression_ A variable that represents a **Shapes** object. ## Parameters |Name|Required/Optional|Data type|Description| |:-----|:-----|:-----|:-----| |PresetTextEffect|Required| **MsoPresetTextEffect**|The preset text effect to use. The values of the **MsoPresetTextEffect** constants correspond to the formats listed in the **WordArt Gallery** dialog box (numbered from left to right and from top to bottom).| |Text|Required| **String**|The text to use for the WordArt object.| |FontName|Required| **String**|The name of the font to use for the WordArt object.| |FontSize|Required| **Variant**|The font size to use for the WordArt object. Numeric values are evaluated in points; strings can be in any units supported by Microsoft Publisher (for example, "2.5 in").| |FontBold|Required| **MsoTriState**|Determines whether to format the WordArt text as bold.| |FontItalic|Required| **MsoTriState**|Determines whether to format the WordArt text as italic.| |Left|Required| **Variant**|The position of the left edge of the shape representing the WordArt object.| |Top|Required| **Variant**|The position of the top edge of the shape representing the WordArt object.| ## Return value Shape ## Remarks For the Left and Top parameters, numeric values are evaluated in points; strings can be in any units supported by Publisher (for example, "2.5 in"). The height and width of the WordArt object is determined by its text and formatting. Use the **[TextEffect](Publisher.Shape.TextEffect.md)** property to return a **[TextEffectFormat](Publisher.TextEffectFormat.md)** object whose properties can be used to edit an existing WordArt object. The PresetTextEffect parameter can be one of the **[MsoPresetTextEffect](Office.MsoPresetTextEffect.md)** constants declared in the Microsoft Office type library. The **msoTextEffectMixed** constant is not supported. The FontBold parameter can be one of the **MsoTriState** constants declared in the Microsoft Office type library and shown in the following table. |Constant|Description| |:-----|:-----| | **msoFalse**|Do not format the WordArt text as bold.| | **msoTrue**|Format the WordArt text as bold.| The FontItalic parameter can be one of the **MsoTriState** constants declared in the Microsoft Office type library and shown in the following table. |Constant|Description| |:-----|:-----| | **msoFalse**| Do not format the WordArt text as italic.| | **msoTrue**|Format the WordArt text as italic.| ## Example The following example adds a WordArt object to the first page of the active publication. ```vb Dim shpWordArt As Shape Set shpWordArt = ActiveDocument.Pages(1).Shapes.AddTextEffect _ (PresetTextEffect:=msoTextEffect7, Text:="Annual Report", _ FontName:="Arial Black", FontSize:=24, _ FontBold:=msoFalse, FontItalic:=msoFalse, _ Left:=144, Top:=72) ``` [!include[Support and feedback](~/includes/feedback-boilerplate.md)]
38.637363
261
0.746303
eng_Latn
0.865378
914ddc6977963a3eacf07066bdc5ac8eda5aabf3
585
markdown
Markdown
website/docs/r/thunder_ip_tcp.html.markdown
a10networks/terraform-provider-thunder
50fe189add4fc51ca17b648945e63685bf350177
[ "BSD-2-Clause" ]
4
2020-10-17T00:07:06.000Z
2021-09-11T21:44:42.000Z
website/docs/r/thunder_ip_tcp.html.markdown
a10networks/terraform-provider-thunder
50fe189add4fc51ca17b648945e63685bf350177
[ "BSD-2-Clause" ]
5
2020-10-09T06:47:26.000Z
2021-09-11T21:44:26.000Z
website/docs/r/thunder_ip_tcp.html.markdown
a10networks/terraform-provider-thunder
50fe189add4fc51ca17b648945e63685bf350177
[ "BSD-2-Clause" ]
3
2020-10-13T06:09:53.000Z
2021-12-03T15:29:08.000Z
--- layout: "thunder" page_title: "thunder: thunder_ip_tcp" sidebar_current: "docs-thunder-resource-ip-tcp" description: |- Provides details about thunder ip tcp resource for A10 --- # thunder\_ip\_tcp `thunder_ip_tcp` Provides details about thunder ip tcp ## Example Usage ```hcl provider "thunder" { address = "192.0.2.65" username = "admin" password = "admin" } resource "thunder_ip_tcp" "Iptcp" { syn_cookie { threshold = 4 } } ``` ## Argument Reference * `uuid` - uuid of the object * `threshold` - SYN cookie expire threshold (seconds (default is 4))
17.205882
68
0.688889
eng_Latn
0.724293
914e4741c8b60c327f55637612ed31a778e66072
2,891
md
Markdown
programador/preferencia-tipos-dominio-especifico.md
jaimerodas/97cosas
c8f2f7967ca53e58d4eb04d73ba89d474f23c5eb
[ "CC-BY-3.0" ]
44
2015-04-02T14:05:21.000Z
2022-02-02T08:34:40.000Z
programador/preferencia-tipos-dominio-especifico.md
jaimerodas/97cosas
c8f2f7967ca53e58d4eb04d73ba89d474f23c5eb
[ "CC-BY-3.0" ]
13
2015-06-17T23:47:28.000Z
2019-10-30T06:23:25.000Z
programador/preferencia-tipos-dominio-especifico.md
esparta/97cosas
f52357df922fea12abe798d9836b5c5121f5732f
[ "CC-BY-3.0" ]
21
2015-04-02T17:49:10.000Z
2021-06-09T00:19:03.000Z
--- layout: programador title: Da preferencia a tipos de Dominio Específico que los tipos primitivos overview: true author: Einar Landre translator: Espartaco Palma original: https://web.archive.org/web/20150106001512/http://programmer.97things.oreilly.com/wiki/index.php/Prefer_Domain-Specific_Types_to_Primitive_Types --- El 23 de septiembre de 1999 el _Mars Climate Orbiter_ de U$327.6 millones se perdió mientras entraba a la órbita alrededor de Marte, debido a un error del software aquí en la Tierra. Error que más tarde fue llamado de “métrica mixta”. El software de la estación en tierra estaba trabajando en libras, mientras que la nave esperaba newtons, llevando a la estación a subestimar el poder de los propulsores de la nave en un factor de 4.45. Éste es uno de los muchos ejemplos de fallas de software que se pudo haber prevenido, si se hubiera aplicado un _tipado_ más fuerte y de dominio específico. Es también un ejemplo del razonamiento detrás de muchas características del lenguaje Ada, uno de sus principales metas de diseño era implementar software de seguridad crítica embebida. Ada estaba fuertemente _tipado_ con revisiones estáticas de ambos: tipos primitivos y tipos definidos por el usuario. type Velocity_In_Knots is new Float range 0.0 .. 500.00; type Distance_In_Nautical_Miles is new Float range 0.0 .. 3000.00; Velocity: Velocity_In_Knots; Distance: Distance_In_Nautical_Miles; Some_Number: Float; Some_Number:= Distance + Velocity; -- Será capturado por el compilador como un error de tipos. Los desarrolladores en dominios menos demandantes también se deberían beneficiar aplicando más _tipado_ de dominio específico, en el que pudieran, de otro modo, continuar usando tipos de datos primitivos ofrecidos por el lenguaje y sus librerías, tales como cadenas y flotantes. En Java, C++, Python y otros lenguajes modernos, los tipos de datos abstractos son conocidos como clases. Usar clases como `Velocity_In_Knots` y `Distance_In_Nautical_Miles` agrega mucho valor con respecto a la calidad del código: * El código se vuelve más legible conforme expresa conceptos de un dominio, no sólo flotantes o cadenas. * El código se vuelve más _testeable_ conforme encapsula su comportamiento, así es fácilmente probado. * El código facilita la reutilización a través de aplicaciones y sistemas. El enfoque es igualmente válido para usuarios de ambos lenguajes de tipo estático y dinámico. La única diferencia es que los desarrolladores que usan lenguajes de _tipado_ estático obtienen más ayuda desde el compilados, mientras aquellos que adoptan lenguajes de _tipado_ dinámico es más común que confíen en sus pruebas unitarias. El estilo de revisión podría ser diferente, pero la motivación y estilo de expresión no. La moraleja es iniciar explorando los tipos de dominio específico con el fin de desarrollar software de calidad.
46.629032
154
0.808025
spa_Latn
0.996094
914e56fe039e6943cc20d26fc4eb02d8d016a707
1,102
md
Markdown
docs/src/SUMMARY.md
roelvdberg/metacontroller
a9ca3730cfba5bfd544f9b65bf42a86f187e478f
[ "Apache-2.0" ]
375
2020-05-21T10:41:15.000Z
2022-03-29T10:23:08.000Z
docs/src/SUMMARY.md
roelvdberg/metacontroller
a9ca3730cfba5bfd544f9b65bf42a86f187e478f
[ "Apache-2.0" ]
391
2020-05-19T09:33:07.000Z
2022-03-31T17:27:18.000Z
docs/src/SUMMARY.md
roelvdberg/metacontroller
a9ca3730cfba5bfd544f9b65bf42a86f187e478f
[ "Apache-2.0" ]
52
2020-05-26T03:03:42.000Z
2022-02-28T03:26:13.000Z
# Summary - [Getting Started](./intro.md) - [Introduction](./intro.md) - [Examples](./examples.md) - [Concepts](./concepts.md) - [Features](./features.md) - [FAQ](./faq.md) - [Pronunciation](./pronunciation.md) - [User Guide](./guide.md) - [Install Metacontroller](./guide/install.md) - [Install Metacontroller via helm](./guide/helm-install.md) - [Configuration](./guide/configuration.md) - [Create a controller](./guide/create.md) - [Constraints and best practices](./guide/best-practices.md) - [Troubleshooting](./guide/troubleshooting.md) - [API Reference](./api.md) - [Apply Semantics](./api/apply.md) - [CompositeController](./api/compositecontroller.md) - [ControllerRevision](./api/controllerrevision.md) - [DecoratorController](./api/decoratorcontroller.md) - [Customize Hook](./api/customize.md) - [Hook](./api/hook.md) - [Design Docs](./design.md) - [MapController](./design/map-controller.md) - [Contributing](./contrib.md) - [Building](./contrib/build.md) - [Local development/debug](./contrib/debug.md)
38
65
0.65245
yue_Hant
0.539141
914e890592492434c7211dafa5cee8f98e0e40dd
1,390
md
Markdown
_listings/aws-cloudformation/action-deletestack-get.md
streamdata-gallery-organizations/aws-cloudformation
4f147d52cf41ec5b07c16a312dff8421187d5ed6
[ "CC-BY-3.0" ]
null
null
null
_listings/aws-cloudformation/action-deletestack-get.md
streamdata-gallery-organizations/aws-cloudformation
4f147d52cf41ec5b07c16a312dff8421187d5ed6
[ "CC-BY-3.0" ]
null
null
null
_listings/aws-cloudformation/action-deletestack-get.md
streamdata-gallery-organizations/aws-cloudformation
4f147d52cf41ec5b07c16a312dff8421187d5ed6
[ "CC-BY-3.0" ]
null
null
null
--- swagger: "2.0" info: title: AWS CloudFormation API version: 1.0.0 schemes: - http produces: - application/json consumes: - application/json paths: /?Action=DeleteStack: get: summary: ' Delete Stack ' description: Deletes a specified stack operationId: deleteStack parameters: - in: query name: RetainResources.member.N description: For stacks in the DELETE_FAILED state, a list of resource logical IDs that are associated with the resources you want to retain type: string - in: query name: RoleARN description: The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes to delete the stack type: string - in: query name: StackName description: The name or the unique stack ID that is associated with the stack type: string responses: 200: description: OK tags: - stacks definitions: [] x-collection-name: AWS CloudFormation x-streamrank: polling_total_time_average: 0 polling_size_download_average: 0 streaming_total_time_average: 0 streaming_size_download_average: 0 change_yes: 0 change_no: 0 time_percentage: 0 size_percentage: 0 change_percentage: 0 last_run: "" days_run: 0 minute_run: 0 ---
25.740741
86
0.661871
eng_Latn
0.721251
914e93e1a084f45b091eaea99f5fbd352f8f8de4
26,470
md
Markdown
common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md
eltociear/common-data-model-and-service
b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd
[ "CC-BY-4.0", "MIT" ]
null
null
null
common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md
eltociear/common-data-model-and-service
b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd
[ "CC-BY-4.0", "MIT" ]
null
null
null
common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md
eltociear/common-data-model-and-service
b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd
[ "CC-BY-4.0", "MIT" ]
1
2020-05-28T15:59:53.000Z
2020-05-28T15:59:53.000Z
--- title: BankStatementEntity - Common Data Model | Microsoft Docs description: undefined author: nenad1002 ms.service: common-data-model ms.reviewer: deonhe ms.topic: article ms.date: 5/5/2020 ms.author: nebanfic --- # Bank statement account statement Latest version of the JSON entity definition is available on <a href="https://github.com/Microsoft/CDM/tree/master/schemaDocuments/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.cdm.json" target="_blank">GitHub</a>. ## Traits <details> <summary>Traits for this entity are listed below. </summary> **is.CDM.entityVersion** <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>versionNumber</td><td>"1.0.0"</td><td>string</td><td>semantic version number of the entity</td></tr></table> **is.application.releaseVersion** <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>releaseVersion</td><td>"10.0.13.0"</td><td>string</td><td>semantic version number of the application introducing this entity</td></tr></table> **is.localized.displayedAs** Holds the list of language specific display text for an object. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>localizedDisplayText</td><td><table><tr><th>languageTag</th><th>displayText</th></tr><tr><td>en</td><td>Bank statement account statement</td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of localized text</td></tr></table> </details> ## Attributes |Name|Description|First Included in Instance| |---|---|---| |[AccountCurrency](#AccountCurrency)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[AccountIdOtherId](#AccountIdOtherId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[AsOfDateModifier](#AsOfDateModifier)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[BankStatementType](#BankStatementType)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[FromDateTime](#FromDateTime)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[GroupControlTotal](#GroupControlTotal)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[GroupStatus](#GroupStatus)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[IBAN](#IBAN)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[Identification](#Identification)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[SwiftNo](#SwiftNo)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[ToDateTime](#ToDateTime)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[TotalAccountsInGroup](#TotalAccountsInGroup)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[TotalRecordsInGroup](#TotalRecordsInGroup)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[TotalEntriesNumberOfEntries](#TotalEntriesNumberOfEntries)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[OriginatorName](#OriginatorName)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[OriginatorId](#OriginatorId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[UltimateReceiverName](#UltimateReceiverName)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[UltimateReceiverId](#UltimateReceiverId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[OriginatorParty](#OriginatorParty)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[UltimateReceiverParty](#UltimateReceiverParty)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[StatementLineNum](#StatementLineNum)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[DocumentLineNum](#DocumentLineNum)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[BankStmtISODocument](#BankStmtISODocument)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[BankStatementRecId](#BankStatementRecId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[BankAccountTable](#BankAccountTable)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[SequenceNumber](#SequenceNumber)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[Relationship_BankStatementDocumentEntityRelationshipId](#Relationship_BankStatementDocumentEntityRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[BackingTable_BankStmtISOAccountStatementRelationshipId](#BackingTable_BankStmtISOAccountStatementRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| |[Relationship_PrimaryCompanyContextRelationshipId](#Relationship_PrimaryCompanyContextRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>| ### <a href=#AccountCurrency name="AccountCurrency">AccountCurrency</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the AccountCurrency attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#AccountIdOtherId name="AccountIdOtherId">AccountIdOtherId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the AccountIdOtherId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#AsOfDateModifier name="AsOfDateModifier">AsOfDateModifier</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the AsOfDateModifier attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#BankStatementType name="BankStatementType">BankStatementType</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the BankStatementType attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#FromDateTime name="FromDateTime">FromDateTime</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the FromDateTime attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#GroupControlTotal name="GroupControlTotal">GroupControlTotal</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the GroupControlTotal attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#GroupStatus name="GroupStatus">GroupStatus</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the GroupStatus attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#IBAN name="IBAN">IBAN</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the IBAN attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#Identification name="Identification">Identification</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the Identification attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#SwiftNo name="SwiftNo">SwiftNo</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the SwiftNo attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#ToDateTime name="ToDateTime">ToDateTime</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the ToDateTime attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#TotalAccountsInGroup name="TotalAccountsInGroup">TotalAccountsInGroup</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the TotalAccountsInGroup attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#TotalRecordsInGroup name="TotalRecordsInGroup">TotalRecordsInGroup</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the TotalRecordsInGroup attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#TotalEntriesNumberOfEntries name="TotalEntriesNumberOfEntries">TotalEntriesNumberOfEntries</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the TotalEntriesNumberOfEntries attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#OriginatorName name="OriginatorName">OriginatorName</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the OriginatorName attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#OriginatorId name="OriginatorId">OriginatorId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the OriginatorId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#UltimateReceiverName name="UltimateReceiverName">UltimateReceiverName</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the UltimateReceiverName attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#UltimateReceiverId name="UltimateReceiverId">UltimateReceiverId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the UltimateReceiverId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#OriginatorParty name="OriginatorParty">OriginatorParty</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the OriginatorParty attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#UltimateReceiverParty name="UltimateReceiverParty">UltimateReceiverParty</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the UltimateReceiverParty attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#StatementLineNum name="StatementLineNum">StatementLineNum</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the StatementLineNum attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#DocumentLineNum name="DocumentLineNum">DocumentLineNum</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the DocumentLineNum attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#BankStmtISODocument name="BankStmtISODocument">BankStmtISODocument</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the BankStmtISODocument attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#BankStatementRecId name="BankStatementRecId">BankStatementRecId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the BankStatementRecId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#BankAccountTable name="BankAccountTable">BankAccountTable</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the BankAccountTable attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#SequenceNumber name="SequenceNumber">SequenceNumber</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table> #### Traits <details> <summary>List of traits for the SequenceNumber attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.nullable** The attribute value may be set to NULL. **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#Relationship_BankStatementDocumentEntityRelationshipId name="Relationship_BankStatementDocumentEntityRelationshipId">Relationship_BankStatementDocumentEntityRelationshipId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table> #### Traits <details> <summary>List of traits for the Relationship_BankStatementDocumentEntityRelationshipId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.dataFormat.guid** **means.identity.entityId** **is.linkedEntity.identifier** Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td>empty table</td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table> **is.dataFormat.guid** **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#BackingTable_BankStmtISOAccountStatementRelationshipId name="BackingTable_BankStmtISOAccountStatementRelationshipId">BackingTable_BankStmtISOAccountStatementRelationshipId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table> #### Traits <details> <summary>List of traits for the BackingTable_BankStmtISOAccountStatementRelationshipId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.dataFormat.guid** **means.identity.entityId** **is.linkedEntity.identifier** Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td><table><tr><th>entityReference</th><th>attributeReference</th></tr><tr><td><a href="../../../Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.md" target="_blank">/core/operationsCommon/Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.cdm.json/BankStmtISOAccountStatement</a></td><td><a href="../../../Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.md#RecId" target="_blank">RecId</a></td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table> **is.dataFormat.guid** **is.dataFormat.character** **is.dataFormat.array** </details> ### <a href=#Relationship_PrimaryCompanyContextRelationshipId name="Relationship_PrimaryCompanyContextRelationshipId">Relationship_PrimaryCompanyContextRelationshipId</a> First included in: CashAndBankManagement/BankStatementEntity (this entity) #### Properties <table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table> #### Traits <details> <summary>List of traits for the Relationship_PrimaryCompanyContextRelationshipId attribute are listed below.</summary> **is.dataFormat.character** **is.dataFormat.big** **is.dataFormat.array** **is.dataFormat.guid** **means.identity.entityId** **is.linkedEntity.identifier** Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td><table><tr><th>entityReference</th><th>attributeReference</th></tr><tr><td><a href="../../../Tables/Finance/Ledger/Main/CompanyInfo.md" target="_blank">/core/operationsCommon/Tables/Finance/Ledger/Main/CompanyInfo.cdm.json/CompanyInfo</a></td><td><a href="../../../Tables/Finance/Ledger/Main/CompanyInfo.md#RecId" target="_blank">RecId</a></td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table> **is.dataFormat.guid** **is.dataFormat.character** **is.dataFormat.array** </details>
35.625841
854
0.730941
eng_Latn
0.316566
914e98746d91e5398269f530d302a428f65ab202
6,046
md
Markdown
src/pages/posts/Life/kvwpt3k0-03700.md
plutoniumm/frontier-next
9487714b636a4f8c00feec15b9fd590cc11d1fe0
[ "MIT" ]
null
null
null
src/pages/posts/Life/kvwpt3k0-03700.md
plutoniumm/frontier-next
9487714b636a4f8c00feec15b9fd590cc11d1fe0
[ "MIT" ]
5
2021-10-12T09:58:43.000Z
2022-02-27T06:43:42.000Z
src/pages/posts/Life/kvwpt3k0-03700.md
plutoniumm/frontier-next
9487714b636a4f8c00feec15b9fd590cc11d1fe0
[ "MIT" ]
null
null
null
--- layout: 'layout:post' --- # Immature Writers: Star Trek Modern movies keep getting more and more shite because they are written and directed by children for children. Or more like people with the emotional maturity and the attention span of children. Consider Star Trek Wrath of Khan (WoK), the undisputed champion of the best Star Trek ever with every metric besides CGI. The main plot is around a 50 year old Kirk who is given back command of the Enterprise to stop Khan who wants chaos everywhere and revenge on Kirk for leaving him stranded on a deserted planet many years ago. The movie starts with a Kirk on his 50th birthday slowly coming to terms with his age, eventually he is even gifted a pair of reading glasses by a member of the crew signifying that his eyesight isn't what is used to be. A VERY VERY RELATABLE BRILLIANT PREMISE, people feel like they are not getting the maximum of their potential and are bored and looking for adventure and that younger and younger people are coming in to replace them everyday. The new Star Trek Into Darkness portrays him as a 36 year old in the prime of his life literally saving the universe and doing his best work ever. But since they need to project him as a reluctant hero they rip off the 50 year old Kirk's struggles and project it onto a 36 year old one. TF!?. In the end of WoK as Spock is dying the old Kirk doesn't start screaming or react uncontrollably but instead just slowly sinks to the grown as he slowly simultaneously copes with the death of a man in action and also the loss of a friend at an old age. Even while delivering the Eulogy as expected of a man leading a starship and a crew he delivers it with stoic composure wavering once slightly at the end. He is expectly to lead by example and does so immaculately. In the new star trek. IT IS A SHITSHOW of characters running around, screaming all the time and discussing relationships in front of superiour officers while responding to high stress situations with emotional outbursts. Even in the end the audience has little time to process what happened as they are forced to come to terms with destruction of a city, a starship, the death of khan and Kirk all within a span of less than 10 minutes. Whatever the fuck just happened? not to mention how in another 5 mins half of this shit will be instantly undone via some magical Trius Ex Machina. ### Interpersonal Conflict At a brief moment during the collapse of the Klingon empire the federation is arguing of what to do. Spock is on one side where he wants to negociate a peace treaty & bring them into the federation and Kirk is on one side who wants to once and for all eliminate the archenemy who for years and years has terrorised the federation. No one side is shown as right or wrong, just two different perspectives are offered. They both argue their case rationally and slowly at a distance from each other across the table. Even briefly as Kirk's bigoted old nature comes to face with his anger he quickly manages to control it and face with reason the situation. Kirk is also demonstrating a difficulting in coming to terms with the scenario given his old mindset which is now slowly aging and becoming outdated. Despite this neither of the two people talk over each other or scream and disparage the other but instead while leaving unresolved, deal with it peacefully. How did the new Star Trek deal with this? Spock and Kirk beat each other up. ### Authority In TNG when Warf openly questions a decision taken by Data when they were left in charge in abscence of the Captain & his First Officer (FO). Data calmly calls Warf into a seperate room where they reason it out quitely in a 2 by 2 exchange. Data asked why he was questioned, Warf denfends that its his job to offer alternatives, data finally explains that after a decision is taken irrespective of his personal feelings the order is to be carried out. In the end Warf recognizes his error and asks data to calmly overlook the incident so that they can continue to be friends. HOLY SHIT THAT WAS IMPECCABLE DIALOGUE. In Star Trek Discovery in stark contrast. The heirarchy and discipline between the people is worse than a liberal arts college. People say anything to anyone, cuss openly, scream shout and denigerate random people. No one from the Admiral to a Science officer has any sense of heirarchy or conflic resolution skill. They are able to do all this and rhetorics and sarcasm with no repurcussions whatsoever. They are all in Immature, Emotional, Generally Incompetent, Impulsive, Hyperactive, Emotionally Unstable and Unprofessional. These teenagers should not be given a small house so much the whole universe to save under their 'responsibility'. ## Hypothesis This is all most likely happening for money. Movies are expensive to make now and if your film doesn't rake in a zillion dollars then you might as well.start learning to code. You need to appeal to as wide an audience as possible particularly the younger demographic, which means lots of action, lots of energy, lots of jokes with dialogue and fast-paced storylines that don't demand too much brain power: no time to waste people, go-go-go. The second reason is this weird trend towards infantilizing modern audiences, carefully shielding them from anything that could be considered difficult scary or threatening: whether it's public information, videos, commercials or government announcements; everything's presented in this weird childish happy-clappy format that looks like the sort of inoffensive crap you'd show to kindergarteners. The final and most influencial reason is writing. Any character is only as smart, capable and resourceful as their writer is. It is no surprise then that today's writers who consider mean tweets to be equivalent to mass murder aren't beacons of tough, stoic, self reliant characters. TNG was written by WW survivors, Late Star Trek post Wok was written by Cold War generation. The new high paced crap is written by Facebook and Snapchat kids. No wonder its shite.
195.032258
959
0.803837
eng_Latn
0.999952
914ea44796c40786c133873559907899eea8b3a3
6,300
md
Markdown
bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md
frederic-junier/DIU
6e3b23d06b69c992e03552f4626676e165f772e9
[ "CC0-1.0" ]
null
null
null
bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md
frederic-junier/DIU
6e3b23d06b69c992e03552f4626676e165f772e9
[ "CC0-1.0" ]
null
null
null
bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md
frederic-junier/DIU
6e3b23d06b69c992e03552f4626676e165f772e9
[ "CC0-1.0" ]
null
null
null
DIU bloc 4 : "Bases de données : création de schémas et normalisation" : TP sur les algorithmes de jointure ==================================================== Dans ce TP, on va s'intéresser **aux algorithmes de jointures**, c'est-à-dire aux algorithmes exécutés par les moteurs des SGBDs quand ils traduisent des requêtes comme la suivante : ```sql SELECT * FROM table1 JOIN table2 ON table1.attr1 == table2.attr2 ``` Il existe plusieurs algorithmes de jointure et l'optimiseur de requêtes du SGBD va tâcher de choisir le _meilleur_, vis-à-vis de statistiques sur les données et surtout des **index** disponibles sur les tables. Le but du TP est ainsi de comprendre ces algorithmes fondamentaux et de les comparer entre eux puis de les comparer face à deux de SQLite 3. **Remarque** la comparaison de performance (_benchmark_) est un exercice complexe car de nombreux paramètres très différents contribuent à la performance finale (matériel, OS, I/O disques ou d'affichage, efficacité de la compilation/interprétation du langage de programmation, caches, temps d'initialisation etc.). Implanter les algorithmes classiques de jointure en Python ----------------------------------------------------------- Le fichier [`join_algorithms.py`](join_algorithms.py) contient le squelette à remplir pour les trois algorithmes, à savoir _nested loop_, _hash join_ et _merge join_. Ces algorithmes font la même chose et ont la même signature `def algo(table1, attr1, table2, attr2):` : * `table1` et `table2` sont des listes (Python) de tuples (Python). Il n'y a pas de garanties d'ordre sur ces listes; * `attr1` (resp. `attr2`) est _l'indice_ (entier) de l'attribut de `table1` (resp. de `tablee`) sur lequel on fait la jointure; * ces algorithmes retournent tous une liste de tuples, comme l'aurait fait la requête SQL. Le fichier [`join_algorithms_test.py`](join_algorithms_test.py) donne un exemple d'entrées et de résultats attendus. **EXERCICE** : compléter la fonction `join_nested_loop` et tester votre implantation avec `pytest-3` et les tests fournis. **EXERCICE** : compléter la fonction `join_hash` et tester votre implantation avec `pytest-3` et les tests fournis. **EXERCICE (POUR ALLER PLUS LOIN)** : compléter la fonction `join_merge` et tester votre implantations avec `pytest-3` et les tests fournis. Vous n'êtes pas obligé de faire cet exercice pour passer à la suite. Comparer algorithmes implantés en Python ---------------------------------------- On peut maintenant comparer la performance des algorithmes avec la fonction fournie `benchmark`.Pour la fonction `join_merge` on compte séparément le temps pris pour le tris des tables. En effet, cette étape peut-être _amortie_ car elle est utile pour d'autre opérations que la jointure, comme les clauses `ORDER BY` ou `GROUP BY`. **EXERCICE** : comprendre ce que fait la fonction `benchmark` (vous pouvez ajouter des commentaires par exempl)e avant de l'exécuter. Avec les paramètres par défaut de `benchmark`, on obtient les résultats suivants sur une machine portable (Dual Core Intel i7-5600U CPU @ 2.60GHz, 8GB RAM). ``` Temps pour une exécution de join_nested_loop : 47.21451419973164 ms Temps pour une exécution de join_hash : 0.8530486993549857 ms Temps pour une exécution des tris : 0.3733556004590355 ms Temps pour une exécution de join_merge : 0.49316930017084815 ms ``` **EXERCICE (POUR ALLER PLUS LOIN)** : jouer avec les paramètres pour trouver un cas qui soit défavorable à `join_hash` mais favorable à `join_merge`. Sans tenir compte du temps de tri, on peut trouver des cas avec un facteur 10x en faveur de `join_merge`. _Indice_ : remarquez que les rôles de `table1` et `table2` sont asymétriques faire en sorte de passer du temps dans l'étape de construction d'index de `join_hash`. **EXERCICE (POUR ALLER PLUS LOIN)** : même question que précédement, mais cette fois si il faut trouver un cas qui est favorable à `join_nested_loop` et dévaforable aux deux autres. _Indice_ faites en sorte que la jointure soit aussi grosse que le produit cartésien. Comparer l'exécution dans Python à celle native dans SQLite ------------------------------------------------------------- Maintenant, on va comparer la performance de ces implantations Python face aux algorithmes jointures de SQLite (qui est écrit en C). Pour cela on va comparer les deux approches suivantes : * **Approche A : jointure en SQLite**, on exécute la requête `SELECT * FROM table1 JOIN table2 ON table1.val == table2.val` puis (depuis Python) on récupère l'intégralité du résultat, c'est la fonction `join_sqlite()` * **Approche B : jointure en Python**, on exécute la requête `SELECT * FROM table1` et on stocke son résultat dans un tableau, de même pour `SELECT * FROM table2` puis on utilise un des algorithmes précedents pour faire le calcul de jointure et enfin on renvoie le résultat, c'est la fonction `join_python()` **EXERCICE** : créer une nouvelle base de données nommée `join_algorithms_versus_sqlite3.db` et exécuter le script SQL `join_algorithms_schema.sql` pour créer le schéma *et* peupler la base avec un jeu de données similaire à celui du benchmark de l'exercice précédent. **EXERCICE** : avec la fonction `join_algorithms_versus_sqlite3()` du programme [`join_algorithms_versus_sqlite3.py`](join_algorithms_versus_sqlite3.py) comparer les temps d'exécution des deux méthodes. Sur ma machine, j'obtiens cet ordre de grandeur : ``` INFO:root:Temps de transfert et de jointure côté Python : 36.440872ms INFO:root:Temps de transfert et de jointure côté Sqlite3 : 108.951388ms ``` **EXERCICE** : reprendre la comparaison mais cette fois avec la requête `SELECT COUNT(*) FROM table1 JOIN table2 ON table1.val == table2.val`. Ici, `join_python()` renverra _la longueur du tableau_ avec `len(join_hash(table1, 1, table2, 0))` pour l'algorithme de jointure par hash. Une différence _importante doit les séparer_ : comment l'expliquer ? Sur ma machine, j'obtiens cet ordre de grandeur : ``` INFO:root:Temps de transfert et de jointure côté Python : 36.784499ms INFO:root:Temps de transfert et de jointure côté Sqlite3 : 6.756878ms ``` **EXERCICE (FACULTATIF ET OUVERT)** : conclure en formulant quelques bonnes pratiques de l'accès à une base de données via un programme (Python).
73.255814
419
0.748095
fra_Latn
0.976679
914ea952efcb9768c90ca31c5a632f425c666142
40
md
Markdown
README.md
issues2markdown/branding
d6f401fe407b0d7f19f59abc5dcc5fa29052bca2
[ "Apache-2.0" ]
1
2018-05-06T00:54:39.000Z
2018-05-06T00:54:39.000Z
README.md
issues2markdown/branding
d6f401fe407b0d7f19f59abc5dcc5fa29052bca2
[ "Apache-2.0" ]
null
null
null
README.md
issues2markdown/branding
d6f401fe407b0d7f19f59abc5dcc5fa29052bca2
[ "Apache-2.0" ]
null
null
null
# branding Branding for issues2markdown
13.333333
28
0.85
eng_Latn
0.989059
914ee5b146839b893d40df9bdad2e82c91dd6d01
7,930
md
Markdown
README.md
libexus/processing-vscode
828325ae0b922d0a881b7fa62b476778e6a06b76
[ "MIT" ]
176
2016-01-31T21:10:43.000Z
2022-03-20T10:11:51.000Z
README.md
libexus/processing-vscode
828325ae0b922d0a881b7fa62b476778e6a06b76
[ "MIT" ]
95
2017-01-08T11:22:45.000Z
2022-03-01T06:06:01.000Z
README.md
libexus/processing-vscode
828325ae0b922d0a881b7fa62b476778e6a06b76
[ "MIT" ]
42
2017-02-11T08:10:27.000Z
2021-12-18T09:38:13.000Z
# Processing for Visual Studio Code [![.github/workflows/main.yml](https://github.com/TobiahZ/processing-vscode/workflows/.github/workflows/main.yml/badge.svg)](https://github.com/TobiahZ/processing-vscode/actions) [![Marketplace Version](https://vsmarketplacebadge.apphb.com/version/Tobiah.language-pde.svg)](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde) [![Installs](https://vsmarketplacebadge.apphb.com/installs/Tobiah.language-pde.svg)](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde) [![Rating](https://vsmarketplacebadge.apphb.com/rating/Tobiah.language-pde.svg)](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde) ## What this extension is This is a Visual Studio Code extension created by Tobiah Zarlez to add Processing language support. ## What this extension isn't This extension does not allow you to debug Java or Processing projects. ### Can you add a feature I want? Possibly! [Let us know](https://github.com/TobiahZ/processing-vscode/issues), we'd love to hear your suggestions. ## Installation Instructions 1. Open [Visual Studio Code](https://code.visualstudio.com/) 1. Open the Command Pallet (`CTRL+SHIFT+P` for Windows/Linux or `CMD+SHIFT+P` on Mac) enter the command “Install Extension” 1. Search for “Processing Language” and click on this extension. 1. Restart Visual Studio Code ## Feature list ### Syntax highlighting Open any .pde file, or choose "Processing" from the drop down menu in the bottom right corner. ### Snippets Once the language has been set, you will see code snippets pop up automatically as you type! ### Commands Installing this extension will add the following commands to your command pallette (`CTRL+SHIFT+P`, or opened by `View -> Command Pallette`). These commands can be selected and run from there, to complete the corresponding tasks. ## Command: Create Task File Adds a `.vscode/tasks.json` file to your project folder, that has the contents of the `ProcessingTasks.json` located in the root folder of this project. When you run this task (Keyboard shortcut: `Ctrl+Shift+B`), it will compile and run your project! If you would like to see output from the compiler, comment out the line `"showOutput": "never",` **NOTE:** Processing must be added to your path, or you must set the "processing.path" setting! Follow [these instructions](#add-processing-to-path) to add Processing to your path, or these [alternate instructions](#alternate-method) instead to modify the path setting. See "[Requirements](#requirements)" for full details. ## Command: Run Processing Project This is a shortcut for running the `.vscode/tasks.json` file. Same as pressing `Ctrl+Shift+B` **Note: Must have ran the "Create Processing Task File" command first, [see above](#command-create-task-file)!** ## Command: Open Extension Documentation Opens this documentation. By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting. ## Command: Open Documentation for Selection Use the pallet command "Processing: Open Documentation for Selection" to open the processing documentation for the current selection. By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting. ## Command: Search Processing Website Use the pallet command "Processing: Search Processing Website" to quickly search whatever you want on the processing website. By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting. By default uses Google for search. Can change to DuckDuckGo if preferred using the `processing.search` setting. ## Requirements Installing the extension will give you instant access to [syntax highlighting](#syntax-highlighting) and [snippets](#snippets). However, in order to compile and run your processing project from Visual Studio Code, you will need to do three things: 1. Set up your `.vscode/tasks.json` file. (See: "[Command: Create Task File](#command-create-task-file)") 1. Add Processing to your path **OR** Modify your `.vscode/tasks.json` file. (See: "[Add Processing to path](#add-processing-to-path)" or "[alternate method](#alternate-method)") 1. Have a `<File>.pde` whose filename matches the name of the project's folder (General Processing Requirement). Your file cannot contain any spaces or it will not run correctly. ## Add Processing to path In order to automatically compile and open from Visual Studio Code, I recommend adding Processing to your path. ### What does that mean? That means you should be able to type the `processing` from anywhere on your machine, and it will open Processing. ### How do I do that? It's easier than you might think! #### Windows * Open the "Advanced System Settings" by running sysdm.cpl * In the "System Properties" window, click on the Advanced tab. * In the "Advanced" section, click the Environment Variables button. * Edit the "Path" variable. Append the processing path (Example: `;C:\Program Files\Processing-3.0.1\`) to the variable value. Each entry is separated with a semicolon. #### Mac Open Processing, and click the `Tools` -> `Install "processing-java"` menu item. **Note:** You will have to install processing-java for all users for this to work #### Linux Set your `PATH` to where your processing application is located. Example: `export PATH=$PATH:/opt/processing/processing-2.0b4` You also need to create an alias for `processing-java` in `/bin/` instead of `/usr/bin/`. Example: `sudo ln -s /opt/processing/processing-java /bin/processing-java` ### Then what? Once you've installed Processing to your path, you need to add the appropriate `.vscode/tasks.json` file to every Processing project. See the command "[Create Task File](#command-create-task-file)" ### Alternate Method What if you cannot, or do not want to add Processing to your path? You can modify the `processing.path` setting to follow the path to wherever processing is installed on your machine. Be sure to remember to keep the `processing-java` at the end of the path! To change settings in VSCode, here is a link to the [official documentation](https://code.visualstudio.com/docs/getstarted/settings). (Remember, for Windows be sure to turn any "`\`" into "`\\`"!) Example: ```json "processing.path": "C:\\Program Files\\processing-3.0.1\\processing-java", ``` **NOTE:** This is untested on Mac and Linux ### If needed: Overwrite default terminal You may need to also overwrite your default terminal in order to get your task file to run correctly. Following [the instructions on the official VSCode documentation](https://code.visualstudio.com/docs/editor/tasks#_common-questions), all you have to do is add a few extra lines to your task file once you generate it. For example, if you are running Windows and want the task file to use Command Prompt ('cmd.exe') you can add an 'options' parameter under the 'windows' portion: ```json "windows": { "options": { "shell": { "executable": "cmd.exe", "args": [ "/d", "/c" ] } }, "args": [ "--force", { "value": "--sketch=${workspaceRoot}", "quoting": "strong" }, { "value": "--output=${workspaceRoot}\\out", "quoting": "strong" }, "--run" ] } ``` ## To Do List * Take nice looking (Animated?) screen shots for README/Instructions ## Credits Syntax highlighting and snippets code based on the [Processing Sublime Text plugin](https://github.com/b-g/processing-sublime). ## Other resources Here are some other resources I recommend: * [Processing's official site](https://processing.org/) * [Tobiah Zarlez Blog](http://www.TobiahZ.com)
40.050505
229
0.731526
eng_Latn
0.953566
914f5a560aa7c944bf528cbb6b77b933b3f50731
9,848
md
Markdown
content/en/docs/quarks-statefulset/development/_index.md
cloudfoundry-incubator/quarks-docs
2bc24a58e81b3a4c333e786ce79f75901dfbdbb2
[ "Apache-2.0" ]
1
2021-03-03T13:33:16.000Z
2021-03-03T13:33:16.000Z
content/en/docs/quarks-statefulset/development/_index.md
cloudfoundry-incubator/quarks-docs
2bc24a58e81b3a4c333e786ce79f75901dfbdbb2
[ "Apache-2.0" ]
4
2020-07-13T05:33:20.000Z
2020-08-28T20:53:18.000Z
content/en/docs/quarks-statefulset/development/_index.md
cloudfoundry-incubator/quarks-docs
2bc24a58e81b3a4c333e786ce79f75901dfbdbb2
[ "Apache-2.0" ]
1
2020-09-08T17:43:20.000Z
2020-09-08T17:43:20.000Z
--- title: "Development" linkTitle: "Development" weight: 4 description: > The QuarksStatefulSet component can be understood as the set of controllers responsible for translating the BOSH manifest `instance_groups` into Kubernetes resources. --- ## Description The QuarksStatefulSet component can be understood as the set of controllers responsible for translating the BOSH manifest `instance_groups` into Kubernetes resources. ## QuarksStatefulset Component The **QuarksStatefulset** component is a categorization of a set of controllers, under the same group. Inside the **QuarksStatefulset** component, we have a set of 2 controllers together with one separate reconciliation loop per controller. Figure 1 illustrates a **QuarksStatefulset** component diagram that covers the set of controllers it uses. ![qsts-component-flow](quarks_estscomponent_flow.png) *Fig. 1: The QuarksStatefulset component* ### **_QuarksStatefulSet Controller_** ![qsts-controller-flow](quarks_estscontroller_flow.png) *Fig. 2: The QuarksStatefulset controller* This controller will generate a Kubernetes statefulset for each `instance_group` defined in the BOSH manifest. This `Statefulset` will also include a set of Kubernetes services, so that each component can be accessed on specific ports. #### Watches in sts controller - `QuarksStatefulset`: Creation - `Configmaps`: Update - `Secrets`: Update #### Reconciliation in sts controller Will generate versioned `Statefulsets` with the required data to make all jobs of the `instance_group` runnable. #### Scaling Restrictions (not implemented) Ability to set restrictions on how scaling can occur: min, max, odd replicas. #### Automatic Restart of Containers When an env value or mount changes due to a `ConfigMap` or `Secret` change, containers are restarted. The operator watches all the `ConfigMaps` and `Secrets` referenced by the `StatefulSet`, and automatically performs the update, without extra workarounds. #### Exposing QuarksStatefulSets Publicly Exposing `quarksstatefulsets` is similar to exposing `statefulsets` in kubernetes. A Kubernetes service makes use of labels to select the pods which should be in the service. We need to use two labels to group the pods of a single instance group. 1. `quarks.cloudfoundry.org/instance-group-name: ((instanceGroupName))` 2. `quarks.cloudfoundry.org/deployment-name: ((deploymentName))` #### Cluster IP Following is the example which creates a service with type **ClusterIP** for a single instance group named `nats` in deployment `nats-deployment` for exposing port 4222. ```yaml apiVersion: v1 kind: Service metadata: name: nats-service spec: type: ClusterIP selector: quarks.cloudfoundry.org/instance-group-name: nats quarks.cloudfoundry.org/deployment-name: nats-deployment ports: - protocol: TCP port: 80 targetPort: 4222 ``` Complete example can be found [here](https://github.com/cloudfoundry-incubator/quarks-operator/tree/master/docs/examples/bosh-deployment/boshdeployment-with-service.yaml). Though, by default, quarks creates three services of type **ClusterIP** as defined [here](../../quarks-operator/concepts/from_bosh_to_kube/#naming-conventions) for any instance group. #### Load Balancer For creating a service type **LoadBalancer**, we just need to change the `.spec.type` to `LoadBalancer` in the above example. The LoadBalancer Ingress is your public IP specified in the output of this command `kubectl describe service nats-service`. #### Ingress **Ingress** doesn't use any labels but just sits on top of services and acts as a smart router. You can create services of different types based on the above examples and use them as values in the ingress Kubernetes spec. An example of Ingress can be found [here](https://kubernetes.io/docs/tasks/access-application-cluster/ingress-minikube/#create-an-ingress-resource) For more information about Kubernetes services, we recommend you to read [this](https://kubernetes.io/docs/concepts/services-networking/service/). #### Extended Upgrade Support When an update needs to happen, a second `StatefulSet` for the new version is deployed, and both coexist until canary conditions are met. Annotated with a version (auto-incremented on each update). The annotation key is `quarks.cloudfoundry.org/version`. Ability to upgrade even though `StatefulSet` pods are not ready. #### Detects if StatefulSet versions are running During upgrades, there is more than one `StatefulSet` version for an `QuarksStatefulSet` resource. The operator lists available versions and keeps track of which are running. A running version means that at least one pod that belongs to a `StatefulSet` is running. When a version **n** is running, any version lower than **n** is deleted. The controller continues to reconcile until there's only one version. #### AZ Support The `zones` key defines the availability zones the `QuarksStatefulSet` needs to span. The `zoneNodeLabel` defines the node label that defines a node's zone. The default value for `zoneNodeLabel` is `failure-domain.beta.kubernetes.io/zone`. The example below defines an `QuarksStatefulSet` that should be deployed in two availability zones, **us-central1-a** and **us-central1-b**. ```yaml apiVersion: quarks.cloudfoundry.org/v1alpha1 kind: QuarksStatefulSet metadata: name: MyQuarksStatefulSet spec: zoneNodeLabel: "failure-domain.beta.kubernetes.io/zone" zones: ["us-central1-a", "us-central1-b"] ... template: spec: replicas: 2 ... ``` The `QuarksStatefulSet` controller creates one `StatefulSet` version for each availability zone, and adds affinity information to the pods of those `StatefulSets`: ```yaml affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchExpressions: - key: "failure-domain.beta.kubernetes.io/zone" operator: In values: ["us-central1-a"] ``` If zones are set for an `QuarksStatefulSet`, the following occurs: - The name of each created `StatefulSet` is generated as `<quarks statefulset name>-z<index of az>`. ```text myquarksstatefulset-z0 ``` - The `StatefulSet` and its `Pods` are labeled with the following: ```yaml quarks.cloudfoundry.org/az-index: "0" quarks.cloudfoundry.org/az-name: "us-central1-a" ``` - The `StatefulSet` and its `Pods` are annotated with an **ordered** JSON array of all the availability zones: ```yaml quarks.cloudfoundry.org/zones: '["us-central1-a", "us-central1-b"]' ``` - As defined above, each pod is modified to contain affinity rules. - Each container and init container of each pod have the following env vars set: ```shell KUBE_AZ="zone name" BOSH_AZ="zone name" CF_OPERATOR_AZ="zone name" AZ_INDEX="zone index" ``` ##### Tolerations Taints and tolerations is a concept defined in kubernetes to repel pods from nodes [link](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/). Defining tolerations is same as defined in the kubernetes docs. Keep in mind the affinity rules added by the controller when az's are defined. An example is specified in the examples folder. #### Restarting on Config Change `QuarksStatefulSets` can be automatically updated when the environment/mounts have changed due to a referenced `ConfigMap` or a `Secret` being updated. This behavior is controlled by the `updateOnConfigChange` flag which defaults to `false`. #### Watches in cleanup controller - `StatefulSet`: Creation/Update #### Reconciliation in cleanup controller It will delete statefulsets with old versions, only after the new statefulset version instances are up and running. ### **_QuarksStatefulSet Active-Passive Controller_** ![qsts-activepassivecontroller-flow](quarks_estsactivepassivecontroller_flow.png) *Fig. 3: The QuarksStatefulset active/passive controller* Active/passive model is application model that have multiple running instances, but only one instance is active and all other instances are passive (standby). If the active instance is down, one of the passive instances will be promoted to active immediately. The `activePassiveProbes` key defines active probe to be performed on a container. The controller examines the active probe periodically to see if the active one is still active. If active pod is down or there isn’t an active pod, the first running pod will be promoted as active and label it as `quarks.cloudfoundry.org/pod-active: active`. ```yaml apiVersion: quarks.cloudfoundry.org/v1alpha1 kind: QuarksStatefulSet metadata: name: myquarksstatefulset spec: activePassiveProbes: busybox: # define a active probe on the container exec: command: - /bin/sh - -c - /root/check-active.sh template: spec: replicas: 2 template: spec: containers: - name: busybox image: busybox command: - sleep - "3600" ... ``` The controller manages this active probing and provides pod designation label to the service's selectors. Any requests sent to the service will then only be sent to the active pod. ## Relationship with the BDPL component ![bpm-qsts-relationship](quarks_bpm_and_ests_flow.png) *Fig. 4: Relationship with the BPM controller* Figure 4 illustrates the interaction of the **BPM** Controller with the **QuarksStatefulSet** Controller. Once the BPM controller consumes the data persisted in secrets from the `QuarksJob` Component, it will use that data to generate new `QuarksStatefulset` instances. When these resources are generated, the QuarksStatefulSet controller will be watching and trigger its reconciliation loop. ## Examples See https://github.com/cloudfoundry-incubator/quarks-operator/tree/master/docs/examples/quarks-statefulset
41.033333
392
0.762998
eng_Latn
0.985505
914fc01bb8fc851df13112b02e3a565dfaa1ea95
3,993
md
Markdown
event-pipeline-evm/README.md
0xProject/0x-event-pipeline
060e7b04bd73fe3d60d9646b7eb03ddabeb2641b
[ "Apache-2.0" ]
11
2020-03-03T19:32:27.000Z
2022-03-30T03:56:53.000Z
event-pipeline-evm/README.md
0xProject/0x-event-pipeline
060e7b04bd73fe3d60d9646b7eb03ddabeb2641b
[ "Apache-2.0" ]
11
2019-12-03T23:57:38.000Z
2021-09-02T03:16:53.000Z
event-pipeline-evm/README.md
0xProject/0x-event-pipeline
060e7b04bd73fe3d60d9646b7eb03ddabeb2641b
[ "Apache-2.0" ]
10
2020-10-22T22:03:28.000Z
2022-03-30T03:56:55.000Z
# 0x-event-pipeline A node.js app that was originally designed for pulling 0x staking events info, but now expanded to all other 0x related events. ## Getting started Test locally: - Step 1 Rename the `.env.exmaple` file to `.env`, or create a new `.env` file. Add the required env variables (see below for configureation details) - Step 2 Set up the database variables in `docker-compose.yml` file to the desired database destination. ``` $ docker-compose up # get postgres up ``` - Step 3 Test build & Debug ``` $ yarn install $ yarn build ``` - Step 4 Build migrations If there are new tables to be created, or schema changes, you will need to create migration files first: ``` yarn migrate:create -n <YourMigrationName> ``` Modify the migration file in `migrations/` folder with necessary changes. Run migration: ``` $ yarn migrate:run ``` To revert migration: ``` $ yarn migrate:revert ``` - Step 5 Start the scraper: ``` $ yarn start ``` ## Configuration ### Environment variables: **Required** `ETHEREUM_RPC_URL` - The RPC URL to use. Must match `CHAIN_ID`. `CHAIN_ID` - The EVM chain id. `EP_DEPLOYMENT_BLOCK` - The block on which the proxy contract was deployed `SCHEMA` - The schema to use to store events in the DB **Optional** `POSTGRES_URI` - The full postgres URI to connect to. Defaults to local development. `START_BLOCK_OFFSET` - How many blocks before the current block to search for events, allowing for updates to previously scraped events that may be in orphaned blocks. `MAX_BLOCKS_TO_PULL` - The maximum number of blocks to pull at once. `MAX_BLOCKS_TO_SEARCH` - The maximum number of blocks to search for events at once. `BLOCK_FINALITY_THRESHOLD` - How many blocks before the current block to end the search, allowing you to limit your event scrape to blocks that are relatively more settled. `MINUTES_BETWEEN_RUNS` - How long to wait between scrapes. `SHOULD_SYNCHRONIZE` - Whether typeorm should synchronize with the database from `POSTGRES_URI`. `STAKING_DEPLOYMENT_BLOCK` - The block on which the staking contract was deployed `STAKING_POOLS_JSON_URL` - The source for the JSON mapping of staking pools to UUIDs (for grabbing metadata about pools). Defaults to the 0x staking pool registry GitHub repo. `STAKING_POOLS_METADATA_JSON_URL` - The source for the JSON mapping of UUIDs to metadata. Defaults to the 0x staking pool registry GitHub repo. `BASE_GITHUB_LOGO_URL` - The base URL for grabbing logos for staking pools. Defaults to the 0x staking pool registry GitHub repo. ## Database snapshots When running the app on a new database it can take a long time to find new events depending on how much time has passed since the contracts were deployed. There are options to dump and restore data from other sources using `pg_dump` ([Documentation](https://www.postgresql.org/docs/9.6/app-pgdump.html)) and `pg_restore` ([Documentation](https://www.postgresql.org/docs/9.2/app-pgrestore.html)). Some examples are outlined below. These examples will require `postgresql` to be installed. ``` $ brew install postgresql ``` ### Getting data from another database If you know of another database that contains up-to-date data, you can `pg_dump` data from the relevant schemas from that database by running: ``` $ pg_dump -h <host> -U <user> -p <port> --schema staking --schema events --data-only --file events.dump --format=c <database name> ``` To save a `pg_dump` archive file named `events.dump`. The command will prompt you for the password. ### Restoring data from a pg_dump If you have access to a `.dump` file you can `pg_restore` data from that file into another database. To restore data into the default development database that is spun up by `docker-compose up`, you can run: ``` $ pg_restore --data-only --dbname events --host localhost --port 5432 -U user events.dump ``` Assuming you have access to an `events.dump` file. The command will prompt you for the password.
31.944
429
0.749562
eng_Latn
0.98751
9150199e8c4ddf912e25e7c581aec7ab62c76ebb
93
md
Markdown
README.md
djoc3/react-markup
d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9
[ "MIT" ]
null
null
null
README.md
djoc3/react-markup
d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9
[ "MIT" ]
5
2020-07-19T08:25:45.000Z
2022-02-13T04:24:54.000Z
README.md
djoc3/react-markdown
d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9
[ "MIT" ]
null
null
null
Markdown Preview is built with React and Marked.js Developed and maintained by Dan O'Connor
23.25
50
0.817204
eng_Latn
0.999521
91502d1e660651cc068a74cce475bd7f5297be53
981
md
Markdown
README.md
m-pawelczyk/skrzynkomaty-domain-explorers
ef80c9e8d2f0abaab2602a35d12b430a36183ed0
[ "MIT" ]
null
null
null
README.md
m-pawelczyk/skrzynkomaty-domain-explorers
ef80c9e8d2f0abaab2602a35d12b430a36183ed0
[ "MIT" ]
null
null
null
README.md
m-pawelczyk/skrzynkomaty-domain-explorers
ef80c9e8d2f0abaab2602a35d12b430a36183ed0
[ "MIT" ]
null
null
null
# SKRZYNKOMATY #DomainExplorers Oryginalny pomysł na ten projekt wyszedł od [Mariusza Gila](https://github.com/mariuszgil/domain-explorers) i jest próbą odpowiedzenia na jego inicjatywę #DomainExplorers Ciekawą zapowiedzią i punktem wyjścia na początek był film opublikowany na YouTube: [Rozpoznanie domeny z Big Picture EventStorming | Domain Explorers](https://youtu.be/LqIF_pGI3wk) ## O repozytorium Ten projekt stanowi zapis analizy, modelowania oraz implementacji w ramach [Domain Explorers](https://explorers.bettersoftwaredesign.pl). Znajdziesz tutaj zarówno artefakty wynikowe poszczególnych sesji EventStormingu, modelowania, jak i (docelowo) powstałą na tej bazie implementację (miejmy nadzieję :D). ## Analiza domeny ### Big Picture EventStorming Pierwsze rozwiązanie - Mikołaj Nowogrodzki i Michał Pawełczyk: ![Big Picture EventStorming - pierwsze rozwiązanie](docs/big-picture-skrzynkomaty.jpg) Board w MIRO - [LINK](https://miro.com/app/board/o9J_lU7sET0=/)
49.05
306
0.808359
pol_Latn
0.998602