hexsha stringlengths 40 40 | size int64 5 1.04M | ext stringclasses 6 values | lang stringclasses 1 value | max_stars_repo_path stringlengths 3 344 | max_stars_repo_name stringlengths 5 125 | max_stars_repo_head_hexsha stringlengths 40 78 | max_stars_repo_licenses listlengths 1 11 | max_stars_count int64 1 368k ⌀ | max_stars_repo_stars_event_min_datetime stringlengths 24 24 ⌀ | max_stars_repo_stars_event_max_datetime stringlengths 24 24 ⌀ | max_issues_repo_path stringlengths 3 344 | max_issues_repo_name stringlengths 5 125 | max_issues_repo_head_hexsha stringlengths 40 78 | max_issues_repo_licenses listlengths 1 11 | max_issues_count int64 1 116k ⌀ | max_issues_repo_issues_event_min_datetime stringlengths 24 24 ⌀ | max_issues_repo_issues_event_max_datetime stringlengths 24 24 ⌀ | max_forks_repo_path stringlengths 3 344 | max_forks_repo_name stringlengths 5 125 | max_forks_repo_head_hexsha stringlengths 40 78 | max_forks_repo_licenses listlengths 1 11 | max_forks_count int64 1 105k ⌀ | max_forks_repo_forks_event_min_datetime stringlengths 24 24 ⌀ | max_forks_repo_forks_event_max_datetime stringlengths 24 24 ⌀ | content stringlengths 5 1.04M | avg_line_length float64 1.14 851k | max_line_length int64 1 1.03M | alphanum_fraction float64 0 1 | lid stringclasses 191 values | lid_prob float64 0.01 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
912784033410cbea044d74258aefe4246b6b2685 | 886 | md | Markdown | _posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md | tatudoquei/tatudoquei.github.io | a3a3c362424fda626d7d0ce2d9f4bead6580631c | [
"MIT"
] | null | null | null | _posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md | tatudoquei/tatudoquei.github.io | a3a3c362424fda626d7d0ce2d9f4bead6580631c | [
"MIT"
] | null | null | null | _posts/2021-06-07-apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba.md | tatudoquei/tatudoquei.github.io | a3a3c362424fda626d7d0ce2d9f4bead6580631c | [
"MIT"
] | 1 | 2022-01-13T07:57:24.000Z | 2022-01-13T07:57:24.000Z | ---
layout: post
item_id: 3351158114
title: >-
Após 19 dias, Salles entrega celular a investigadores da operação Akuanduba
author: Tatu D'Oquei
date: 2021-06-07 18:45:32
pub_date: 2021-06-07 18:45:32
time_added: 2021-06-07 19:04:57
category:
tags: []
image: https://www.cartacapital.com.br/wp-content/uploads/2021/05/ricardosalles.jpg
---
O ministro do Meio Ambiente, Ricardo Salles, informou ao Supremo Tribunal Federal, nesta segunda-feira 7, que entregou seu celular aos os investigadores da Polícia Federal, no âmbito da Operação Akuanduba, que investiga sua possível atuação em prol de empresas na exportação ilegal de madeir
**Link:** [https://www.cartacapital.com.br/politica/apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba/](https://www.cartacapital.com.br/politica/apos-19-dias-salles-entrega-celular-a-investigadores-da-operacao-akuanduba/)
| 46.631579 | 291 | 0.784424 | por_Latn | 0.82637 |
9127a5ff9bdafd7c26d128524ed18eb560875fba | 673 | md | Markdown | sandstone/pattern-virtualgridlist-api/README.md | enyojs/enact-samples | e360b50eeed552613f8dc4f5b1d18d299ee5a7dd | [
"Apache-2.0"
] | null | null | null | sandstone/pattern-virtualgridlist-api/README.md | enyojs/enact-samples | e360b50eeed552613f8dc4f5b1d18d299ee5a7dd | [
"Apache-2.0"
] | null | null | null | sandstone/pattern-virtualgridlist-api/README.md | enyojs/enact-samples | e360b50eeed552613f8dc4f5b1d18d299ee5a7dd | [
"Apache-2.0"
] | null | null | null | ## VirtualGridList add/remove/select/deselect pattern // My Gallery
A sample Enact application that shows off how to add/remove/select/deselect items of VirtualGridList
Run `npm install` then `npm run serve` to have the app running on [http://localhost:8080](http://localhost:8080), where you can view it in your browser.
#### Enact Components Used
- `sandstone/Button`
- `sandstone/ImageItem`
- `sandstone/Panels/Header`
- `sandstone/VirtualList/VirtualGridList`
In this app, you can add, remove, select, deselect, selectAll, and deselectAll items using redux.
---
This project was bootstrapped with the Enact [cli](https://github.com/enactjs/cli).
| 37.388889 | 153 | 0.744428 | eng_Latn | 0.956692 |
912862e895a04c35a9cd6b8534c8800642e505ad | 2,428 | md | Markdown | README.md | es-shims/Promise.try | 23546f147e25515660c9b20f346be3dca513dd74 | [
"MIT"
] | 5 | 2016-08-21T12:18:28.000Z | 2019-10-27T22:28:44.000Z | README.md | es-shims/Promise.try | 23546f147e25515660c9b20f346be3dca513dd74 | [
"MIT"
] | 2 | 2019-11-11T07:57:12.000Z | 2019-11-19T09:48:45.000Z | README.md | es-shims/Promise.try | 23546f147e25515660c9b20f346be3dca513dd74 | [
"MIT"
] | 2 | 2016-08-21T09:35:10.000Z | 2018-01-17T09:33:54.000Z | # promise.try <sup>[![Version Badge][npm-version-svg]][package-url]</sup>
[![Build Status][travis-svg]][travis-url]
[![dependency status][deps-svg]][deps-url]
[![dev dependency status][dev-deps-svg]][dev-deps-url]
[![License][license-image]][license-url]
[![Downloads][downloads-image]][downloads-url]
[![npm badge][npm-badge-png]][package-url]
[![browser support][testling-svg]][testling-url]
ES Proposal spec-compliant shim for `Promise.try`. Invoke its "shim" method to shim `Promise.try` if it is unavailable or noncompliant. **Note**: a global `Promise` must already exist: the [es6-shim](https://github.com/es-shims/es6-shim) is recommended.
This package implements the [es-shim API](https://github.com/es-shims/api) interface. It works in an ES3-supported environment that has `Promise` available globally, and complies with the [proposed spec](https://github.com/ljharb/proposal-promise-try).
Most common usage:
```js
var assert = require('assert');
var promiseTry = require('promise.try');
promiseTry(function () {
throw 42;
}).catch(function (e) {
assert.equal(e, 42);
});
promiseTry(function () {
return Infinity;
}).then(function (x) {
assert.equal(x, Infinity);
});
promiseTry.shim(); // will be a no-op if not needed
Promise.try(function () {
throw 42;
}).catch(function (e) {
assert.equal(e, 42);
});
Promise.try(function () {
return Infinity;
}).then(function (x) {
assert.equal(x, Infinity);
});
```
## Tests
Simply clone the repo, `npm install`, and run `npm test`
[package-url]: https://npmjs.com/package/promise.try
[npm-version-svg]: http://versionbadg.es/es-shims/Promise.try.svg
[travis-svg]: https://travis-ci.org/es-shims/Promise.try.svg
[travis-url]: https://travis-ci.org/es-shims/Promise.try
[deps-svg]: https://david-dm.org/es-shims/Promise.try.svg
[deps-url]: https://david-dm.org/es-shims/Promise.try
[dev-deps-svg]: https://david-dm.org/es-shims/Promise.try/dev-status.svg
[dev-deps-url]: https://david-dm.org/es-shims/Promise.try#info=devDependencies
[testling-svg]: https://ci.testling.com/es-shims/Promise.try.png
[testling-url]: https://ci.testling.com/es-shims/Promise.try
[npm-badge-png]: https://nodei.co/npm/promise.try.png?downloads=true&stars=true
[license-image]: http://img.shields.io/npm/l/promise.try.svg
[license-url]: LICENSE
[downloads-image]: http://img.shields.io/npm/dm/promise.try.svg
[downloads-url]: http://npm-stat.com/charts.html?package=promise.try
| 36.238806 | 253 | 0.719522 | kor_Hang | 0.266022 |
9128a347536e4cd1119bce7afa140e842578a0cb | 142 | md | Markdown | content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md | technology-toolbox/website | 9d845dc68e650ee164959da418fde24eacecf1c9 | [
"MIT"
] | null | null | null | content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md | technology-toolbox/website | 9d845dc68e650ee164959da418fde24eacecf1c9 | [
"MIT"
] | 109 | 2021-03-25T11:16:17.000Z | 2022-01-23T20:55:51.000Z | content/blog/jjameson/2008/04/08/resources/table-1-popout/_index.md | technology-toolbox/website | 9d845dc68e650ee164959da418fde24eacecf1c9 | [
"MIT"
] | null | null | null | ---
layout: popout
title: Table 1 - MOSS 2007 Feature Definitions
date: 2008-04-08T18:39:00-06:00
---
{{< include-html "../table-1.html" >}}
| 17.75 | 46 | 0.65493 | kor_Hang | 0.46492 |
912951379577e4f7b23fcae87b7916c8adf06b40 | 349 | md | Markdown | README.md | Ccode-lang/pydrumplugin | ab65688288a6f2f464fe1ca69560c975a750f194 | [
"MIT"
] | null | null | null | README.md | Ccode-lang/pydrumplugin | ab65688288a6f2f464fe1ca69560c975a750f194 | [
"MIT"
] | null | null | null | README.md | Ccode-lang/pydrumplugin | ab65688288a6f2f464fe1ca69560c975a750f194 | [
"MIT"
] | null | null | null | # pydrumplugin
A template for a drumbash plugin in python
# setup
1. Change filler in `build.sh` to whatever you want to name the plugin
2. Write code using the imported api
# building
Run:
```bash
./build.sh
```
The outputed drumfile will be in `./artifact`.
# drum api
look at how to use at https://github.com/Ccode-lang/drumbash/wiki/py-api-docs
| 24.928571 | 77 | 0.74212 | eng_Latn | 0.978943 |
912958e5c9c66c83392f1299f852e467080be1d4 | 738 | md | Markdown | _posts/2020-01-21-cartagena-taller3.md | EducacionSiglo21/jekyll | f04b07efe35a5c2c61b290592e3b9b63d2e16552 | [
"MIT"
] | 3 | 2018-11-26T15:11:12.000Z | 2019-02-12T06:51:17.000Z | _posts/2020-01-21-cartagena-taller3.md | EducacionSiglo21/startbootstrap-clean-blog-jekyll | db55bc500681697614b32c530ae29bdf58305b72 | [
"MIT"
] | null | null | null | _posts/2020-01-21-cartagena-taller3.md | EducacionSiglo21/startbootstrap-clean-blog-jekyll | db55bc500681697614b32c530ae29bdf58305b72 | [
"MIT"
] | 3 | 2018-12-16T10:55:27.000Z | 2021-01-22T21:19:25.000Z | ---
layout: post
title: "Realidad virtual y realidad aumentada"
subtitle: "Taller"
background: "/img/posts/bg-cartagena.jpg"
eventdate: 2020-02-20 10:00:00 +0100
category: "local"
tags: "cartagena"
placeName: "CEIP Bethoven, Cartagena"
speakers:
- name: Paqui Rosique
---
Descripción taller: La realidad virtual (VR) se coló en la vida de más de uno desde el mundo de los videojuegos. No
obstante, sus aplicaciones van ya mucho más allá del propio entretenimiento y son algo más que útiles para distintos
sectores en el día a día de cualquier persona o profesional. El taller de Realidad Virtual y Realidad Aumentada per -
mitirá a los participantes sentir la inmersión de la Realidad Virtual a través de las gafas Oculus Ri
| 33.545455 | 117 | 0.768293 | spa_Latn | 0.993285 |
912a636db978d531b25d027896114e19feb331b1 | 227 | md | Markdown | README.md | ShaneMcC/DMDirc-Util | ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b | [
"MIT"
] | null | null | null | README.md | ShaneMcC/DMDirc-Util | ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b | [
"MIT"
] | 6 | 2015-01-17T21:58:27.000Z | 2017-01-15T05:09:08.000Z | README.md | ShaneMcC/DMDirc-Util | ff6906a5ce69f654d6f7b8b716dfe354c1c4ec8b | [
"MIT"
] | 2 | 2019-05-02T22:30:56.000Z | 2019-05-08T05:51:13.000Z | # Util
[](https://www.codacy.com/app/DMDirc/Util?utm_source=github.com&utm_medium=referral&utm_content=DMDirc/Util&utm_campaign=badger) | 113.5 | 220 | 0.823789 | yue_Hant | 0.886751 |
912b0439871eb5c3da9cb4c27038da1001ee31e6 | 204 | md | Markdown | README.md | Roragok/namafia-anime | 6d804163152bd002f198034734b84704356eb199 | [
"MIT"
] | null | null | null | README.md | Roragok/namafia-anime | 6d804163152bd002f198034734b84704356eb199 | [
"MIT"
] | 15 | 2019-12-17T16:49:32.000Z | 2022-02-18T16:48:13.000Z | README.md | Roragok/namafia-anime | 6d804163152bd002f198034734b84704356eb199 | [
"MIT"
] | null | null | null | # namafia-anime
Webpage for anime streamings
To run this yourself install node and yarn, clone the repo and run yarn start.
Useful Links:
- https://www.typescriptlang.org/docs/handbook/basic-types.html
| 25.5 | 78 | 0.789216 | eng_Latn | 0.52797 |
912b1587f74aa7894fca0a1d7e118eab34590f4e | 16,211 | md | Markdown | _episodes/05-loop.md | statkclee/shell-novice-kr | c8c59be87a1c1f4e9310baca5f209eaac6801d15 | [
"CC-BY-4.0"
] | null | null | null | _episodes/05-loop.md | statkclee/shell-novice-kr | c8c59be87a1c1f4e9310baca5f209eaac6801d15 | [
"CC-BY-4.0"
] | null | null | null | _episodes/05-loop.md | statkclee/shell-novice-kr | c8c59be87a1c1f4e9310baca5f209eaac6801d15 | [
"CC-BY-4.0"
] | null | null | null | ---
title: "루프(Loops)"
teaching: 40
exercises: 10
questions:
- "다른 파일이 많은데 어떻게 동일한 동작을 수행시킬 수 있을까?"
objectives:
- "파일 집합의 각 파일에 따로 따로 나누어서 하나 혹은 명령어 다수를 적용하는 루프를 작성한다."
- "루프가 실행되는 동안에 루프 변수가 취하는 값을 추적한다."
- "변수명과 변수값 차이에 대해 설명한다."
- "왜 공백과 일부 구두점 문자는 파일 이름에 사용되지 말아야 되는지 설명한다."
- "어떤 명령어가 최근에 실행되었는지를 확인하는 방법을 시범으로 보여준다."
- "명령어를 다시 타이핑하지 않고 최근에 실행된 명령어를 다시 실행한다."
keypoints:
- "`for` 루프는 리스트의 모든 원소에 대해서 명령어를 한번씩 모두 반복한다."
- "모든 `for` 루프는 변수를 사용해서 현재 연산작업하는 것을 추적한다."
- "`$name`을 사용해서 변수(즉, 변수값을 얻는데)를 확장한다. `${name}`도 사용될 수 있다."
- "파일명에 공백, 인용부호, '*' 혹은 '?'와 같은 와일드카드 문자를 사용하지 않는다. 왜냐햐면, 변수 확장을 난해하게 된다."
- "파일에 일관된 명칭을 부여해서 와일드카드 패턴으로 매칭되기 쉽게 하고 루프를 돌릴 때 선택도 쉽게 만든다."
- "이전 명령어를 편집하고 반복실행하는데 키보드 윗방향 화살표를 사용한다."
- "`Ctrl-R` 명령어를 사용해서 이전에 입력한 명령어를 검색한다."
- "`history` 명령어를 사용해서 가장 최근 명령어를 화면에 출력하고, `!number`을 사용해서 해당 숫자 명령어를 반복실행한다."
---
반복적으로 명령어를 실행하게 함으로써 자동화를 통해서 **루프**는 생산성 향상에 핵심이 된다.
와일드카드와 탭 자동완성과 유사하게, 루프를 사용하면 타이핑 상당량(타이핑 실수)을 줄일 수 있다.
와일드카드와 탭 자동완성은 타이핑을 (타이핑 실수를) 줄이는 두가지 방법이다.
또다른 것은 쉘이 반복해서 특정 작업을 수행하게 하는 것이다.
`basilisk.dat`, `unicorn.dat` 등으로 이름 붙여진 게놈 데이터 파일이 수백개 있다고 가정하자.
이번 예제에서,
단지 두개 예제 파일만 있는 `creatures` 디렉토리를 사용할 것이지만 동일한 원칙은 훨씬 더 많은 파일에 즉시 적용될 수 있다.
디렉토리에 있는 파일을 변경하고 싶지만,
원본 파일을 `original-basilisk.dat`와 `original-unicorn.dat`으로 이름을 변경해서 저장한다.
하지만 다음 명령어를 사용할 수 없다:
~~~
$ cp *.dat original-*.dat
~~~
{: .language-bash}
왜냐하면 상기 두 파일 경우에 전개가 다음과 같이 될 것이기 때문이다:
~~~
$ cp basilisk.dat unicorn.dat original-*.dat
~~~
{: .language-bash}
상기 명령어는 파일을 백업하지 않고 대신에 오류가 발생된다:
~~~
cp: target `original-*.dat' is not a directory
~~~
{: .error}
`cp` 명령어는 입력값 두개 이상을 받을 때 이런 문제가 발생한다.
이런 상황이 발생할 때, 마지막 입력값을 디렉토리로 예상해서 모든 파일을 해당 디렉토리로 넘긴다.
`creatures` 디렉토리에는 `original-*.dat` 라고 이름 붙은 하위 디렉토리가 없기 때문에, 오류가 생긴다.
대신에, 리스트에서 한번에 연산작업을 하나씩 수행하는
**루프(loop)**를 사용할 수 있다.
교대로 각 파일에 대해 첫 3줄을 화면에 출력하는 단순한 예제가 다음에 나와 있다:
~~~
$ for filename in basilisk.dat unicorn.dat
> do
> head -n 3 $filename # 루프 내부에 들여쓰기는 가독성에 도움을 준다.
> done
~~~
{: .language-bash}
~~~
COMMON NAME: basilisk
CLASSIFICATION: basiliscus vulgaris
UPDATED: 1745-05-02
COMMON NAME: unicorn
CLASSIFICATION: equus monoceros
UPDATED: 1738-11-24
~~~
{: .output}
> ## `for` 루프 내부에 코드 들여쓰기
> `for` 루프 내부의 코드를 들여쓰는 것이 일반적인 관행이다.
> 들여쓰는 유일한 목적은 코드를 더 읽기 쉽게 하는 것 밖에 없다 -- `for` 루프를 실행하는데는 꼭 필요하지는 않다.
{: .callout}
쉘이 키워드 `for`를 보게 되면,
쉘은 리스트에 있는 각각에 대해 명령문 하나(혹은 명령문 집합)을 반복할 것이라는 것을 알게 된다.
루프를 반복할 때마다(iteration 이라고도 한다),
현재 작업하고 있는 파일 이름은 `filename`으로 불리는 **변수(variable)**에 할당된다.
리스트의 다음 원소로 넘어가기 전에 루프 내부 명령어가 실행된다.
루프 내부에서, 변수 이름 앞에 `$` 기호를 붙여 변수 값을 얻는다:
`$` 기호는 쉘 해석기가 변수명을 텍스트나 외부 명령어가 아닌 **변수**로 처리해서 값을 해당 위치에 치환하도록 지시한다.
이번 경우에 리스트는 파일이름이 두개다: `basilisk.dat`, `unicorn.dat`.
매번 루프가 돌 때마다 파일명을 `filename` 변수에 할당하고 `head` 명령어를 실행시킨다.
즉, 루프가 첫번째 돌 때 `$filename` 은 `basilisk.dat`이 된다.
쉘 해석기는 `basilisk.dat` 파일에 `head` 명령어를 실행시켜서
`basilisk.dat` 파일의 첫 3줄을 화면에 출력시킨다.
두번째 반복에서, `$filename`은 `unicorn.dat`이 된다.
이번에는 쉘이 `head` 명령어를 `unicorn.dat` 파일에 적용시켜
`unicorn.dat` 파일 첫 3줄을 화면에 출력시킨다.
리스트에 원소가 두개라서, 쉘은 `for` 루프를 빠져나온다.
변수명을 분명히 구분하는데, 중괄호 내부에 변수명을 넣어서 변수로 사용하는 것도 가능하다:
`$filename` 은 `${filename}`와 동치지만, `${file}name`와는 다르다.
이 표기법을 다른 사람 프로그램에서 찾아볼 수 있다.
> ## 루프 내부의 변수
>
> 이번 예제는 `data-shell/molecules` 디렉토리를 가정한다.
> `ls` 명령어를 던지면 출력결과는 다음과 같다:
>
> ~~~
> cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> ~~~
> {: .output}
>
> 다음 코드의 출력결과는 어떻게 나오는가?
>
> ~~~
> $ for datafile in *.pdb
> > do
> > ls *.pdb
> > done
> ~~~
> {: .language-bash}
>
> 이제 다음 코드의 출력결과는 무엇인가?
>
> ~~~
> $ for datafile in *.pdb
> > do
> > ls $datafile
> > done
> ~~~
> {: .language-bash}
>
> 왜 상기 두 루프 실행결과는 다를까?
>
> > ## 해답
> > 첫번째 코드 블록은 루프를 돌릴 때마다 동일한 출력결과를 출력한다.
> > 배쉬는 루프 몸통 내부 와일드카드 `*.pdb`을 확장해서 `.pdb`로 끝나는
> > 모든 파일을 매칭시킨다.
> > 확장된 루프는 다음과 같이 생겼다:
> > ```
> > $ for datafile in cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > > do
> > > ls cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > > done
> > ```
> > {: .language-bash}
> >
> > ```
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > cubane.pdb ethane.pdb methane.pdb octane.pdb pentane.pdb propane.pdb
> > ```
> > {: .output}
> >
> > 두번째 코드 블록은 루프를 돌 때마다 다른 파일을 출력한다.
> > `datafile` 파일 변수값이 `$datafile`을 통해 평가되고
> > `ls` 명령어를 사용해서 파일 목록을 출력하게 된다.
> >
> > ```
> > cubane.pdb
> > ethane.pdb
> > methane.pdb
> > octane.pdb
> > pentane.pdb
> > propane.pdb
> > ```
> > {: .output}
> {: .solution}
{: .challenge}
> ## 프롬프트 따라가기
>
> 루프안에서 타이핑을 할 때, 쉘 프롬프트가 `$`에서 `>`으로 바뀐다.
> 두번째 프롬프트는, `>`, 온전한 명령문 타이핑이 끝마치지 않았음을 상기시키려고 다르게 표기된다.
> 세미콜론 `;` 을 사용해서 두 명령어로 구성된 문장을 단일 명령줄로 단순화한다.
{: .callout}
> ## 동일한 기호, 하지만 다른 의미
>
> 쉘 프롬프트로 `>` 기호가 사용되는 것을 확인했지만,
> `>` 기호는 출력결과를 방향변경(redirect) 하는데도 사용된다.
> 유사하게 `$` 기호를 쉘 프롬프트로 사용했지만, 앞에서 살펴봤듯이,
> 쉘로 하여금 변수값을 추출하는데도 사용된다.
>
> *쉘*이 `>` 혹은 `$` 기호를 출력하게 되면, 사용자가 뭔가 타이핑하길 기대하고 있다는 것으로
> 해당 기호는 프롬프트를 의미한다.
>
> *사용자 본인*이 `>` 혹은 `$` 기호를 타이핑하게 되면,
> 출력결과를 방향변경하거나 변수 값을 끄집어내는 지시를 쉘에 전달하게 된다.
{: .callout}
`data-shell/creatures` 디렉토리의 예제로 돌아가자.
사람 코드를 읽는 독자에게 목적을 좀더 명확히 하기 위해서 루프의 변수명을 `filename`로 했다.
쉘 자체는 변수명이 어떻게 작명되든지 문제삼지 않는다. 만약 루프를 다음과 같이 작성하거나:
~~~
$ for x in basilisk.dat unicorn.dat
> do
> head -n 3 $x
> done
~~~
{: .language-bash}
혹은:
~~~
$ for temperature in basilisk.dat unicorn.dat
> do
> head -n 3 $temperature
> done
~~~
{: .language-bash}
둘다 정확하게 동일하게 동작한다.
*이렇게는 절대 하지 마세요*.
사람이 프로그램을 이해할 수 있을 때만 프로그램이 유용하기 때문에,
(`x`같은) 의미없는 이름이나, (`temperature`같은) 오해를 줄 수 있는 이름은
오해를 불러일으켜서 독자가 생각하기에 당연히 프로그램이 수행해야 할 작업을 프로그램이 수행하지 못하게 할 가능성을 높인다.
> ## 파일 집합 제한걸기
>
> `data-shell/molecules` 디렉토리에서 다음 루프를 실행하게 되면 출력결과는 어떻게 될까?
>
> ~~~
> $ for filename in c*
> > do
> > ls $filename
> > done
> ~~~
> {: .language-bash}
>
> 1. 어떤 파일도 출력되지 않는다.
> 2. 모든 파일이 출력된다.
> 3. `cubane.pdb`, `octane.pdb`, `pentane.pdb` 파일만 출력된다.
> 4. `cubane.pdb` 파일만 출력된다.
>
> > ## 해답
> > 정답은 4. 와일드카드 `*` 문자는 0 혹은 그 이상 문자를 매칭하게 된다.
> > 따라서, 문자 `c`로 시작하는 문자 다음에 0 혹은 그 이상 문자를 갖는 모든 파일이 매칭된다.
> {: .solution}
>
> How would the output differ from using this command instead?
>
> ~~~
> $ for filename in *c*
> > do
> > ls $filename
> > done
> ~~~
> {: .language-bash}
>
> 1. The same files would be listed.
> 2. All the files are listed this time.
> 3. No files are listed this time.
> 4. The files `cubane.pdb` and `octane.pdb` will be listed.
> 5. Only the file `octane.pdb` will be listed.
>
> > ## 해답
> > 정답은 4. 와일드카드 `*` 문자는 0 혹은 그 이상 문자를 매칭하게 된다.
> > 따라서, `c` 앞에 0 혹은 그 이상 문자가 올 수 있고, `c` 문자 다음에 0 혹은 그 이상 문자가 모두 매칭된다.
> {: .solution}
{: .challenge}
`data-shell/creatures` 디렉토리에서 예제를 계속해서 학습해보자.
다음에 좀더 복잡한 루프가 있다:
~~~
$ for filename in *.dat
> do
> echo $filename
> head -n 100 $filename | tail -n 20
> done
~~~
{: .language-bash}
쉘이 `*.dat`을 전개해서 쉘이 처리할 파일 리스트를 생성한다.
그리고 나서 **루프 몸통(loop body)** 부분이 파일 각각에 대해 명령어 두개를 실행한다.
첫 명령어 `echo`는 명령 라인 매개변수를 표준 출력으로 화면에 뿌려준다.
예를 들어:
~~~
$ echo hello there
~~~
{: .language-bash}
prints:
~~~
hello there
~~~
{: .output}
이 사례에서, 쉘이 파일 이름으로 `$filename`을 전개했기 때문에,
`echo $filename`은 단지 파일 이름만 화면에 출력한다. 다음과 같이 작성할 수 없다는 것에 주의한다:
~~~
$ for filename in *.dat
> do
> $filename
> head -n 100 $filename | tail -n 20
> done
~~~
{: .language-bash}
왜냐하면, `$filename`이 `basilisk.dat`으로 전개될 때 루프 처음에 쉘이 프로그램으로 인식한 `basilisk.dat`를 실행하려고 하기 때문이다.
마지막으로, `head`와 `tail` 조합은 어떤 파일이 처리되든 81-100줄만 선택해서 화면에 뿌려준다.
(파일이 적어도 100줄로 되었음을 가정)
> ## 파일, 디렉토리, 변수 등 이름에 공백
>
> 공백(whitespace)을 사용해서 루프를 돌릴 때 리스트의 각 원소를 구별했다.
> 리스트 원소중 일부가 공백을 갖는 경우, 해당 원소를 인용부호로 감싸서 사용해야 된다.
> 데이터 파일이 다음과 같은 이름으로 되었다고 가정하자:
>
> ~~~
> red dragon.dat
> purple unicorn.dat
> ~~~
> {: .source}
>
> 다음을 사용하여 파일을 처리하려고 한다면:
>
> ~~~
> $ for filename in "red dragon.dat" "purple unicorn.dat"
> > do
> > head -n 100 "$filename" | tail -n 3
> > done
> ~~~
> {: .language-bash}
>
> 파일명에 공백(혹은 다른 특수 문자)를 회피하는 것이 더 단순하다.
>
> 상기 파일은 존재하지 않는다. 그래서 상기 코드를 실행하게 되면, `head` 명령어는
> 파일을 찾을 수가 없어서 예상되는 파일명을 보여주는 오류 메시지가 반환된다:
> ```
> head: cannot open ‘red dragon.dat’ for reading: No such file or directory
> head: cannot open ‘purple unicorn.dat’ for reading: No such file or directory
> ```
> {: .output}
> 상기 루프 내부 `$filename` 파일명 주위 인용부호를 제거하고 공백 효과를 살펴보자.
> `creatures` 디렉토리에서 코드를 실행시키게 되면 `unicorn.dat` 파일에 대한 결과를 루프 명령어 실행 결과를 얻게 됨에 주목한다:
> ```
> head: cannot open ‘red’ for reading: No such file or directory
> head: cannot open ‘dragon.dat’ for reading: No such file or directory
> head: cannot open ‘purple’ for reading: No such file or directory
> CGGTACCGAA
> AAGGGTCGCG
> CAAGTGTTCC
> ```
> {: . output}
{: .callout}
원래 파일 복사문제로 되돌아가서, 다음 루프를 사용해서 문제를 해결해 보자:
~~~
$ for filename in *.dat
> do
> cp $filename original-$filename
> done
~~~
{: .language-bash}
상기 루프는 `cp` 명령문을 각 파일이름에 대해 실행한다.
처음에 `$filename`이 `basilisk.dat`로 전개될 때, 쉘은 다음을 실행한다:
~~~
cp basilisk.dat original-basilisk.dat
~~~
{: .language-bash}
두번째에는 명령문은 다음과 같다:
~~~
cp unicorn.dat original-unicorn.dat
~~~
{: .language-bash}
`cp` 명령어는 아무런 출력결과도 만들어내지 않기 때문에,
루프가 제대로 돌아가는지 확인하기 어렵다.
`echo`로 명령문 앞에 위치시킴으로써, 명령문 각각이 제대로
동작되고 있는 확인하는 것이 가능하다.
다음 도표를 통해서 스크립트가 동작할 때 어떤 작업이 수행하고 있는지 상술하고 있다.
또한 `echo` 명령어를 사려깊이 사용하는 것이 어떻게 훌륭한 디버깅 기술이 되는지도 보여주고 있다.

## Nelle의 파이프라인: 많은 파일 처리하기
Nelle은 이제 `goostats` 프로그램(논문 지도교수가 작성한 쉘 스크립트)을 사용해서 데이터 파일을 처리할 준비가 되었다.
`goostats` 프로그램은 표본추출 단백질 파일에서 통계량을 산출하는데 인자를 두개 받는다:
1. 입력파일 (원본 데이터를 포함)
2. 출력파일 (산출된 통계량을 저장)
아직 쉘을 어떻게 사용하는지 학습단계에 있기 때문에,
단계별로 요구되는 명령어를 차근히 작성하기로 마음먹었다.
첫번째 단계는 적합한 파일을 선택했는지를 확인하는 것이다
--- 'Z'가 아닌 'A' 혹은 'B'로 파일이름이 끝나는 것이 적합한 파일이라는 것을 명심한다.
홈 디렉토리에서 시작해서, 박사과정 Nelle이 다음과 같이 타이핑한다:
~~~
$ cd north-pacific-gyre/2012-07-03
$ for datafile in NENE*[AB].txt
> do
> echo $datafile
> done
~~~
{: .language-bash}
~~~
NENE01729A.txt
NENE01729B.txt
NENE01736A.txt
...
NENE02043A.txt
NENE02043B.txt
~~~
{: .output}
다음 단계는 `goostats` 분석 프로그램이 생성할 파일이름을 무엇으로 할지 결정하는 것이다.
"stats"을 각 입력 파일에 접두어로 붙이는 것이 간단해 보여서, 루프를 변경해서 작업을 수행하도록 한다:
~~~
$ for datafile in NENE*[AB].txt
> do
> echo $datafile stats-$datafile
> done
~~~
{: .language-bash}
~~~
NENE01729A.txt stats-NENE01729A.txt
NENE01729B.txt stats-NENE01729B.txt
NENE01736A.txt stats-NENE01736A.txt
...
NENE02043A.txt stats-NENE02043A.txt
NENE02043B.txt stats-NENE02043B.txt
~~~
{: .output}
`goostats`을 아직 실행하지는 않았지만,
이제 확신할 수 있는 것은 올바른 파일을 선택해서,
올바른 출력 파일이름을 생성할 수 있다는 점이다.
명령어를 반복적으로 타이핑하는 것은 귀찮은 일이지만,
더 걱정이 되는 것은 Nelle이 타이핑 실수를 하는 것이다.
그래서 루프를 다시 입력하는 대신에 위쪽 화살표를 누른다.
위쪽 화살표에 반응해서 컴퓨터 쉘은 한줄에 전체 루프를 다시 보여준다.
(스크립트 각 부분이 구분되는데 세미콜론이 사용됨):
~~~
$ for datafile in NENE*[AB].txt; do echo $datafile stats-$datafile; done
~~~
{: .language-bash}
왼쪽 화살표 키를 사용해서,
Nelle은 `echo`명령어를 `bash goostats`으로 변경하고 백업한다:
~~~
$ for datafile in NENE*[AB].txt; do bash goostats $datafile stats-$datafile; done
~~~
{: .language-bash}
엔터키를 누를 때, 쉘은 수정된 명령어를 실행한다.
하지만, 어떤 것도 일어나지 않는 것처럼 보인다 --- 출력이 아무것도 없다.
잠시뒤에 Nelle은 작성한 스크립트가 화면에 아무것도 출력하지 않아서,
실행되고 있는지, 얼마나 빨리 실행되는지에 대한 정보가 없다는 것을 깨닫는다.
컨트롤+C(Control-C)를 눌러서 작업을 종료하고,
반복할 명령문을 위쪽 화살표로 선택하고,
편집해서 다음과 같이 작성한다:
~~~
$ for datafile in NENE*[AB].txt; do echo $datafile; bash goostats $datafile stats-$datafile; done
~~~
{: .language-bash}
> ## 시작과 끝
>
> 쉘에 ^A, 콘트롤+A(Control-A, `Ctrl-a`)를 타이핑해서 해당 라인 처음으로 가고,
> ^E (`Ctrl-e`, Control-E)를 쳐서 라인의 끝으로 이동한다.
{: .callout}
이번에 프로그램을 실행하면, 매 5초간격으로 한줄을 출력한다:
~~~
NENE01729A.txt
NENE01729B.txt
NENE01736A.txt
...
~~~
{: .output}
1518 곱하기 5초를 60으로 나누면,
작성한 스크립트를 실행하는데 약 2시간 정도 소요된다고 볼 수 있다.
마지막 점검으로, 또다른 터미널 윈도우를 열어서,
`north-pacific-gyre/2012-07-03` 디렉토리로 가서,
`cat stats-NENE01729B.txt`을 사용해서 출력파일 중 하나를 면밀히 조사한다.
출력결과가 좋아보인다.
그래서 커피를 마시고 그동안 밀린 논문을 읽기로 한다.
> ## 역사(history)를 아는 사람은 반복할 수 있다.
>
> 앞선 작업을 반복하는 또다른 방법은 `history` 명령어를 사용하는 것이다.
> 실행된 마지막 수백개 명령어 리스트를 얻고 나서,
> 이들 명령어 중 하나를 반복실행하기 위해서 `!123`("123"은 명령 숫자로 교체된다.)을 사용한다.
> 예를 들어 Nelle이 다음과 같이 타이핑한다면:
>
> ~~~
> $ history | tail -n 5
> ~~~
> {: .language-bash}
> ~~~
> 456 ls -l NENE0*.txt
> 457 rm stats-NENE01729B.txt.txt
> 458 bash goostats NENE01729B.txt stats-NENE01729B.txt
> 459 ls -l NENE0*.txt
> 460 history
> ~~~
> {: .output}
>
> 그리고 나서, 단순히 `!458`을 타이핑함으로써,
> `NENE01729B.txt` 파일에 `goostats`을 다시 실행할 수 있게 된다.
{: .callout}
> ## 다른 이력(history) 명령어
>
> 이력(history)에 접근하는 단축 명령어가 다수 존재한다.
>
> - `Ctrl-R` 탄축키는 "reverse-i-search" 이력 검색모드로
> 입력한 텍스트와 매칭되는 가장 최슨 명령어를 이력에서 찾아서 제시한다.
> `Ctrl-R` 단축키를 한번 혹은 그 이상 누르게 되면 그 이전 매칭을 검색해 준다.
> - `!!` 명령어는 바로 직전 명령어를 불러온다.
> (키보드 윗화살표를 사용하는 것보다 더 편리할수도 편리하지 않을 수도 있다.)
> - `!$` 명령어는 마지막 명령문의 마지막 단어를 불러온다.
> 기대했던 것보다 훨씬 유용할 수 있다:
> `bash goostats NENE01729B.txt stats-NENE01729B.txt` 명령문을 실행한 후에
> `less !$`을 타이핑하게 되면 `stats-NENE01729B.txt` 파일을 찾아준다.
> 키보드 위화살표를 눌러 명령라인을 편집하는 것보다 훨씬 빠르다.
{: .callout}
> ## 루프 내부에서 파일에 저장하기 - 1부
>
> `data-shell/molecules` 디렉토리에 있다고 가정하자.
> 다음 루프의 효과는 무엇인가?
>
> ~~~
> $ for alkanes in *.pdb
> > do
> > echo $alkanes
> > cat $alkanes > alkanes.pdb
> > done
> ~~~
> {: .language-bash}
>
> 1. `fructose.dat`, `glucose.dat`, `sucrose.dat`을 출력하고, `sucrose.dat`에서 나온 텍스트를 `xylose.dat`에 저장된다.
> 2. `fructose.dat`, `glucose.dat`, `sucrose.dat`을 출력하고, 모든 파일 3개에서 나온 텍스트를 합쳐 `xylose.dat`에 저장된다.
> 3. `fructose.dat`, `glucose.dat`, `sucrose.dat`, `xylose.dat`을 출력하고, `sucrose.dat`에서 나온 텍스트를 `xylose.dat`에 저장된다.
> 4. 위 어느 것도 아니다.
>
> > ## 해답
> > 1. 순차적으로 각 파일의 텍스트가 `alkanes.pdb` 파일에 기록된다.
> > 하지만, 루프가 매번 반복될 때마다 파일에 덮어쓰기가 수행되어서 마지막 `alkanes.pdb` 파일 텍스트만
> > `alkanes.pdb` 파일에 기록된다.
> {: .solution}
{: .challenge}
> ## 루프 내부에서 파일에 저장하기 - 2부
>
> 이번에도 `data-shell/molecules` 디렉토리에 있다고 가정하고, 다음 루프 실행 출력결과는 무엇일까?
>
> ~~~
> $ for datafile in *.pdb
> > do
> > cat $datafile >> all.pdb
> > done
> ~~~
> {: .language-bash}
>
> 1. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb` 파일에 나온 모든 모든 텍스트가 하나로 붙여져서
> `all.pdb` 파일에 저장된다.
> 2. `ethane.pdb` 파일에 나온 텍스트만 `all.pdb` 파일에 저장된다.
> 3. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb`, `propane.pdb` 파일에서 나온 모든 텍스트가
> 하나로 풑여져서 `all.pdb` 파일에 저장된다.
> 4. `cubane.pdb`, `ethane.pdb`, `methane.pdb`, `octane.pdb`, `pentane.pdb`, `propane.pdb` 파일에서 나온
> 모든 텍스트가 화면에 출력되고 `all.pdb` 파일에 저장된다.
>
> > ## 해답
> > 정답은 3. 명령어 실행 출력결과를 방향변경하여 덮었는 것이 아니라 `>>` 기호는 파일에 덧붙인다.
> > `cat` 명령어에서 나온 출력결과가 파일로 방향변경되어 어떤 출력결과도 화면에 출력되지는 않는다.
> {: .solution}
{: .challenge}
> ## 시운전(Dry Run)
>
> 루프는 한번에 많은 작업을 수행하는 방식이다 --- 만약 잘못된 것이 있다면,
> 한번에 실수를 대단히 많이 범하게 된다.
> 루프가 수행하는 작업을 점검하는 한 방법이 실제로 루프를 돌리는 대신에
> `echo` 명령어를 사용하는 것이다.
>
> 실제로 명령어를 실행하지 않고, 다음 루프가 실행할 명령어를 머릿속으로 미리보고자 한다고 가정한다:
>
> ~~~
> $ for file in *.pdb
> > do
> > analyze $file > analyzed-$file
> > done
> ~~~
> {: .language-bash}
>
> 아래 두 루프 사이에 차이는 무엇이고, 어느 것을 시운전으로 실행하고 싶은가?
>
> ~~~
> # Version 1
> $ for file in *.pdb
> > do
> > echo analyze $file > analyzed-$file
> > done
> ~~~
> {: .language-bash}
>
> ~~~
> # Version 2
> $ for file in *.pdb
> > do
> > echo "analyze $file > analyzed-$file"
> > done
> ~~~
> {: .language-bash}
>
> > ## 해답
> > 두번째 버젼을 실행하면 좋을 것이다.
> > 달러 기호로 접두명을 주었기 때문에 루프 변수를 확장해서 인용부호로 감싼 모든 것을 화면에 출력한다.
> >
> > 첫번째 버전은 `echo analyze $file` 명령을 수행해서 `analyzed-$file` 파일로
> > 출력결과를 방향변경하여 저장시킨다. 따라서 파일이 쭉 자동생성된다:`analyzed-cubane.pdb`,
> > `analyzed-ethane.pdb` ...
> >
> > 두가지 버젼을 직접 실행해보고 출력결과를 살펴보자!
> > `analyzed-*.pdb` 파일을 열어서 파일에 기록된 내용도 살펴본다.
> {: .solution}
{: .challenge}
> ## 중첩루프(Nested Loops)
>
> 다른 화합물과 다른 온도를 갖는 조합을 해서, 각 반응율 상수를 측정하는
> 실험을 조직하도록 이에 상응하는 디렉토리 구조를 갖추고자 한다.
> 다음 코드 실행결과는 어떻게 될까?
>
> ~~~
> $ for species in cubane ethane methane
> > do
> > for temperature in 25 30 37 40
> > do
> > mkdir $species-$temperature
> > done
> > done
> ~~~
> {: .language-bash}
>
> > ## 해답
> > 중첩 루프(루프 내부에 루프가 포함됨)를 생성하게 된다.
> > 외부 루프에 각 화학물이, 내부 루프(중첩된 루프)에 온도 조건을 반복하게 되서,
> > 화학물과 온도를 조합한 새로운 디렉토리가 쭉 생성된다.
> >
> > 직접 코드를 실행해서 어떤 디렉토리가 생성되는지 확인한다!
> {: .solution}
{: .challenge}
| 23.700292 | 114 | 0.637283 | kor_Hang | 1.00001 |
912b1e112548b60e6bc6100387aff62084aa29e9 | 3,646 | md | Markdown | desktop-src/SecGloss/b-gly.md | velden/win32 | 94b05f07dccf18d4b1dbca13b19fd365a0c7eedc | [
"CC-BY-4.0",
"MIT"
] | 552 | 2019-08-20T00:08:40.000Z | 2022-03-30T18:25:35.000Z | desktop-src/SecGloss/b-gly.md | velden/win32 | 94b05f07dccf18d4b1dbca13b19fd365a0c7eedc | [
"CC-BY-4.0",
"MIT"
] | 1,143 | 2019-08-21T20:17:47.000Z | 2022-03-31T20:24:39.000Z | desktop-src/SecGloss/b-gly.md | velden/win32 | 94b05f07dccf18d4b1dbca13b19fd365a0c7eedc | [
"CC-BY-4.0",
"MIT"
] | 1,287 | 2019-08-20T05:37:48.000Z | 2022-03-31T20:22:06.000Z | ---
description: Contains definitions of security terms that begin with the letter B.
ROBOTS: NOINDEX, NOFOLLOW
ms.assetid: 2e570727-7da0-4e17-bf5d-6fe0e6aef65b
title: B (Security Glossary)
ms.topic: article
ms.date: 05/31/2018
---
# B (Security Glossary)
[A](a-gly.md) B [C](c-gly.md) [D](d-gly.md) [E](e-gly.md) F [G](g-gly.md) [H](h-gly.md) [I](i-gly.md) J [K](k-gly.md) [L](l-gly.md) [M](m-gly.md) [N](n-gly.md) [O](o-gly.md) [P](p-gly.md) Q [R](r-gly.md) [S](s-gly.md) [T](t-gly.md) [U](u-gly.md) [V](v-gly.md) [W](w-gly.md) [X](x-gly.md) Y Z
<dl> <dt>
<span id="_security_backup_authority_gly"></span><span id="_SECURITY_BACKUP_AUTHORITY_GLY"></span>**backup authority**
</dt> <dd>
A trusted application running on a secure computer that provides secondary storage for the session keys of its clients. The backup authority stores session keys as key BLOBs that are encrypted with the backup authority's public key.
</dd> <dt>
<span id="_security_base_content_type_gly"></span><span id="_SECURITY_BASE_CONTENT_TYPE_GLY"></span>**base content type**
</dt> <dd>
A type of data contained in a PKCS \#7 message. Base content types only contain data, no cryptographic enhancements such as hashes or signatures. Currently, the only base content type is the Data content type.
</dd> <dt>
<span id="_security_base_cryptographic_functions_gly"></span><span id="_SECURITY_BASE_CRYPTOGRAPHIC_FUNCTIONS_GLY"></span>**base cryptographic functions**
</dt> <dd>
The lowest level of functions in the CryptoAPI architecture. They are used by applications and other high-level CryptoAPI functions to provide access to CSP-provided cryptographic algorithms, secure key generation, and secure storage of secrets.
See also [*cryptographic service providers*](c-gly.md).
</dd> <dt>
<span id="_security_basic_encoding_rules_gly"></span><span id="_SECURITY_BASIC_ENCODING_RULES_GLY"></span>**Basic Encoding Rules**
</dt> <dd>
(BER) The set of rules used to encode ASN.1 defined data into a stream of bits (zeros or ones) for external storage or transmission. A single ASN.1 object may have several equivalent BER encodes. BER is defined in CCITT Recommendation X.209. This is one of the two encoding methods currently used by CryptoAPI.
</dd> <dt>
<span id="_security_ber_gly"></span><span id="_SECURITY_BER_GLY"></span>**BER**
</dt> <dd>
See *Basic Encoding Rules*.
</dd> <dt>
<span id="_security_big_endian_gly"></span><span id="_SECURITY_BIG_ENDIAN_GLY"></span>**big-endian**
</dt> <dd>
A memory or data format in which the most significant byte is stored at the lower address or arrives first.
See also [*little-endian*](l-gly.md).
</dd> <dt>
<span id="_security_blob_gly"></span><span id="_SECURITY_BLOB_GLY"></span>**BLOB**
</dt> <dd>
A generic sequence of bits that contain one or more fixed-length header structures plus context specific data.
See also [*key BLOBs*](k-gly.md), [*certificate BLOBs*](c-gly.md), [*certificate name BLOBs*](c-gly.md), and [*attribute BLOBs*](a-gly.md).
</dd> <dt>
<span id="_security_block_cipher_gly"></span><span id="_SECURITY_BLOCK_CIPHER_GLY"></span>**block cipher**
</dt> <dd>
A cipher algorithm that encrypts data in discrete units (called blocks), rather than as a continuous stream of bits. The most common block size is 64 bits. For example, DES is a block cipher.
See also [*stream cipher*](s-gly.md).
</dd> <dt>
<span id="_security_bulk_encryption_key_gly"></span><span id="_SECURITY_BULK_ENCRYPTION_KEY_GLY"></span>**bulk encryption key**
</dt> <dd>
A session key derived from a master key. Bulk encryption keys are used in [*Schannel*](s-gly.md) encryption.
</dd> </dl>
| 39.204301 | 310 | 0.729292 | eng_Latn | 0.872546 |
912b3493a48828811ee290d29f546625c794918f | 705 | md | Markdown | docs/api/ESCWalkContext.md | StraToN/unofficial-escoria-reloaded | ccb34e319b716b4d3afc540fbb970348d872ffbf | [
"MIT"
] | 7 | 2021-03-09T08:13:45.000Z | 2021-09-20T07:12:08.000Z | docs/api/ESCWalkContext.md | StraToN/unofficial-escoria-reloaded | ccb34e319b716b4d3afc540fbb970348d872ffbf | [
"MIT"
] | 17 | 2021-05-15T16:10:14.000Z | 2021-07-04T17:00:05.000Z | docs/api/ESCWalkContext.md | StraToN/unofficial-escoria-reloaded | ccb34e319b716b4d3afc540fbb970348d872ffbf | [
"MIT"
] | null | null | null | <!-- Auto-generated from JSON by GDScript docs maker. Do not edit this document directly. -->
# ESCWalkContext
**Extends:** [Object](../Object)
## Description
The walk context describes the target of a walk command and if that command
should be executed fast
## Property Descriptions
### target\_object
```gdscript
var target_object: ESCObject
```
Target object that the walk command tries to reach
### target\_position
```gdscript
var target_position: Vector2 = "(0, 0)"
```
The target position
### fast
```gdscript
var fast: bool
```
Wether to move fast
## Method Descriptions
### \_init
```gdscript
func _init(p_target_object: ESCObject, p_target_position: Vector2, p_fast: bool)
```
| 15.326087 | 93 | 0.719149 | eng_Latn | 0.903082 |
912ba768002369b6d06e5b9801803f10a0ad41ee | 9,164 | md | Markdown | content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md | lnds/lnds-site | c7d8483a764c91f1653c77ab6934c4f34d847f62 | [
"MIT"
] | null | null | null | content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md | lnds/lnds-site | c7d8483a764c91f1653c77ab6934c4f34d847f62 | [
"MIT"
] | null | null | null | content/post/2009/2009-05-26-las-maquinas-del-fin-del-mundo-intermedio/index.md | lnds/lnds-site | c7d8483a764c91f1653c77ab6934c4f34d847f62 | [
"MIT"
] | null | null | null | ---
comments: true
date: 2009-05-26 20:53:24
layout: post
slug: las-maquinas-del-fin-del-mundo-intermedio
title: Las máquinas del fin del mundo (intermedio)
wordpress_id: 191
categories:
- General
- Paradigmas
---
Ya hemos visto [una posición](http://www.lnds.net/2009/05/el-desafio-del-nuevo-ludita.html), en uno de los extremos, sobre que hacer ante las amenazas de la tecnología para nuestra sociedad.
En 2006 escribí este artículo, que he decidido retocar y adaptar a mi serie actual sobre los desafíos que nos impone la tecnología, creo que es apropiado publicarlo en este momento, como una suerte de puente hacia la otra propuesta extrema, sobre que hacer ante estos escenarios, en que podemos ser sobrepasados por la tecnología.
Las máquinas del fin del mundo
Leí con asombro un post de mi hermano, sobre una cadena que [anuncia el fin del mundo para el 6 de junio de este año (6/6/6)](http://www.ricardodiaz.org/archives/2006/04/fin.html) [^1].
Aparte de ser una interesante coincidencia numérica, ¿qué razones hay para pensar de que el mundo debe acabar?
## El fin por culpa de la tecnología
La verdad es que el fin del mundo ha sido anunciado muchas veces, y no sólo por fanáticos religiosos.

Las visiones apocalípticas normalmente anuncian todo tipo de cataclismos, pero, como sería de esperar, en el último tiempo, se han agregado otras causas para el fin del mundo, muchas consecuencia de nuestro abuso de la tecnología:
* El fin del mundo por el uso de armas bioquímicas
* Guerra Nuclear total
* La explosión de plantas nucleares
* [Una super máquina que crea un vórtice](http://www.lnds.net/2008/09/buuum.html), o un agujero negro
* El calentamiento global y una nueva era glacial
* El surgimiento de inteligencias artificiales que destruyen a su creador
Esta última "amenaza" es la que me interesa explorar ahora, porque tiene que ver con temas que domino más.
## El miedo a la inteligencia artificial

El temor al alzamiento de las máquinas ha sido discutido en taquilleras películas como Terminator y Matrix, y de una forma más elegante en "2001 Odisea Espacial".
Cuando se habla de esta visión de la máquina que se alza contra su creador, muchos comparan con el [Frankenstein](http://www.literature.org/authors/shelley-mary/frankenstein/), de Mary Shelley. En realidad no es lo mismo, pues la "criatura" es un humano vuelt
o a la vida (bueno, un cuarpo humano reanimado, construido a partir de partes de muchos cadaveres).
En Matrix no sabemos cómo son las "máquinas", vemos robots en forma de calamares, pero son sólo "trajes", para acomodar inteligencias artificiales. Las inteligencias artificiales viven en la matrix, un mundo virtual, que tiene un soporte físico (hardware), alimentado por la energía bioquímica de millones de seres humanos.
En Terminator la idea es similar, una inteligencia artificial surge desde una red militar (skynet), y esta intelgencia decide destruir al hombre, para eso se sirve de robots, conocidos como exterminadores, pero nuevamente, las IA están en otra parte.
John Connor sabe que debe destruir Skynet antes que se forme, porque probablemente el acceso a Skynet es imposible. Neo se dirige hacia una suerte de nucleo de la Matrix, porque sabe que aunque maten a todos los robots calamares estos no son el problema.

La brillante saga de ciencia ficción Hyperión, de Dan Simmons, nos muestra una guerra entre las IA y las personas, con una interesante variante, las IA han decidido sacrificarse a si mismas para crear IA más evolucionadas (¿seríamos capaces de sacrificarnos por la especie que nos suceda?).
En todos estos casos, ls enemigos, no vistos, son Inteligencias Artificiales que viven, flotan, por decirlo de alguna manera, en una red, o soporte físico. Son programas, algoritmos, heurísticas, en cierta manera, pero con capacidad de aprender y manipular su entorno con el fin de protegerse, por ejemplo, construyendo robots.
No son los robots los peligrosos, estos son herramientas, son las inteligencias artificiales las que comandan y las peligrosas.
Pero, por su naturaleza, las IA son escurridizas. En 2001, HAL es eliminado, pues tiene un soporte físico de fácil acceso. Pero en Matrix y Terminator, eso no es posible, y los heroes de esa sagas, deben tratar de evitar que surjan, en el caso de Terminator, o atacar el núcleo donde se encuentran, como en Matrix.
Sin embargo, la estrategia de John Connor es más astuta que la de Neo, porque John Connor quiere evitar que Skynet llegue a existir, pero se produce una interesante paradoja, que si John Connor destruye Skynet, el dejará de existir. En el caso de Matrix, sabemos que Neo no puede destruir a las IA, que la lucha es eterna y se repetirá por siempre, porque Neo en cierta manera es un mecanismo de la misma matrix para reorganizarse.
Quizás por eso Matrix como trilogía decepciona tanto. Porque esperabamos un desenlace más emocionante, pero en realidad nos llevan a un final decepcionante, en que las IA ganan. Pero también, nos queda claro, que en cierta manera, la matrix sigue dependiendo de humanos libres, una élite acaso, que viven en Zión.
Hemos visto, que sólo en 2001 el héroe logra eliminar la amenaza de la IA. HAL falla porque es incapaz en protegerse a si mismo.
## Vida Artificial
Actualmente se postula que antes de construir una IA, debemos contruir niveles inferiores de inteligencia. Estos rdimentos de inteligencia artificial ya están disponibles, [incluso al alcance de cualquiera de nosotros](http://www.lnds.net/2009/03/el-sueno-de-las-ovejas-electricas.html). Estos rudimentos, constituyen los "instintos" de las IA.
Sin esos instintos básicos, que le permitan a la IA desarrollarse, no será posible que esta llegue a ser conciente.
Una Inteligencia Artificial debería
1. Preservarse
2. Reproducirse
3. Evolucionar
Estos tres requisitos básicos son las bases de la denominada [tesis fuerte de la vida artificia](http://vidaartificial.com/index.php?title=Vida_Artificial)l, en esencia esta dice que una vida es un modelo operacional, y si satisface las tres operaciones mencionadas, entonces da lo mismo el soporte físico de la vida.
La satisfacción de estas tres necesidades de las IAs son las que justificarían su hostilidad hacia nosotros. Efectivamente, si hacemos precaria su preservación, si tienen que competir con nosotros por los recursos, o somos un estorbo dentro de su evolución, las IAs pueden volverse contra sus creadores.
¿Dónde surgirá la Inteligencia Artificial?
Lo interesante, es saber si las inteligencias artificiales son capaces de desarrollarse fuera de su soporte físico, o antes de llegar a tenerlo, es decir, [¿pueden las IA surgir dentro de la red?](http://www.lnds.net/2008/12/y-si-las-inteligencias-artificiales-ya-e.html).
En Terminator, lo primero que se desarrolla es Skynet, una red informática, que empieza a controlar las factorías necesarias para desarrollar las armas y robots exterminadores.
Entonces, además de preocuparnos del mundo industrial, y de la amenazas de las máquinas, que principalmente son amenazas ambientales, es interesante concentrarse en entender las amenazas que representan las tecnologías de la información, es decir, la amenaza de que se desarrolle una IA hostil.
Consideren el siguiente fragmento:
> "¿Saben lo que es barato en el siglo 21? El tiempo de computación. ¿Saben lo que es caro? El juicio humano. Ambos no son intercambiables. Los humanos son buenos entendiendo cosas, los computadores son buenos contando cosas, pero los humanos apestan al contar, y os computadores apestan al entender."
> "¿Conocen los algoritmos genéticos? Tomen cualquier problema y generen 10 billones de pogramas de computador al azar y póngalos a resolver el problema. Tomen el 10% de los que lo hacen mejor, usen variantes al azar de ellos y repítanlos otras 10 billones de veces. Repítanlo 10 billones de veces más, y vuelvan en un día o dos, y descubrirán que el computador ha "evolucionado" a una extraña y bizarra respuesta que a ningún humano jamás se le habría ocurrido."
>
> "Ningún humano va a lograr escribir el código para una Inteligencia Artificial (IA). La inteligencia es una propiedad emergente de factores evolutivos, no de la planificación central. Es anarquía, no estalinismo. ¿Lo entienden?"[^2].
¿Que tal si el surgimiento de la inteligencia artificial es algo inevitable, y eso hace que tengamos que empezar a aceptar que tendremos que lidiar con seres potencialmente superiores a nosotros?
Hay una propuesta extrema, que nos dice lo que debemos hacer en este caso, y que al contrario de lo propuesto por los neoluditas, y los que llaman a destruir, o abandonar la tecnología, toma un camino radicalmente opuesto, y es el que vamos a explorar en el siguiente post.
[^1]: La primera parte de este post fue escrita en abril de 2006, bajo el título de [Las máquinas del fin del mundo](/2006/04/las-maquinas-del-fin-del-mundo.html).
[^2]: Cory Doctorow, "Pester Power", relato publicado en Communications of ACM, diciembre 2008.
| 78.324786 | 463 | 0.791248 | spa_Latn | 0.997531 |
912c31695f32b5ae8ab45d6507060e2a7a205424 | 1,480 | md | Markdown | intl.en-US/Product Introduction/Benefits.md | vlgnaw/emapreduce | 37918944befffc3895b53fe8d8ae5d793331f401 | [
"MIT"
] | null | null | null | intl.en-US/Product Introduction/Benefits.md | vlgnaw/emapreduce | 37918944befffc3895b53fe8d8ae5d793331f401 | [
"MIT"
] | null | null | null | intl.en-US/Product Introduction/Benefits.md | vlgnaw/emapreduce | 37918944befffc3895b53fe8d8ae5d793331f401 | [
"MIT"
] | null | null | null | # Benefits {#concept_j4w_dky_w2b .concept}
E-MapReduce has some practical strength over self-built clusters. For example, it provides some convenient and controllable means to manage its clusters. In addition, it also has the following strengths:
- Usability
User can select the required ECS types and disks and select the required software for automatic deployment.
Users can apply for cluster resources at the corresponding position according to the geographical location where users or the data source are located. Now, Alibaba Cloud ECS supports regions, including China East 1, China East 2, China North 1, China North 2, China South 1, Singapore, Hong Kong, US East 1 and US West 1. E-MapReduce supports regions including China North 2, China East 1, China East 2 and China South 1, and later it will extend to all the regions supported by Alibaba Cloud ECS.
- Low price
The user can create a cluster as needed, that is, it can release the cluster after an offline task running is completed and add a node dynamically when needed.
- Deep integration
E-MapReduce can be subject to deep integration with other Alibaba Cloud products, so that they can be used as the input source or output destination of Hadoop or Spark computing engine in E-MapReduce.
- Security
E-MapReduce integrates Alibaba Cloud RAM resource permission management system, so that it can isolate the service permissions through the primary account or sub-accounts.
| 61.666667 | 501 | 0.787162 | eng_Latn | 0.999365 |
912c7f22c8273129c8aa65f19f04180f50e966d9 | 1,549 | md | Markdown | business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md | nschonni/dynamics365smb-docs | 619182073e912c1373c58db16c20f0770aefc1b3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md | nschonni/dynamics365smb-docs | 619182073e912c1373c58db16c20f0770aefc1b3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | business-central/LocalFunctionality/UnitedKingdom/how-to-print-direct-sales-and-purchase-details-reports.md | nschonni/dynamics365smb-docs | 619182073e912c1373c58db16c20f0770aefc1b3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: How to Print Direct Sales and Purchase Details Reports | Microsoft Docs
description: The **Direct Sales Details** and **Direct Purchase Details** reports include headers with order numbers and descriptions from sales and purchase documents.
services: project-madeira
documentationcenter: ''
author: SorenGP
ms.service: dynamics365-business-central
ms.topic: article
ms.devlang: na
ms.tgt_pltfrm: na
ms.workload: na
ms.search.keywords:
ms.date: 10/01/2018
ms.author: sgroespe
---
# Print Direct Sales and Purchase Details Reports
The **Direct Sales Details** and **Direct Purchase Details** reports include headers with order numbers and descriptions from sales and purchase documents.
You can filter or select data for these reports based on:
- General ledger account number
- Document number
- Posting date
## To print direct sales and direct purchase details reports
1. Choose the  icon, enter **Direct Sales Details** or **Direct Purchase Details**, and then choose the related link.
2. On the **G/L Entry** FastTab, select the appropriate filters.
3. Choose the **Print** button to print the report or choose the **Preview** button to view it on the screen.
## See Also
[United Kingdom Local Functionality](united-kingdom-local-functionality.md)
[Sales](../../sales-manage-sales.md)
[Purchasing](../../purchasing-manage-purchasing.md)
| 41.864865 | 221 | 0.723047 | eng_Latn | 0.96782 |
912cdb13d5de0c13eb0d4826d9af66f3f2c938dd | 50 | md | Markdown | README.md | jorge-matricali/jwt-crack | 16262f521ee871a0e2581e8b2016c9477c884be0 | [
"MIT"
] | 1 | 2019-12-10T23:52:16.000Z | 2019-12-10T23:52:16.000Z | README.md | jorge-matricali/jwt-crack | 16262f521ee871a0e2581e8b2016c9477c884be0 | [
"MIT"
] | null | null | null | README.md | jorge-matricali/jwt-crack | 16262f521ee871a0e2581e8b2016c9477c884be0 | [
"MIT"
] | 1 | 2019-11-05T16:47:23.000Z | 2019-11-05T16:47:23.000Z | # jwt-crack
JWT brute force cracker written in C.
| 16.666667 | 37 | 0.76 | eng_Latn | 0.986595 |
912cde06c9d7598ad824441dddc898c518243f1a | 573 | md | Markdown | README.md | rmurai0610/DArgs | 170b7d354c90c3212886eb57e04999c2f16d7de0 | [
"MIT"
] | null | null | null | README.md | rmurai0610/DArgs | 170b7d354c90c3212886eb57e04999c2f16d7de0 | [
"MIT"
] | null | null | null | README.md | rmurai0610/DArgs | 170b7d354c90c3212886eb57e04999c2f16d7de0 | [
"MIT"
] | null | null | null | # DArgs - Dumb Argument Parser for C++
DArgs is a minimal, simple argument parser for C++.
DArgs parses the arguments as the options are defined, enabling the user to use DArgs with minimal lines of code.
## Example
```
DArgs::DArgs dargs(argc, argv);
std::string dataset = dargs("--dataset", "Path to the dataset to load", std::string("dataset/default.txt"));
int n_iters = dargs("--n_iters", "Number of iterations", 200);
double alpha = dargs("--alpha", "alpha value", 0.0001f);
if (!dargs.check()) {
dargs.print_help();
exit(EXIT_FAILURE);
}
```
| 35.8125 | 113 | 0.675393 | eng_Latn | 0.714656 |
912d3f6be97e0a30352980eab51df33a4d60bfa1 | 2,930 | md | Markdown | docs/framework/wcf/guidelines-and-best-practices.md | nicolaiarocci/docs.it-it | 74867e24b2aeb9dbaf0a908eabd8918bc780d7b4 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/wcf/guidelines-and-best-practices.md | nicolaiarocci/docs.it-it | 74867e24b2aeb9dbaf0a908eabd8918bc780d7b4 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/wcf/guidelines-and-best-practices.md | nicolaiarocci/docs.it-it | 74867e24b2aeb9dbaf0a908eabd8918bc780d7b4 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Linee guida e suggerimenti
ms.date: 03/30/2017
helpviewer_keywords:
- WCF, guidelines
- best practices [WCF], application design
- Windows Communication Foundation, best practices
- WCF, best practices
- Windows Communication Foundation, guidelines
ms.assetid: 5098ba46-6e8d-4e02-b0c5-d737f9fdad84
ms.openlocfilehash: 1c16c0936f729636aaa56e6e50c33e18c2b0f788
ms.sourcegitcommit: 5b6d778ebb269ee6684fb57ad69a8c28b06235b9
ms.translationtype: MT
ms.contentlocale: it-IT
ms.lasthandoff: 04/08/2019
ms.locfileid: "59181342"
---
# <a name="guidelines-and-best-practices"></a>Linee guida e suggerimenti
In questa sezione contiene argomenti che forniscono le linee guida per la creazione di applicazioni Windows Communication Foundation (WCF).
## <a name="in-this-section"></a>In questa sezione
[Procedure consigliate: Controllo delle versioni dei contratti dati](../../../docs/framework/wcf/best-practices-data-contract-versioning.md)
Viene illustrato come e quando creare contratti dati che non determinano interruzioni quando vengono create versioni successive.
[Controllo delle versioni dei servizi](../../../docs/framework/wcf/service-versioning.md)
Viene illustrato come controllare la versione in WCF. Dopo la distribuzione potrebbe essere necessario modificare i servizi (e gli endpoint che espongono) per soddisfare, ad esempio, la modifica di requisiti aziendali o di requisiti IT o per correggere problemi. Ogni modifica introduce una nuova versione del servizio.
[Bilanciamento del carico](../../../docs/framework/wcf/load-balancing.md)
Elenca le linee guida per il bilanciamento del carico con una Web farm.
[Controllo dell'utilizzo di risorse e miglioramento delle prestazioni](../../../docs/framework/wcf/controlling-resource-consumption-and-improving-performance.md)
Descrive le proprietà progettate per impedire il consumo indebito di risorse e per migliorare la sicurezza e fa riferimento a informazioni più complete riguardo al loro utilizzo.
[Distribuzione di applicazioni WCF con ClickOnce](../../../docs/framework/wcf/deploying-wcf-applications-with-clickonce.md)
Descrive le considerazioni da fare in caso di utilizzo della funzionalità ClickOnce.
## <a name="reference"></a>Riferimenti
<xref:System.ServiceModel>
<xref:System.Runtime.Serialization>
## <a name="related-sections"></a>Sezioni correlate
[Panoramica dei concetti](../../../docs/framework/wcf/conceptual-overview.md)
[Programmazione WCF di base](../../../docs/framework/wcf/basic-wcf-programming.md)
## <a name="see-also"></a>Vedere anche
- [Informazioni su Windows Communication Foundation](../../../docs/framework/wcf/whats-wcf.md)
- [Esempi di Windows Communication Foundation (WCF)](./samples/index.md)
- [Panoramica dei concetti](../../../docs/framework/wcf/conceptual-overview.md)
- [Creazione di client](../../../docs/framework/wcf/building-clients.md)
| 55.283019 | 322 | 0.768259 | ita_Latn | 0.943743 |
912db653b18767cd57ca5921077b47029a577796 | 1,947 | md | Markdown | dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | dynamicsax2012-technet/salestransaction-channelreferenceid-property-microsoft-dynamics-commerce-runtime-datamodel.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: SalesTransaction.ChannelReferenceId Property (Microsoft.Dynamics.Commerce.Runtime.DataModel)
TOCTitle: ChannelReferenceId Property
ms:assetid: P:Microsoft.Dynamics.Commerce.Runtime.DataModel.SalesTransaction.ChannelReferenceId
ms:mtpsurl: https://technet.microsoft.com/en-us/library/microsoft.dynamics.commerce.runtime.datamodel.salestransaction.channelreferenceid(v=AX.60)
ms:contentKeyID: 49834319
ms.date: 05/18/2015
mtps_version: v=AX.60
f1_keywords:
- Microsoft.Dynamics.Commerce.Runtime.DataModel.SalesTransaction.ChannelReferenceId
dev_langs:
- CSharp
- C++
- VB
---
# ChannelReferenceId Property
Gets or sets the channel reference identifier.
**Namespace:** [Microsoft.Dynamics.Commerce.Runtime.DataModel](microsoft-dynamics-commerce-runtime-datamodel-namespace.md)
**Assembly:** Microsoft.Dynamics.Commerce.Runtime.Entities (in Microsoft.Dynamics.Commerce.Runtime.Entities.dll)
## Syntax
``` vb
'Declaration
<DataMemberAttribute> _
<ColumnAttribute("CHANNELREFERENCEID")> _
Public Property ChannelReferenceId As String
Get
Set
'Usage
Dim instance As SalesTransaction
Dim value As String
value = instance.ChannelReferenceId
instance.ChannelReferenceId = value
```
``` csharp
[DataMemberAttribute]
[ColumnAttribute("CHANNELREFERENCEID")]
public string ChannelReferenceId { get; set; }
```
``` c++
[DataMemberAttribute]
[ColumnAttribute(L"CHANNELREFERENCEID")]
public:
property String^ ChannelReferenceId {
String^ get ();
void set (String^ value);
}
```
#### Property Value
Type: [System.String](https://technet.microsoft.com/en-us/library/s1wwdcbf\(v=ax.60\))
Returns [String](https://technet.microsoft.com/en-us/library/s1wwdcbf\(v=ax.60\)).
## See Also
#### Reference
[SalesTransaction Class](salestransaction-class-microsoft-dynamics-commerce-runtime-datamodel.md)
[Microsoft.Dynamics.Commerce.Runtime.DataModel Namespace](microsoft-dynamics-commerce-runtime-datamodel-namespace.md)
| 27.422535 | 146 | 0.787365 | yue_Hant | 0.800152 |
912db6eab825327c7e96e88ea5dd1ef514eb9007 | 1,171 | md | Markdown | AlchemyInsights/plan-passwordless-deployment.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2020-05-19T19:07:24.000Z | 2020-05-19T19:07:24.000Z | AlchemyInsights/plan-passwordless-deployment.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2022-02-09T06:52:18.000Z | 2022-02-09T06:52:35.000Z | AlchemyInsights/plan-passwordless-deployment.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2019-10-09T20:27:31.000Z | 2019-10-09T20:27:31.000Z | ---
title: Planowanie wdrożenia bez haseł
ms.author: pebaum
author: pebaum
manager: scotv
ms.date: 04/14/2021
audience: Admin
ms.topic: article
ms.service: o365-administration
ROBOTS: NOINDEX, NOFOLLOW
localization_priority: Priority
ms.collection: Adm_O365
ms.custom:
- "10394"
- "9005762"
ms.openlocfilehash: a167e33a5d9cb7900aabceda37466d731d784345
ms.sourcegitcommit: 8bc60ec34bc1e40685e3976576e04a2623f63a7c
ms.translationtype: MT
ms.contentlocale: pl-PL
ms.lasthandoff: 04/15/2021
ms.locfileid: "51837434"
---
# <a name="plan-your-passwordless-deployment"></a>Planowanie wdrożenia bez haseł
Uaktualnianie do logowania alternatywnego przy użyciu metody uwierzytelniania bez użycia hasła. Weryfikacja dwuskładnikowa, logowanie telefoniczne lub generatory kodów mogą zwiększyć bezpieczeństwo i utrudnić innym osobom włamanie się na konto.
Do wyboru są: funkcja Windows Hello dla firm, aplikacja uwierzytelniania firmy Microsoft lub klucze zabezpieczeń FIDO2. Aby ustalić, która metoda jest najlepsza i jak wdrożyć je na urządzeniu, skorzystaj z [przewodnika wdrażania bez haseł.](https://admin.microsoft.com/adminportal/home?#/modernonboarding/passwordlesssetup)
| 40.37931 | 324 | 0.824082 | pol_Latn | 0.993109 |
912dcc2e7a7029ce548e782c115d146f512c7804 | 237 | md | Markdown | kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md | JonathanxD/Kores | 236f7db6eeef7e6238f0ae0dab3f3b05fc531abb | [
"MIT-0",
"MIT"
] | 1 | 2019-04-16T10:42:02.000Z | 2019-04-16T10:42:02.000Z | kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md | koresframework/Kores | b6ab31b1d376ab501fd9f481345c767cb0c37d04 | [
"MIT-0",
"MIT"
] | 8 | 2020-12-12T06:48:34.000Z | 2021-08-15T22:34:49.000Z | kdocs/-kores/com.github.jonathanxd.kores/-mutable-instructions/-mutable-instructions.md | koresframework/Kores | b6ab31b1d376ab501fd9f481345c767cb0c37d04 | [
"MIT-0",
"MIT"
] | null | null | null | //[Kores](../../../index.md)/[com.github.jonathanxd.kores](../index.md)/[MutableInstructions](index.md)/[MutableInstructions](-mutable-instructions.md)
# MutableInstructions
[jvm]\
fun [MutableInstructions](-mutable-instructions.md)()
| 33.857143 | 151 | 0.734177 | kor_Hang | 0.255148 |
912dd20adea8785d1f902da7a972d6c78d75bfe8 | 2,966 | md | Markdown | docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md | cawrites/sql-docs | 58158eda0aa0d7f87f9d958ae349a14c0ba8a209 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2020-05-07T19:40:49.000Z | 2020-09-19T00:57:12.000Z | docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md | cawrites/sql-docs | 58158eda0aa0d7f87f9d958ae349a14c0ba8a209 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/2014/relational-databases/lesson-2-create-a-sql-server-credential-using-a-shared-access-signature.md | cawrites/sql-docs | 58158eda0aa0d7f87f9d958ae349a14c0ba8a209 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2020-03-11T20:30:39.000Z | 2020-05-07T19:40:49.000Z | ---
title: "Lesson 3: Create a SQL Server Credential | Microsoft Docs"
ms.custom: ""
ms.date: "06/13/2017"
ms.prod: "sql-server-2014"
ms.reviewer: ""
ms.technology: "database-engine"
ms.topic: conceptual
ms.assetid: 29e57ebd-828f-4dff-b473-c10ab0b1c597
author: MikeRayMSFT
ms.author: mikeray
manager: craigg
---
# Lesson 3: Create a SQL Server Credential
In this lesson, you will create a credential to store security information used to access the Azure storage account.
A SQL Server credential is an object that is used to store authentication information required to connect to a resource outside of SQL Server. The credential stores the URI path of the storage container and the shared access signature key values. For each storage container used by a data or log file, you must create a SQL Server Credential whose name matches the container path.
For general information about credentials, see [Credentials (Database Engine)](security/authentication-access/credentials-database-engine.md).
> [!IMPORTANT]
> The requirements for creating a SQL Server credential described below are specific to the [SQL Server Data Files in Azure](databases/sql-server-data-files-in-microsoft-azure.md) feature. For information on creating credentials for backup processes in Azure storage, see [Lesson 2: Create a SQL Server Credential](../tutorials/lesson-2-create-a-sql-server-credential.md).
To create a SQL Server Credential, follow these steps:
1. Connect to SQL Server Management Studio.
2. In Object Explorer, connect to the instance of Database Engine installed.
3. On the Standard tool bar, click New Query.
4. Copy and paste the following example into the query window, modify as needed. The following statement will create a SQL Server Credential to store your storage container's Shared Access Certificate.
```sql
USE master
CREATE CREDENTIAL credentialname - this name should match the container path and it must start with https.
WITH IDENTITY='SHARED ACCESS SIGNATURE', -- this is a mandatory string and do not change it.
SECRET = 'sharedaccesssignature' -- this is the shared access signature key that you obtained in Lesson 2.
GO
```
For detailed information, see [CREATE CREDENTIAL (Transact-SQL)](/sql/t-sql/statements/create-credential-transact-sql) in SQL Server Books Online.
5. To see all available credentials, you can run the following statement in the query window:
```sql
SELECT * from sys.credentials
```
For more information on sys.credentials, see [sys.credentials (Transact-SQL)](/sql/relational-databases/system-catalog-views/sys-credentials-transact-sql) in SQL Server Books Online.
**Next Lesson:**
[Lesson 4: Create a database in Azure Storage](lesson-3-database-backup-to-url.md)
| 50.271186 | 384 | 0.726231 | eng_Latn | 0.964667 |
912e13cd99b8478e64737398e0079e8381342732 | 622 | md | Markdown | src/pages/companies/2018-02-13-supermeat.md | arvenjadeaguilar/cellagri-cms | f177977e4d859f540ed1fc455594629b30f42bd2 | [
"MIT"
] | null | null | null | src/pages/companies/2018-02-13-supermeat.md | arvenjadeaguilar/cellagri-cms | f177977e4d859f540ed1fc455594629b30f42bd2 | [
"MIT"
] | null | null | null | src/pages/companies/2018-02-13-supermeat.md | arvenjadeaguilar/cellagri-cms | f177977e4d859f540ed1fc455594629b30f42bd2 | [
"MIT"
] | null | null | null | ---
templateKey: company-post
path: /supermeat
date: '2018-02-15T11:00:00-05:00'
title: SuperMeat
location: 'Tel Aviv, Israel'
website: supermeat.com
socialMedia:
- media: Twitter
url: 'https://twitter.com/_SuperMeat_'
logo: /img/supermeat logo.jpg
thumbnail: /img/supermeat logo.jpg
description: >-
SuperMeat is a cellular agriculture startup based in Tel Aviv, Israel.
SuperMeat was founded by Ido Savir, Koby Barak and Shir Friedman in 2015 to
make cultured chicken meat. Through cellular agriculture, SuperMeat hopes to
provides a sustainable and healthy source of chicken meat for processed foods.
---
| 31.1 | 80 | 0.765273 | eng_Latn | 0.809162 |
912e708b1c2bf65f27991b730490981a9fed93a4 | 1,352 | md | Markdown | catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md | htron-dev/baka-db | cb6e907a5c53113275da271631698cd3b35c9589 | [
"MIT"
] | 3 | 2021-08-12T20:02:29.000Z | 2021-09-05T05:03:32.000Z | catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md | zzhenryquezz/baka-db | da8f54a87191a53a7fca54b0775b3c00f99d2531 | [
"MIT"
] | 8 | 2021-07-20T00:44:48.000Z | 2021-09-22T18:44:04.000Z | catalog/boku-no-mama-chan-43-kaihatsu-nikki/en-US_boku-no-mama-chan-43-kaihatsu-nikki.md | zzhenryquezz/baka-db | da8f54a87191a53a7fca54b0775b3c00f99d2531 | [
"MIT"
] | 2 | 2021-07-19T01:38:25.000Z | 2021-07-29T08:10:29.000Z | # Boku no Mama-chan (43) Kaihatsu Nikki

- **type**: manga
- **volumes**: 1
- **chapters**: 6
- **original-name**: 僕のママちゃん(43)開発日記
- **start-date**: 2016-10-19
## Tags
- yaoi
## Authors
- Neko Hakushaku (Story & Art)
## Sinopse
When seasoned actor Mamoru "Mama-chan" Maeda first stepped onto the set of his new crime drama, Dark Panther, he thought the only part he would have to play was that of experienced detective Inspector Kinoda. However, his rather persistent co-star, Christopher Shirasu, has other ideas.
Shirasu is an avid follower of fan culture, spending hours scrolling through his social media tags, looking at art created by his fans, and obsessing over reading fanfiction. After spending so much time engaging with the desires of the drama's most enthusiastic demographic, Shirasu has become rather strict with Maeda about his expectations—he needs to give fans what they want.
Ensuring that the two toe the perfect line between close friendship and homosexual romance, Shirasu is determined to make sure that the hashtag "#shiramama" will continue to trend—regardless of any "ship wars."
[Source My Anime List]
## Links
- [My Anime list](https://myanimelist.net/manga/103294/Boku_no_Mama-chan_43_Kaihatsu_Nikki)
| 42.25 | 379 | 0.756657 | eng_Latn | 0.99339 |
912ea325833ebaae5e7ab8b9a6dacec03587355c | 813 | md | Markdown | README.md | Phocacius/kmeans | 3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d | [
"MIT"
] | null | null | null | README.md | Phocacius/kmeans | 3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d | [
"MIT"
] | null | null | null | README.md | Phocacius/kmeans | 3bd1ff32a1cd37988d7f2d94acd1213c5ed9e74d | [
"MIT"
] | null | null | null | # K-Means Simulator
Provides a step-by-step visualisation of the k-means algorithm for unsupervised clustering of 2D Data. Click on the canvas to add points and choose your desired number of clusters (k). You can then run the algorithm step-by-step manually using "Assign Clusters" and "Recenter Centroids" or automatically where there's a half second delay between each step using "Autostart". "Place centroids randomly" resets the clusters and centroids.

## Required Software for Development
The javascript files are compiled using browserify.
- [NodeJS](https://nodejs.org/en/download/)
- [Browserify](http://browserify.org/): `npm install -g browserify`
## Building
- install dependencies (currently only jquery): `npm install`
- compile JS: `browserify js/main.js -o js/bundle.js`
| 45.166667 | 436 | 0.771218 | eng_Latn | 0.968846 |
912f240e2130569d98515738df087ed19fd5f0ee | 476 | md | Markdown | guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md | SweeneyNew/freeCodeCamp | e24b995d3d6a2829701de7ac2225d72f3a954b40 | [
"BSD-3-Clause"
] | 10 | 2019-08-09T19:58:19.000Z | 2019-08-11T20:57:44.000Z | guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md | SweeneyNew/freeCodeCamp | e24b995d3d6a2829701de7ac2225d72f3a954b40 | [
"BSD-3-Clause"
] | 2,056 | 2019-08-25T19:29:20.000Z | 2022-02-13T22:13:01.000Z | guide/arabic/certifications/javascript-algorithms-and-data-structures/es6/use-destructuring-assignment-to-assign-variables-from-arrays/index.md | SweeneyNew/freeCodeCamp | e24b995d3d6a2829701de7ac2225d72f3a954b40 | [
"BSD-3-Clause"
] | 5 | 2018-10-18T02:02:23.000Z | 2020-08-25T00:32:41.000Z | ---
title: Use Destructuring Assignment to Assign Variables from Arrays
localeTitle: استخدم Destructuring Assignment لتعيين متغيرات من صفائف
---
## استخدم Destructuring Assignment لتعيين متغيرات من صفائف
علينا اتخاذ بعض الاحتياطات في هذه الحالة.
1. لا حاجة للثابتة \[ب ، أ\] لأنها ستحافظ على تأثير الواجب المحلي.
2. const \[b، a\] = \[a، b\] سيؤدي إلى قيمة a ، مثل b غير محددة (قاعدة تخصيص بسيطة من اليسار إلى اليمين).
ومن هنا حل هذه المشكلة \[b، a\] = \[a، b\] | 34 | 106 | 0.707983 | arb_Arab | 0.997059 |
912f4e0ad3415be2311ea7f9f504919fd6bc3f70 | 2,554 | md | Markdown | README.md | breglerj/cloud-foundry-tools-api | 883de7da0c233fabc366c86b62c72a2720139295 | [
"Apache-2.0"
] | null | null | null | README.md | breglerj/cloud-foundry-tools-api | 883de7da0c233fabc366c86b62c72a2720139295 | [
"Apache-2.0"
] | null | null | null | README.md | breglerj/cloud-foundry-tools-api | 883de7da0c233fabc366c86b62c72a2720139295 | [
"Apache-2.0"
] | null | null | null | 
[](https://circleci.com/gh/SAP/cloud-foundry-tools-api)
[](https://coveralls.io/github/SAP/cloud-foundry-tools-api?branch=master)
[](https://lgtm.com/projects/g/SAP/cloud-foundry-tools-api/context:javascript)
[](http://commitizen.github.io/cz-cli/)
[](https://api.reuse.software/info/github.com/SAP/cloud-foundry-tools-api)
[](https://dependabot.com/)
# Overview
This package provides a set of APIs to help you develop applications in Cloud Foundry. You can use these APIs to manage apps, service instances, orgs, spaces, and users in your environment. Mostly, this is a wrapper of the CF command line client, which runs a particular command and parses the output to the suitable JSON file. If an error or failure occurs, the runtime exception throws with relevant problem information.
## Prerequisite
_from version_ `v1.0.0` _and above_:
Make sure you have installed the CF CLI [v7](https://github.com/cloudfoundry/cli/blob/master/doc/installation-instructions/installation-instructions-v7.md#installers-and-compressed-binaries) tool in your environment.
_below version_ `v1.0.0`:
Make sure you have installed the CF CLI [v6](https://github.com/cloudfoundry/cli/blob/master/doc/installation-instructions/installation-instructions-v6.md#installers-and-compressed-binaries) tool in your environment.
## Examples of usage
Example 1
```
try {
const result = await cfLogin("https://api.cf.....com", "user", "password");
if (result === "OK") {
// successful
}
} catch (e) {
// display or/and log error
}
```
Example 2
```
try {
const spaces = await cfGetAvailableSpaces("myOrg");
for (const space of spaces) {
console.log("Space label is " + space.label + " guid is " + space.guid);
}
} catch (e) {
// display or/and log error
}
```
## Contributing
Contributing information can be found in the [CONTRIBUTING.md](CONTRIBUTING.md) file.
| 49.115385 | 423 | 0.755286 | eng_Latn | 0.619575 |
912fa5fd54b2114e2cd8ca1fdf098ceebabbb7ad | 6,151 | md | Markdown | README.md | DazaSeal/wrapper-infinite | 7079f33791246d023ac204a42f89f1b9d42b39fc | [
"MIT"
] | null | null | null | README.md | DazaSeal/wrapper-infinite | 7079f33791246d023ac204a42f89f1b9d42b39fc | [
"MIT"
] | null | null | null | README.md | DazaSeal/wrapper-infinite | 7079f33791246d023ac204a42f89f1b9d42b39fc | [
"MIT"
] | 1 | 2022-03-12T15:00:11.000Z | 2022-03-12T15:00:11.000Z | ## Wrapper: Infinite
Wrapper: Infinite is a project **(unofficially)** carrying on the torch of [VisualPlugin's Wrapper project](https://github.com/GoAnimate-Wrapper) after it's shutdown in 2020. Unlike the original project, Infinite can not be shut down by Vyond. Why? It's because of our twist on the Wrapper formula! Everything is hosted and programmed to function locally, no internet access required, barring the initial download (and TTS at this time). This project is important for archival purposes, as the ability to use this legacy editor and themes would be irritating and unproductive without it. Besides the main feature of Infinite, we have also added other additional features and improvements compared to the original Wrapper project, and there will be more to come, as Infinite will continue to get additional updates in the future.
## Running / Installation
To start Infinite on Windows, open start_wrapper.bat. It'll automate just about everything for you and, well, start Wrapper: Infinite. On your first run, you will likely need to right-click it and click "Run as Administrator". This allows it to properly install what it needs to run. After your initial run, you shouldn't need to do that again, you can start it as normal.
If you want to import videos and characters from the original Wrapper or any other clones of it, open its folder and drag the "_SAVED" folder into Infinite's "wrapper" folder. If you have already made any videos or characters, this will not work. Please only import on a new install with no saved characters or videos, or take the "_SAVED" folder in Infinite out before dragging the old one in. If you want to import character IDs from the original LVM, you can paste `&original_asset_id=[ID HERE]` at the end of the link for the matching character creator.
## Updates & Support
### This information is outdated and does not pertain to Wrapper: Infinite.
For support, the first thing you should do is read through faq.txt, it most likely has what you want to know. If you can't find what you need, you can join the [Discord server](https://discord.gg/Kf7BzSw). Joining the server is recommended, as there is a whole community to help you out. If you don't use Discord, you can email mailbenson@protonmail.com to get in contact with me directly, but don't expect nearly as quick of a response.
## Dependencies
This program relies on Flash and Node.js to work properly. SilentCMD is also used to suppress all the extra logging noise you'd only need for troubleshooting and development. These all have been included with the project (utilities folder) to ensure full offline operation and will be installed if missing. The "wrapper" folder has it's own dependencies, but they are included as well.
## License
Most of this project is free/libre software[1] under the MIT license. You have the freedom to run, change, and share this as much as you want.
This includes:
- Files in the "wrapper" folder
- Batch files included with Wrapper: Infinite
- Node.js
- SilentCMD
- Chromium Web Store
ungoogled-chromium is under the BSD 3-Clause license, which grants similar rights, but has some differences from MIT. MediaInfo has a similar BSD 2-Clause license. 7zip's license is mostly LGPL, but some parts are under the BSD 3-clause License, and some parts have an unRAR restriction. Stylus is under the GNU GPLv3 license. These licenses can be found in each program's folder in utilities\sourcecode.
The source code for compiled programs are all stored in utilities\sourcecode, and you can modify these as you wish. Parts of Infinite that run from their source code directly (such as batch scripts) are not included in that folder, for obvious reasons.
Flash Player (utilities folder) and GoAnimate's original assets (server folder) are proprietary and do not grant you these rights, but if they did, this project wouldn't need to exist. Requestly, an addon included in Infinite's browser, is sadly proprietary software, but you're free to remove the Chromium profile and use a fresh one if this bothers you. Requestly is primarily included because of how popular it is with our community.
While completely unnecessary, if you decide to use your freedom to change the software, it would be greatly appreciated if you sent it to me so I can implement it into the main program! With credit down here of course :)
## Credits
**Please do not contact anyone on the list for support, use the Discord server.**
Original Wrapper credits:
| Name | Contribution |
| ------------ | -------------------- |
| VisualPlugin | GoAnimate Wrapper |
| xomdjl_ | Custom/Modded Assets |
| CLarramore | Bug fixes |
| PoleyMagik | Asset Store Archive |
No members of the original team are officially working on Infinite, even if they have contributed. Some members of the original team have asked to not be given credit, and they have been removed.
Developers will have (D) on their name, to show they are officially working on W:I. Everyone else has simply contributed something at some point, or left the development team.
Wrapper: Infinite credits:
| Name | Contribution |
| ------------- | -------------------- |
| Octanuary | Wrapper: Infinite |
| Daza (D) | Improved CSS |
These are unaffiliated people that they haven't directly done anything for the project (and probably don't even know it exists) but still deserve credit for their things. Kinda like a shoutout but in a project's readme. ***Please do not contact them about Wrapper: Infinite.***
| Name | Contribution |
| ------------------ | -------------------------------- |
| Vyond | Creators of the themes we love |
| Stephan Brenner | Creator of SilentCMD |
| vocatus | Some of TronScript's batch code |
| ss64.com | Incredible CMD info resource |
| robvanderwoude.com | Also amazing CMD info resource |
| VisualPlugin | Text to speech voices |
## Footnotes
[1] - See <https://www.gnu.org/philosophy/free-sw.html> for a better definition of free software.
| 93.19697 | 828 | 0.738579 | eng_Latn | 0.999434 |
912fe6ae1c7ab206900cc994d8593fa76c63f95f | 1,547 | md | Markdown | results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md | eliMakeouthill/AutoEq | b16c72495b3ce493293c6a4a4fdf45a81aec9ca0 | [
"MIT"
] | 3 | 2022-02-25T08:33:08.000Z | 2022-03-13T11:27:29.000Z | results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md | billclintonwong/AutoEq | aa25ed8e8270c523893fadbda57e9811c65733f1 | [
"MIT"
] | null | null | null | results/referenceaudioanalyzer/referenceaudioanalyzer_siec_harman_in-ear_2019v2/DUNU DN16 Hephaes/README.md | billclintonwong/AutoEq | aa25ed8e8270c523893fadbda57e9811c65733f1 | [
"MIT"
] | null | null | null | # DUNU DN16 Hephaes
See [usage instructions](https://github.com/jaakkopasanen/AutoEq#usage) for more options and info.
### Parametric EQs
In case of using parametric equalizer, apply preamp of **-7.5dB** and build filters manually
with these parameters. The first 5 filters can be used independently.
When using independent subset of filters, apply preamp of **-7.5dB**.
| Type | Fc | Q | Gain |
|:--------|:---------|:-----|:--------|
| Peaking | 26 Hz | 0.14 | -7.8 dB |
| Peaking | 289 Hz | 0.41 | -9.7 dB |
| Peaking | 1149 Hz | 0.26 | 9.0 dB |
| Peaking | 5784 Hz | 1.05 | -6.7 dB |
| Peaking | 11495 Hz | 0.81 | 7.3 dB |
| Peaking | 3005 Hz | 4.4 | 2.2 dB |
| Peaking | 3494 Hz | 3.64 | -1.8 dB |
| Peaking | 9276 Hz | 6.08 | 1.3 dB |
| Peaking | 15919 Hz | 1.97 | 4.7 dB |
| Peaking | 20191 Hz | 0.63 | -9.7 dB |
### Fixed Band EQs
In case of using fixed band (also called graphic) equalizer, apply preamp of **-7.9dB**
(if available) and set gains manually with these parameters.
| Type | Fc | Q | Gain |
|:--------|:---------|:-----|:--------|
| Peaking | 31 Hz | 1.41 | -8.4 dB |
| Peaking | 62 Hz | 1.41 | -6.6 dB |
| Peaking | 125 Hz | 1.41 | -7.9 dB |
| Peaking | 250 Hz | 1.41 | -7.6 dB |
| Peaking | 500 Hz | 1.41 | -1.2 dB |
| Peaking | 1000 Hz | 1.41 | 5.0 dB |
| Peaking | 2000 Hz | 1.41 | 6.6 dB |
| Peaking | 4000 Hz | 1.41 | -0.6 dB |
| Peaking | 8000 Hz | 1.41 | 2.4 dB |
| Peaking | 16000 Hz | 1.41 | 6.0 dB |
### Graphs
 | 38.675 | 98 | 0.553975 | eng_Latn | 0.746417 |
912ff8cf5d480d6cfe9dc42d772158a1b1b9c609 | 513 | md | Markdown | _course_files/Chapter 6 Services/Step.md | boriphuth/k8s-fleetman | 5bc31a72d7fadd7f0550d0390fcab9525055663b | [
"MIT"
] | 1 | 2019-10-23T09:14:35.000Z | 2019-10-23T09:14:35.000Z | _course_files/Chapter 6 Services/Step.md | boriphuth/k8s-fleetman | 5bc31a72d7fadd7f0550d0390fcab9525055663b | [
"MIT"
] | null | null | null | _course_files/Chapter 6 Services/Step.md | boriphuth/k8s-fleetman | 5bc31a72d7fadd7f0550d0390fcab9525055663b | [
"MIT"
] | null | null | null | $ kubectl apply -f first-pod.yaml
$ kubectl apply -f webapp-service.yaml
## Connect to Service
$ minikube service fleetman-webapp
$ kubectl get po --show-labels
NAME READY STATUS RESTARTS AGE LABELS
webapp 1/1 Running 0 92m app=webapp,release=0
webapp-release-0-5 1/1 Running 0 48m app=webapp,release=0-5
$ kubectl get po --show-labels -l release=0
$ kubectl get po --show-labels -l release=1
No resources found in default namespace. | 36.642857 | 78 | 0.647173 | yue_Hant | 0.764128 |
9130714c260b6a6b44a06f07c4f320519420ba98 | 5,310 | md | Markdown | articles/cognitive-services/Content-Moderator/review-api.md | changeworld/azure-docs.nl-nl | bdaa9c94e3a164b14a5d4b985a519e8ae95248d5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/cognitive-services/Content-Moderator/review-api.md | changeworld/azure-docs.nl-nl | bdaa9c94e3a164b14a5d4b985a519e8ae95248d5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/cognitive-services/Content-Moderator/review-api.md | changeworld/azure-docs.nl-nl | bdaa9c94e3a164b14a5d4b985a519e8ae95248d5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Recensies, werk stromen en taken concepten-Content Moderator
titleSuffix: Azure Cognitive Services
description: In dit artikel vindt u meer informatie over de basis concepten van het hulp programma voor beoordeling. Beoordelingen, werk stromen en taken.
services: cognitive-services
author: PatrickFarley
manager: nitinme
ms.service: cognitive-services
ms.subservice: content-moderator
ms.topic: conceptual
ms.date: 03/14/2019
ms.author: pafarley
ms.openlocfilehash: 1aba86efb9ea76fbf060e80b47f9f2f6cdf8ee71
ms.sourcegitcommit: 7b25c9981b52c385af77feb022825c1be6ff55bf
ms.translationtype: MT
ms.contentlocale: nl-NL
ms.lasthandoff: 03/13/2020
ms.locfileid: "79221148"
---
# <a name="content-moderation-reviews-workflows-and-jobs"></a>Beoordelingen, werk stromen en taken voor content toezicht
Content Moderator combineert de door de machine ondersteunde toezicht met de mogelijkheden van Human-in-the-loop om een optimaal toezicht proces te maken voor praktijk scenario's. Dit wordt gedaan via het [beoordelings programma](https://contentmoderator.cognitive.microsoft.com)op basis van de Cloud. In deze hand leiding vindt u informatie over de belangrijkste concepten van het hulp programma voor beoordeling: Recensies, werk stromen en taken.
## <a name="reviews"></a>Beoordelingen
Bij een beoordeling wordt inhoud geüpload naar het hulp programma beoordeling en wordt deze weer gegeven onder het tabblad **controleren** . Hier kunnen gebruikers de toegepaste labels wijzigen en hun eigen aangepaste labels Toep assen. Wanneer een gebruiker een beoordeling verzendt, worden de resultaten verzonden naar een opgegeven eind punt van de retour aanroep en wordt de inhoud verwijderd van de site.

Raadpleeg de [hand leiding voor het beoordelings programma](./review-tool-user-guide/review-moderated-images.md) om te beginnen met het maken van recensies of raadpleeg de [rest API gids](./try-review-api-review.md) voor informatie over hoe u dit kunt doen.
## <a name="workflows"></a>Werkstromen
Een werk stroom is een op de cloud gebaseerd gepersonaliseerd filter voor inhoud. Werk stromen kunnen verbinding maken met verschillende services om inhoud op verschillende manieren te filteren en vervolgens de juiste actie ondernemen. Met de Content Moderator-connector kunt u met een werk stroom automatisch toezicht Tags Toep assen en beoordelingen maken met verzonden inhoud.
### <a name="view-workflows"></a>Werk stromen weer geven
Als u uw bestaande werk stromen wilt weer geven, gaat u naar het [hulp programma controleren](https://contentmoderator.cognitive.microsoft.com/) en selecteert u **instellingen** > **werk stromen**.

Werk stromen kunnen volledig worden beschreven als JSON-teken reeksen, waardoor ze via een programma toegankelijk zijn. Als u de optie **bewerken** selecteert voor uw werk stroom en vervolgens het tabblad **JSON** selecteert, ziet u een JSON-expressie zoals het volgende:
```json
{
"Type": "Logic",
"If": {
"ConnectorName": "moderator",
"OutputName": "isAdult",
"Operator": "eq",
"Value": "true",
"Type": "Condition"
},
"Then": {
"Perform": [
{
"Name": "createreview",
"CallbackEndpoint": null,
"Tags": []
}
],
"Type": "Actions"
}
}
```
Raadpleeg de [hand leiding voor het hulp programma](./review-tool-user-guide/workflows.md) voor het maken en gebruiken van werk stromen of raadpleeg de [rest API gids](./try-review-api-workflow.md) voor informatie over hoe u dit kunt doen met programma code.
## <a name="jobs"></a>Taken
Een toezicht taak fungeert als een soort wrapper voor de functionaliteit van toezicht op inhoud, werk stromen en Beoordelingen. De taak scant uw inhoud met behulp van de Content Moderator afbeeldings toezicht-API of tekst toezicht-API en controleert deze vervolgens op de aangewezen werk stroom. Op basis van de werk stroom resultaten kan de inhoud niet worden gecontroleerd in het [beoordelings programma](./review-tool-user-guide/human-in-the-loop.md). Hoewel zowel beoordelingen als werk stromen kunnen worden gemaakt en geconfigureerd met hun respectieve Api's, kunt u met de taak-API een gedetailleerd rapport verkrijgen van het hele proces (dat kan worden verzonden naar een opgegeven call back-eind punt).
Raadpleeg de [rest API-hand leiding](./try-review-api-job.md) om aan de slag te gaan met taken.
## <a name="next-steps"></a>Volgende stappen
* Test de [taak-API-console](try-review-api-job.md)en gebruik de rest API code voorbeelden. Als u bekend bent met Visual Studio en C#raadpleegt u ook de [taken .net Quick](moderation-jobs-quickstart-dotnet.md)start.
* Ga voor recensies naar aan de slag met de [API-console voor controle](try-review-api-review.md)en gebruik de rest API code voorbeelden. Zie de sectie Recensies van [.net Quick](dotnet-sdk-quickstart.md)start.
* Gebruik voor video revisies de [Snelstartgids voor video revisie](video-reviews-quickstart-dotnet.md)en leer hoe u [Transcripten kunt toevoegen aan de video beoordeling](video-transcript-reviews-quickstart-dotnet.md).
| 66.375 | 712 | 0.775141 | nld_Latn | 0.999047 |
91308e7fdc37435df478c923542ea311fc37b53f | 7,397 | md | Markdown | business-central/readiness/readiness-learning-sales.md | MicrosoftDocs/dynamics365smb-docs-pr.nb-no | f57ffe1865b515a2240b7e4d1401263a33d2a535 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2020-05-18T17:20:08.000Z | 2021-04-20T21:13:47.000Z | business-central/readiness/readiness-learning-sales.md | MicrosoftDocs/dynamics365smb-docs-pr.nb-no | f57ffe1865b515a2240b7e4d1401263a33d2a535 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | business-central/readiness/readiness-learning-sales.md | MicrosoftDocs/dynamics365smb-docs-pr.nb-no | f57ffe1865b515a2240b7e4d1401263a33d2a535 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2019-10-12T19:50:37.000Z | 2020-09-30T16:51:21.000Z | ---
title: Læringskatalog for salg og markedsføring for partner
description: Finn alle tilgjengelige læringsressurser for salg og markedsføringsroller for partner i Business Central.
author: loreleishannonmsft
ms.date: 04/01/2021
ms.topic: conceptual
ms.author: margoc
ms.openlocfilehash: 9830e4e842cc7fe3febcbe809547ad26701f527d
ms.sourcegitcommit: 766e2840fd16efb901d211d7fa64d96766ac99d9
ms.translationtype: HT
ms.contentlocale: nb-NO
ms.lasthandoff: 03/31/2021
ms.locfileid: "5781841"
---
# <a name="partner-sales-and-marketing-learning-catalog"></a>Læringskatalog for salg og markedsføring for partner
Er du ansvarlig for å hjelpe kundene med å kjøpe riktig løsning?
Følgende katalog er ordnet fra kjernekunnskap til bestemte domener, og fra de mest grunnleggende til de fleste avanserte. Hvis innholdet finnes i flere formater, gir vi deg beskjed slik at du kan velge det opplæringsformatet som best dekker dine behov.
## <a name="business-value"></a>Forretningsverdi<a name="busvalue"></a>
| Innhold | Beskrivelse | Format | Lengde |
|-------------|----------------|---------------------------------------|---------------------|
| [Dynamics 365 Business Central (Blitz)](https://mbspartner.microsoft.com/D365/Videos/101760) | (Bare partnere) Denne økten vil lage en disposisjon av verdien for Business Central. Business Central tilbyr god verdi fordi den er integrert med andre Microsoft-skytjenester, inkludert Microsoft 365, og kan tilpasses eller utvides for bestemte bransjebehov med Power Apps, Power Automate og Power BI. | Video | 8 minutter 5 sekunder |
| [Utforske Dynamics 365 Business Central-funksjoner](https://dynamics.microsoft.com/business-central/capabilities/) | Utforske Dynamics 365 Business Central-funksjoner | Nettsted | |
| [Dynamics 365 Business Central](/learn/modules/dynamics-365-business-central/) | Denne modulen viser programfunksjoner, fordeler og eksempler på Dynamics 365 Business Central i bruk. | Gratis, elektronisk opplæringsbane i eget tempo | 24 minutter |
## <a name="getting-started"></a>Komme i gang<a name="get-started"></a>
| Innhold | Beskrivelse | Format | Lengde |
|-----------|---------------|---------------------------------------|---------------------|
| [Kom i gang med Microsoft Dynamics 365 Business Central](/learn/paths/get-started-dynamics-365-business-central/) | Er du interessert i Business Central? Hvis dette er tilfellet, får du en god innføring i denne opplæringsbanen. Den drøfter hvordan du konfigurerer en prøveversjon, gir grunnleggende informasjon om hva som er tilgjengelig, og viser noen av tilpasningsalternativene du kan bruke for å tilpasse Business Central for egen bruk. | Gratis, elektronisk opplæringsbane i eget tempo | 3 timer 4 minutter |
| [Arbeide med Microsoft Dynamics 365 Business Central-brukergrensesnittet](/learn/paths/work-with-user-interface-dynamics-365-business-central/) | Vil du lære om brukergrensesnittet for Business Central? Denne opplæringsbanen viser brukergrensesnittet, og også hvordan du kan tilpasse det. Det drøfter også alternative grensesnitt, for eksempel en smarttelefon. | Gratis, elektronisk opplæringsbane i eget tempo | 2 timer 27 minutter |
| [Arbeid som en profesjonell med data i Business Central](/learn/paths/work-pro-data-dynamics-365-business-central) | Har du nettopp kommet i gang med Business Central? Denne opplæringsbanen gjør det enklere for deg å forberede deg til datatypene i Business Central og finne informasjon du trenger i programmet. | Gratis, elektronisk opplæringsbane i eget tempo | 2 timer 27 minutter |
| [Business Central YouTube-spilliste](https://www.youtube.com/playlist?list=PLcakwueIHoT-wVFPKUtmxlqcG1kJ0oqq4) | YouTube-målside for Dynamics 365 Business Central-videoer | YouTube-videoer | |
## <a name="grow-your-business"></a>Utvikle bedriften<a name="grow"></a>
| Innhold | Beskrivelse | Format | Lengde |
|-----------|--------------------|---------------------------------------|-----------------------|
| [Dynamics 365 Business Central - Overbevisende samtaler med kundeemner (Blitz)](https://mbspartner.microsoft.com/D365/Videos/101761) | (Bare partnere) Bli med Errol Schoenfish, direktøren for produktmarkedsføring, mens han snakker om hvordan du kan ha overbevisende samtaler med kundeemner på Microsoft Dynamics 365 Business Central. | Video | 11 minutter 15 sekunder |
| [Program for leverandør av skytjenester](https://partner.microsoft.com/membership/cloud-solution-provider) | (Bare partnere) Om program for leverandør av skyløsninger | Nettsted | |
| [Markedsføring i en skyverden](https://mbspartner.microsoft.com/OSS/Topic/19) | (Bare partner) Hvordan gjøre det mulig å optimalisere markedsføringspraksisen og justere markedsføringsprinsippene for en skytjeneste er en primær faktor for å finne ut hvor vellykket eller mislykket organisasjonen går fremover. Denne serien med videoer og ressurser inneholder gode fremgangsmåter for markedsføring, verktøy for oppretting av innhold, nyttige tips og verktøy for å bruke prinsipper for markedsføring i skyen til å skape økt etterspørsel etter løsningene dine. | Emneside | 35 minutter |
| [Program for leverandør av skyløsninger – selge etterspurte skyløsninger](/partner-center/csp-overview) | Programmet for leverandør av skyløsninger (CSP) hjelper deg med å gå lenger enn å videreselge lisenser til å være mer involvert i kundens virksomhet. | Nettsted | |
| [Program for leverandør av skyløsninger - sammenligne partnerstøtteprogrammer](https://partner.microsoft.com/support/partnersupport) | Se denne videoen for å se hvordan du kan velge den riktige planen for partnerstøtte for bedriften. Deretter ser du en side ved side-sammenligning nedenfor. | Nettsted | |
| [Cloud SureStep-akselerert salg](https://mbspartner.microsoft.com/OSS/Topic/20) | (Bare partnere) Prosessen Cloud SureStep SureStep-akselerert salg er bygget rundt behovene til moderne kjøpere og søker å utløse en emosjonell, erfaringsdrevet strategisk vendepunkt tidlig i salgsprosessen. Denne serien med videoer og ressurser forklarer en normativ salgsprosess som er spesielt utformet for å motvirke den tradisjonelle ressursintensive, demoorienterte tilnærmingen til salg som vanligvis fører til svært tilpassede produktdemonstrasjoner, komplekse salgssykluser og lange implementeringsprosjekter. Ressursblioteket omfatter en rekke tilpassbare maler, slik at du kan opprette ditt eget sett med salgsressurser for å støtte akselerert salg. | Nettsted | |
| [Tillegg til strategiplanen for overgangspartner (juli 2019)](https://assetsprod.microsoft.com/d365-business-central-supplement-to-transition-partner-playbook.pptx) | Dette tillegget til strategiplanen for overgangspartner gir oppdatert og ny veiledning for partnere for å sikre den eksisterende kundebasen mot konkurransen og nye SaaS-konkurrenter. | PowerPoint-lysbildefremvisning | 15 lysbilder |
[!INCLUDE[footer-include](../includes/footer-banner.md)] | 145.039216 | 785 | 0.723807 | nob_Latn | 0.982268 |
91317037239b34b820e8c646595604362f7e4e07 | 5,130 | md | Markdown | basics/plugin_structure/plugin_configuration_file.md | pettermahlen/intellij-sdk-docs | 1fd57016f2bf34afb3277d7094da091e0d84876b | [
"Apache-2.0"
] | 1 | 2021-08-18T09:44:04.000Z | 2021-08-18T09:44:04.000Z | basics/plugin_structure/plugin_configuration_file.md | pettermahlen/intellij-sdk-docs | 1fd57016f2bf34afb3277d7094da091e0d84876b | [
"Apache-2.0"
] | null | null | null | basics/plugin_structure/plugin_configuration_file.md | pettermahlen/intellij-sdk-docs | 1fd57016f2bf34afb3277d7094da091e0d84876b | [
"Apache-2.0"
] | null | null | null | ---
title: Plugin Configuration File - plugin.xml
---
The following is a sample plugin configuration file. This sample showcases and describes all elements that can be used in the plugin.xml file.
```xml
<!-- url="" specifies the URL of the plugin homepage (displayed in the Welcome Screen and in "Plugins" settings dialog) -->
<idea-plugin url="http://www.jetbrains.com/idea">
<!-- Plugin name -->
<name>VssIntegration</name>
<!-- Unique identifier of the plugin.
Cannot be changed between the plugin versions.
If not specified, assumed to be equal to <name>. -->
<id>VssIntegration</id>
<!-- Description of the plugin. -->
<description>Vss integration plugin</description>
<!-- Description of changes in the latest version of the plugin.
Displayed in the "Plugins" settings dialog and in the
plugin repository Web interface. -->
<change-notes>Initial release of the plugin.</change-notes>
<!-- Plugin version -->
<version>1.0</version>
<!-- The vendor of the plugin.
The optional "url" attribute specifies the URL of the vendor homepage.
The optional "email" attribute specifies the e-mail address of the vendor.-->
<vendor url="http://www.jetbrains.com" email="support@jetbrains.com" />
<!-- The unique identifiers of the plugins on which this plugin depends. -->
<depends>MyFirstPlugin</depends>
<!-- Optional dependency on another plugin.
If the plugin with the "MySecondPlugin" ID is installed,
the contents of mysecondplugin.xml (the format of this file
conforms to the format of plugin.xml) will be loaded. -->
<depends optional="true" config-file="mysecondplugin.xml">MySecondPlugin</depends>
<!-- Allows a plugin to integrate its help system (in JavaHelp format)
with the IDEA help system. The "file" attribute specifies the name
of the JAR file in the "help" subdirectory of the plugin directory.
The "path" attribute specifies the name of the helpset file within
the JAR file.-->
<helpset file="myhelp.jar" path="/Help.hs" />
<!-- Minimum and maximum build of IDEA compatible with the plugin -->
<idea-version since-build="3000" until-build="3999"/>
<!-- Resource bundle from which the text of plugin descriptions,
action names and etc. will be loaded -->
<resource-bundle>messages.MyPluginBundle</resource-bundle>
<!-- Plugin's application components -->
<application-components>
<component>
<!-- Component's interface class -->
<interface-class>com.foo.Component1Interface</interface-class>
<!-- Component's implementation class -->
<implementation-class>com.foo.impl.Component1Impl</implementation-class>
</component>
</application-components>
<!-- Plugin's project components -->
<project-components>
<component>
<!-- Interface and implementation classes are the same -->
<interface-class>com.foo.Component2</interface-class>
<!-- If the "workspace" option is set "true", the component
saves its state to the .iws file instead of the .ipr file.
Note that the <option> element is used only if the component
implements the JDOMExternalizable interface. Otherwise, the
use of the <option> element takes no effect. -->
<option name="workspace" value="true" />
<!-- If the "loadForDefaultProject" tag is present, the project component is instantiated also for the default project. -->
<loadForDefaultProject>
</component>
</project-components>
<!-- Plugin's module components -->
<module-components>
<component>
<interface-class>com.foo.Component3</interface-class>
</component>
</module-components>
<!-- Actions -->
<actions>
<action id="VssIntegration.GarbageCollection" class="com.foo.impl.CollectGarbage" text="Collect _Garbage" description="Run garbage collector">
<keyboard-shortcut first-keystroke="control alt G" second-keystroke="C" keymap="$default"/>
</action>
</actions>
<!-- Extension points defined by the plugin.
Extension points are registered by a plugin so that other
plugins can provide this plugin with certain data. The
"beanClass" attribute specifies the class the implementations
of which can be used for the extension point. -->
<extensionPoints>
<extensionPoint name="testExtensionPoint" beanClass="com.foo.impl.MyExtensionBean"/>
</extensionPoints>
<!-- Extensions which the plugin adds to extension points
defined by the IDEA core or by other plugins.
The "defaultExtensionNs " attribute must be set to the
ID of the plugin defining the extension point, or to
"com.intellij" if the extension point is defined by the
IDEA core. The name of the tag within the <extensions>
tag matches the name of the extension point, and the
"implementation" class specifies the name of the class
added to the extension point. -->
<extensions xmlns="VssIntegration">
<testExtensionPoint implementation="com.foo.impl.MyExtensionImpl"/>
</extensions>
</idea-plugin>
```
| 41.370968 | 146 | 0.696491 | eng_Latn | 0.96917 |
9131a9f6f961a049b6249b5cf547e4a13f841134 | 8,094 | markdown | Markdown | _posts/python/2020-08-13-virtual-environment.markdown | daesungRa/namu | 8a6e5b74a20189fb56d498155e81f55daeb03f52 | [
"MIT"
] | null | null | null | _posts/python/2020-08-13-virtual-environment.markdown | daesungRa/namu | 8a6e5b74a20189fb56d498155e81f55daeb03f52 | [
"MIT"
] | 78 | 2020-10-02T12:50:55.000Z | 2022-03-27T08:08:46.000Z | _posts/python/2020-08-13-virtual-environment.markdown | daesungRa/namu | 8a6e5b74a20189fb56d498155e81f55daeb03f52 | [
"MIT"
] | null | null | null | ---
title: Python 가상환경을 만드는 방법
date: 2020-08-13 20:45:48 +0900
author: namu
categories: python
permalink: "/python/:year/:month/:day/:title"
image: https://cdn.pixabay.com/photo/2017/07/31/14/56/wall-2558279_1280.jpg
image-view: true
image-author: StockSnap
image-source: https://pixabay.com/ko/users/stocksnap-894430/
---
---
[목차]
1. [들어가며](#들어가며)
2. [pip](#pip)
3. [virtualenv](#virtualenv)
4. [pipenv](#pipenv)
5. [gitignore 에 포함할 것들](#gitignore-에-포함할-것들)
[참조]
1. [python.org, "pip and virtual environments"](https://packaging.python.org/guides/installing-using-pip-and-virtual-environments/)
2. [velog.io](https://velog.io/@doondoony/pipenv-101)
---
<br>
### 들어가며
파이썬에서 **_가상환경 개념_**은 매우 중요하다.
여러 개의 프로젝트를 동시에 개발할 때 각각 고유한 파이썬 개발환경을 구축하고(의존성 포함)
상호간에 영향을 미치지 않도록 해야 하기 때문이다.
파이썬으로 개발한다면 언제나 가상환경부터 구축해야 한다는 점을 유념해야 한다.
먼저 os global 영역에 파이썬이 설치되어 있다고 가정하고 출발해보자.
python 3.x 버전이라면 무엇이든 가능하다.
### pip
pip 는 파이썬 패키지 매니저이다.
현재 사용하는 파이썬 버전에 의존적인 모든 패키지들을 [pypi.org](https://pypi.org/) 아카이브로부터 설치 및 업데이트 해준다.
os global 영역에 파이썬이 설치되어 있다면 기본적으로 pip 도 설치되어 있다.
- 파이썬과 pip 버전확인 및 pip 업그레이드
```text
# windows
> py --version
> py -m pip --version
> py -m pip install --upgrade pip
# linux
$ python --version
$ python -m pip --version
$ python -m pip install --upgrade pip
```
### virtualenv
virtualenv 모듈은 **_프로젝트 내에서 가상환경을 만들어 파이썬 패키지들을 관리_**해준다.
위의 global pip 를 활용해 virtualenv 를 설치한 후, 내 프로젝트를 위한 가상환경을 생성해 보자.
- virtualenv 설치 및 업그레이드
```text
# windows
> py -m pip install virtualenv
> py -m pip install --upgrade virtualenv # 만약 필요하다면 업그레이드!
# linux
$ python -m pip install virtualenv
$ python -m pip install --upgrade virtualenv # 리눅스도 마찬가지.
```
> tip 1
> > 당연하겠지만 환경변수에 global python 인터프리터를 등록해 두자. pip 까지 자유롭게 쓰게.
- 가상환경 생성 후 실행
<br>프로젝트 root 디렉토리로 이동한 후, 그곳에 가상환경을 만들자.
```text
# windows
> cd D:\project\root\
D:\project\root> py -m virtualenv venv
D:\project\root> call .\venv\Scripts\activate # 가상환경 실행
(venv) D:\project\root> where pip # 가상환경 내 pip 경로
# linux
$ cd /DATA/project/root/
/DATA/project/root$ python -m virtualenv venv
/DATA/project/root$ source ./venv/bin/activate # 가상환경 실행
(venv) /DATA/project/root$ which pip # 가상환경 내 pip 경로
```
- 가상환경 내에서 패키지 설치 > ```requests```
```text
# windows
(venv) D:\project\root> pip install requests # ... Successfully installed ...
(venv) D:\project\root> pip install requests==2.18.4 # 특정 버전 지정
# linux
(venv) /DATA/project/root$ pip install requests # ... Successfully installed ...
(venv) /DATA/project/root$ pip install requests==2.18.4 # 여기도.
```
이로써 가상환경 내 파이썬 인터프리터에 ```requests``` 패키지가 설치되었다. python import 를 통해 확인해 보자.
- 패키지 의존성 freezing! > requirements.txt 활용하기
<br>requirements.txt 는 현재 가상환경에 설치된 패키지들의 정적 의존성 정보를 텍스트 파일로 저장한다.
이렇게 해두면 나중에 일일히 설치할 필요가 없어진다.
```text
# windows
(venv) D:\project\root> pip freeze > requirements.txt # 이 파일이 생성된다.
# linux
(venv) /DATA/project/root$ pip freeze > requirements.txt
```
텍스트 파일을 열어보면 다음과 같다.
```text
asgiref==3.2.10
certifi==2020.6.20
chardet==3.0.4
Django==3.1
idna==2.10
pytz==2020.1
requests==2.24.0
sqlparse==0.3.1
urllib3==1.25.10
```
```requests``` 패키지가 버전 의존성까지 포함해서 기록되어 있다.
이제 이것을 활용해보면,
```text
# windows
(venv) D:\project\root> pip install -r requirements.txt # 위 정보를 기반으로 패키지 자동설치.
# linux
(venv) /DATA/project/root$ pip install -r requirements.txt
```
이제 어디서든 가상환경을 쉽게 구축할 수 있다!
- 가상환경 빠져나오기
```text
# windows
(venv) D:\project\root> deactivate
D:\project\root> # 빠져나왔다!
# linux
(venv) /DATA/project/root$ deactivate
/DATA/project/root$
```
### pipenv
이것도 가상환경을 구축하고 관리한다는 점에서 동일하다.<br>
하나 차이점은, ```pip + virtualenv == pipenv``` 라는 것!
> Tip 2
> > pipenv 는 더 편리하지만, 업데이트가 느리고 블랙박스 동작이 있다고 한다. 그래도 현업에 사용하기에 큰 무리는 없다.
> > 내가 그렇게 사용하고 있으니.. 그러므로 pipenv 를 사용하고자 한다면, 먼저 pip, virtualenv 사용법을 확실히 알자.
- 주요 명령어
<br>pip 와 virtualenv 가 합쳐진 만큼, pipenv 하나만으로 가상환경 생성과 가상환경 내 패키지 관리가 가능하다.
이를 위한 기본 명령어들을 살펴보자.
- pipenv --python 3.6 : 3.6 버전으로 가상환경 생성(단순 생성!).
- pipenv **shell** : 가상환경 쉘 진입. 없다면 생성 후 진입.
- pipenv **install** [package-name] : 패키지 설치. ```--dev``` 옵션을 붙이면 개발용으로 설치(Pipfile).
같은 방식으로 버전정보를 추가할 수 있다```~=1.2```.
- pipenv **lock** : 현재 버전 의존성 그대로 locking(Pipfile.lock).
이는 **결정론적 빌드**를 보장한다(어떤 상황에서도 같은 환경 보장).
- pipenv **sync** : locking 된 버전 의존성 그대로 패키지 설치(최신 버전이 아닐 수 있음).
- pipenv **update** : 설치된 패키지들을 최신 버전으로 업데이트. locking 이 자동으로 이루어진다.
업데이트를 원하는 특정 패키지 명시 가능.
shell 진입 시 ```Pipfile``` 및 ```Pipfile.lock``` 파일이 생성되며,
install, lock, sync, update 명령에 따라 패키지 및 패키지 의존성 정보가 기록된다.
Pipfile 예시는 다음과 같다.
```text
[[source]]
name = "pypi"
url = "https://pypi.org/simple"
verify_ssl = true
[dev-packages] # 개발용은 여기
django = "*"
[packages] # 배포용은 여기
requests = "*"
django = "*"
[requires] # 파이썬 버전 및 환경정보
python_version = "3.6"
[scripts] # scripts 섹션을 만들고
start = "python app.py" # pipenv run start
```
- pipenv 설치 및 업그레이드
<br>virtualenv 와 마찬가지로 일단은 global 영역에 설치 및 업그레이드한다.
```text
# windows
> py -m pip install pipenv
> py -m pip install --upgrade pipenv
# linux
$ python -m pip install pipenv
$ python -m pip install --upgrade pipenv
```
- 가상환경 쉘 진입 > 원하는 패키지 설치 > lock
```text
# windows
> cd D:\project\root\
D:\project\root> py -m pipenv shell # pipenv 쉘로 진입. 없다면 생성.
(root-x8NlcEbx) D:\project\root> pipenv install requests
(root-x8NlcEbx) D:\project\root> pipenv install django
(root-x8NlcEbx) D:\project\root> pipenv install django --dev
(root-x8NlcEbx) D:\project\root> pipenv lock
# linux
$ cd /DATA/project/root/
/DATA/project/root$ python -m pipenv shell
(root-x8NlcEbx) /DATA/project/root$ pipenv install requests
(root-x8NlcEbx) /DATA/project/root$ pipenv install django
(root-x8NlcEbx) /DATA/project/root$ pipenv install django --dev
(root-x8NlcEbx) /DATA/project/root$ pipenv lock
```
- 가상환경은 어디에 생성되나??
<br>다음을 실행해 보자.
```text
# windows
(root-x8NlcEbx) D:\project\root> pipenv --py
C:\Users\[USER]\.virtualenvs\root-x8NlcEbx\Scripts\python.exe
# linux
(root-x8NlcEbx) /DATA/project/root$ pipenv --py
$HOME/.local/share/virtualenvs/root-x8NlcEbx/bin/python
```
virtualenv 와 다르게(프로젝트 내 venv), pipenv 의 가상환경은 **현재 계정 home 의 숨겨진 디렉토리 내에 생성**된다.
그러므로 내가 실행하는 프로젝트는 그것이 연계된 home 하위 가상환경의 인터프리터를 활용한다고 생각하면 된다.
- 가상환경 삭제 및 빠져나오기
```text
# windows
(root-x8NlcEbx) D:\project\root> pipenv --rm
Removing virtualenv (C:\Users\[USER]\.virtualenvs\root-x8NlcEbx) …
(root-x8NlcEbx) D:\project\root> exit
D:\project\root>
# linux
(root-x8NlcEbx) /DATA/project/root$ pipenv --py
Removing virtualenv ($HOME/.local/share/virtualenvs/root-x8NlcEbx) …
(root-x8NlcEbx) /DATA/project/root$ exit
/DATA/project/root$
```
```pipenv --rm``` 이후 위에서 언급한 **현재 계정 home 의 숨겨진 디렉토리**에서 확인해보면
방금전까지 있던 가상환경이 삭제된 것을 볼 수 있다. 사실상 ```rm -rf``` 로 실제 경로를 삭제해도 되기는 하다.
### gitignore 에 포함할 것들
팀 전체가 공유하는 프로젝트 github repository 에 가상환경 전체가 올라가 버린다면 참으로 비효율적일 것이다.
그래서 사용하는 것이 최소한의 정보만 기록된 requirements.txt 나 Pipfile, Pipfile.lock 인 것이다.
pipenv 는 전혀 다른 공간에 가상환경을 생성하므로 상관 없지만,
virtualenv 를 사용한다면 ```.gitignore``` 에 ```venv``` 혹은 ```가상환경 디렉토리명```을 추가하자.
끝~!
<br><br>
---
<br>
아 그리고 혹시라도 [**결정론적 빌드**](https://reproducible-builds.org/docs/deterministic-build-systems/)가 무엇인지 궁금하다면
링크를 따라가 문서를 읽어보자. 짧게 설명하자면, 최신 버전의 특정 패키지가 현존하는 프로젝트 의존성을 파괴할지도 모르는 상황 속에서,
언제 어디서나 같은 의존성을 보장하는 재현 가능한 빌드를 상상해보면 된다.
결정론적 빌드 시스템!
> **"Be able to get the exact same set of dependencies on multiple machines"**,
>[Using locked packages](https://docs.npmjs.com/files/package-locks#using-locked-packages)
진짜끝!!
| 26.98 | 131 | 0.628614 | kor_Hang | 0.999892 |
9131ee56d21d543ec50a4735087372d2e2ee7840 | 1,376 | md | Markdown | docs/NugetDocumentation.md | LorenzCK/Pseudo-i18n | 08fae4570fb89a832020299a91d1620f703e467c | [
"MIT"
] | 6 | 2016-12-09T01:31:18.000Z | 2019-01-23T17:56:51.000Z | docs/NugetDocumentation.md | LorenzCK/Pseudo-i18n | 08fae4570fb89a832020299a91d1620f703e467c | [
"MIT"
] | null | null | null | docs/NugetDocumentation.md | LorenzCK/Pseudo-i18n | 08fae4570fb89a832020299a91d1620f703e467c | [
"MIT"
] | null | null | null | # Pseudo-i18n
*Simple pseudo-internationalization utility library.*
The library allows you to convert any latin alphabet string to a pseudo-language in order to test whether your application is localization-ready. The generated pseudo-string will try to respect links, tags, and other markup in your original strings.
## Usage
Use the `Translator` class to translate strings at runtime:
```cs
PseudoInternationalization.Translator.ConvertToFakeInternationalized("Hello world");
```
This will return the string `[Ħęľľő ŵőřľđ !!! !!!]`.
An extension method can also be used:
```cs
using PseudoInternationalization;
"Hello world".ToPseudo();
```
The following rules will be applied:
* Strings containing URLs will not be translated,
* Strings shorter than 10 characters will grow by 400%, longer strings will grow by 30% (`!` is used as a padding character),
* Translated strings will always start with `[` and end with `]`,
* Tags, braces, and other markup will not be translated.
## Links
A command-line utility that converts resource files is available from the [library’s Github page](https://github.com/LorenzCK/Pseudo-i18n).
Based on [John Robbin’s Pseudoizer](http://msdn.microsoft.com/msdnmag/issues/04/04/Bugslayer/default.aspx) and [Scott Hanselman’s implementation](http://www.hanselman.com/blog/PsuedoInternationalizationAndYourASPNETApplication.aspx).
| 37.189189 | 249 | 0.77689 | eng_Latn | 0.967275 |
9132dcfe623428ec49b932a5bd286781295c16b8 | 30 | md | Markdown | README.md | khairnaramol/Angular5 | 7915ef1dee3f908b323295f12ef9a588e3b3dbd5 | [
"MIT"
] | null | null | null | README.md | khairnaramol/Angular5 | 7915ef1dee3f908b323295f12ef9a588e3b3dbd5 | [
"MIT"
] | null | null | null | README.md | khairnaramol/Angular5 | 7915ef1dee3f908b323295f12ef9a588e3b3dbd5 | [
"MIT"
] | null | null | null | # Angular5
angular 5 learning
| 10 | 18 | 0.8 | eng_Latn | 0.781067 |
9132fda89a07f12c1729e42ee318d1814c7fd396 | 131 | md | Markdown | README.md | LiveTiles/PageGallery | 829d187da64775795d09b0f6ea5d85fea117336c | [
"MIT"
] | null | null | null | README.md | LiveTiles/PageGallery | 829d187da64775795d09b0f6ea5d85fea117336c | [
"MIT"
] | null | null | null | README.md | LiveTiles/PageGallery | 829d187da64775795d09b0f6ea5d85fea117336c | [
"MIT"
] | null | null | null | # PageGallery
A collection of demo pages intended to guide folks in using (and in some cases hacking) LiveTiles to fit their needs
| 43.666667 | 116 | 0.801527 | eng_Latn | 0.99683 |
913328e76928236576f7f3b53c8297c59213d60f | 3,460 | md | Markdown | posts/blog/2015/06/15-c88.en.md | danmaq/danmaq.article | ab5626c7a8053175d33044a38404bd8f873f79ef | [
"MIT"
] | null | null | null | posts/blog/2015/06/15-c88.en.md | danmaq/danmaq.article | ab5626c7a8053175d33044a38404bd8f873f79ef | [
"MIT"
] | null | null | null | posts/blog/2015/06/15-c88.en.md | danmaq/danmaq.article | ab5626c7a8053175d33044a38404bd8f873f79ef | [
"MIT"
] | null | null | null | ---
title: Comic Market 88 Exhibition Information
post_id: '6827'
date: '2015-06-15T02:25:22+09:00'
draft: true
tags: []
---
We _successfully won the_ comic this summer _at Sunday East Q - 24a_ ! ˶\> ◡ <˶ Although it is quickly, I will announce the distribution etc!
## \[\[Newly released game\] MATH.SC (tentative name)
\] ( [http://math.sc/](http://math.sc/) ) Pre-Alpha trial version 100 yen → FREE! _For the convenience of completeness this time, I will distribute it for free!_ This new publication is a quiz & card battle game original, smart new sense. It is for Windows (Vista SP 2 or later required). Currently working intensively, we will guide you on [special](http://math.sc/) occasions as soon as there is progress! I will be happy to have the new publication as it is going to bring a certain amount of bamboo shoots, so please come and visit the big players slowly. However, _due to various circumstances on_ that _day, we will withdraw after 13:00_ .
## \[\[Pre-released game\] Concert fantasy mirror - mythical mirror
\] ( [http://kagaminer.in/](http://kagaminer.in/) ) 1,000 yen ※ This is the player's replay. [It](http://dangoya.moo.jp/) is a vertical barrage STG using cooperation with Mr. [DANGOYA](http://dangoya.moo.jp/) , the view of the world in the east. Familiar characters such as Cirno and Suwako are like Gensokyo and are not Gensokyo, but have gone into the mysterious world, but such a thing rampage without a doubt! ORIYARA will also appear! It is for Windows (XP SP3 or later required).
## \[\[Ranube Before\] EVIL MAGIC ~ The Witch of Witchcraft
\] (/ evilmagic) 800 yen → 500 yen (Comiket special price)  It is an authentic fantasy light novel of 180 pages. Girl Enna who can use magical abilities and evil cuckoo of brother and childhood rampage! But the worse man, the lord swept away a boy with a childhood friend! Knowing things, hero's heavenly enel challenges the lord against a game ...! What? [If you are interested please try the trial version which can read only the beginning 69 pages!](https://danmaq.com/filez/em_trial.pdf) (4 MB | PDF Attention)
## \[\[Pre-release music CD\] danmaq soundtrack
\] (/! / dst /) ¥ 100  30 songs, a large volume of 75 minutes! It is an _outtake collection_ of songs that danmaq made from 1999 to 2007. Mainly in the eastern direction BEMANI Arrange and many original collections. I recorded only Bocaro and just a little ...!
* [This](https://danmaq.com/!/dst/Hey!.mp3)
* [Songs and so on](https://danmaq.com/filez/music/new_psm.mp3)
* [A lot](https://danmaq.com/filez/music/pcb.mp3)
* [It is included!](http://lama.danmaq.com/lamarisa/mp3/15.mp3) **\[\[Previously released music CD\] Leila Prismriver** \] (/! / Leila /) 300 yen → 100 yen (Comiket special price)  It is an eastward arrangement music CD of 13 songs 64 minutes. We concentrated on secret club music and tried remixing with synthesizer KORG OASYS!
* [Listen to all song cross fade demo](https://danmaq.com/!/leila/x.mp3) (7: 11 | 8 MB) The above announcement is _all scheduled_ . Please note that the distribution may be added or changed by the day. ※ Redistribution of summer in Dream Purple Dream PACK is severe due to circumstances of budget> <I am sorry!
| 104.848485 | 645 | 0.738439 | eng_Latn | 0.986061 |
913461ab034ea0fe626d2990b170c6da5050a9d6 | 2,320 | md | Markdown | index.md | AliNite/spreadsheets-socialsci | ee98eadef71b9f45775e7115cffc397e0e5dbef5 | [
"CC-BY-4.0"
] | null | null | null | index.md | AliNite/spreadsheets-socialsci | ee98eadef71b9f45775e7115cffc397e0e5dbef5 | [
"CC-BY-4.0"
] | null | null | null | index.md | AliNite/spreadsheets-socialsci | ee98eadef71b9f45775e7115cffc397e0e5dbef5 | [
"CC-BY-4.0"
] | null | null | null | ---
layout: lesson
root: .
---
Good data organization is the foundation of any research project. Most
researchers have data in spreadsheets, so it's the place that many research
projects start.
Typically we organize data in spreadsheets in ways that we as humans want to work with the data. However
computers require data to be organized in particular ways. In order
to use tools that make computation more efficient, such as programming
languages like R or Python, we need to structure our data the way that
computers need the data. Since this is where most research projects start,
this is where we want to start too!
In this lesson, you will learn:
- Good data entry practices - formatting data tables in spreadsheets
- How to avoid common formatting mistakes
- Approaches for handling dates in spreadsheets
- Basic quality control and data manipulation in spreadsheets
- Exporting data from spreadsheets
In this lesson, however, you will *not* learn about data analysis with spreadsheets.
Much of your time as a researcher will be spent in the initial 'data wrangling'
stage, where you need to organize the data to perform a proper analysis later.
It's not the most fun, but it is necessary. In this lesson you will
learn how to think about data organization and some practices for more
effective data wrangling. With this approach you can better format current data
and plan new data collection so less data wrangling is needed.
> ## Getting Started
>
> Data Carpentry's teaching is hands-on, so participants are encouraged to use
> their own computers to ensure the proper setup of tools for an efficient
> workflow. <br>**These lessons assume no prior knowledge of the skills or tools.**
>
> To get started, follow the directions in the "[Setup](setup.html)" tab to
> download data to your computer and follow any installation instructions.
>
> #### Prerequisites
>
> This lesson requires a working copy of spreadsheet software, such as Microsoft
> Excel or LibreOffice or OpenOffice.org (see more details in "[Setup](setup/)").
> <br>To most effectively use these materials, please make sure to install
> everything *before* working through this lesson.
{: .prereq}
> ## For Instructors
> If you are teaching this lesson in a workshop, please see the
> [Instructor notes](guide/).
{: .prereq}
| 42.181818 | 105 | 0.775 | eng_Latn | 0.99958 |
91346216640df47b297cb0e4df4cc03f68a1e107 | 3,632 | markdown | Markdown | _posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown | martinlowinski/halfthetruth.de | a9513ca95cb07dad58f625b3b56fa56fbf40e946 | [
"MIT"
] | null | null | null | _posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown | martinlowinski/halfthetruth.de | a9513ca95cb07dad58f625b3b56fa56fbf40e946 | [
"MIT"
] | null | null | null | _posts/2010-03-11-how-a-certification-authority-handles-whois-data.markdown | martinlowinski/halfthetruth.de | a9513ca95cb07dad58f625b3b56fa56fbf40e946 | [
"MIT"
] | null | null | null | ---
wordpress_id: 131
author_login: admin
layout: post
comments: []
author: martinlowinski
title: How a certification authority handles whois data
published: true
tags: []
date: 2010-03-11 17:09:16 +01:00
categories:
- Website
author_email: martin@goldtopf.org
wordpress_url: http://halfthetruth.de/2010/03/11/how-a-certification-authority-handles-whois-data/
author_url: http://goldtopf.org
status: publish
---
I just got a call from someone at <a href="http://comodo.com/">comodo.com</a> and the call went something like this<sup><a href="../2010/03/11/how-a-certification-authority-handles-whois-data/#fn12870367704caa288765c55">1</a></sup>:
<em>co:</em> I am [some name here] from comodo and looking for the owner of halfthetruth.de.
<em>me:</em> That’s me.
<em>co:</em> Great. First I want to let you know that the ssl-certificate for halfthetruth.de has expired, did you know that?
<em>me:</em> Jep.
Then a never ending story about what ssl-certs can do for you and your business and how many clients you’ll get by buying a “so great and cheap certificate from comodo.com” – for only 300 euros (I forgot the exact number) per year.
So I told her, that halfthetruth.de is, as you can see, no business and I don’t have clients to that not existing business and I just don’t need a certificate from comodo.
<em>co:</em> Ok, can I ask you how much you paid for your certificate?
<em>me:</em> Nothing, it’s for free.
<em>*pause</em>
<em>co:</em> Ok well, and how much do you pay for renewing the certificate?
<em>me:</em> It’s for free as well, I can renew it anytime for free.[2]
<em>*pause</em>
<em>co:</em> Than you obviously don’t need a certificate from comodo.
<em>me:</em> Absolutely right, just one question: Where did you get my phone number?
<em>co:</em> We have a research team doing that kind of stuff.
<em>me:</em> But I didn’t post my phone number on my website or somewhere else publicly available.
<em>co:</em> Our research team is doing this, I got your phone number from them..
Then she asked me, if I want to give her my contact information so that she can forward some information regarding comodo and certificates and stuff. I declined politly and hang up. Apparently, they got my phone number from the whois service, but the information from the whois is not meant for advertising as you can see from the terms of use:
<blockquote>Terms and Conditions of Use
The data in this record is provided by DENIC for informational purposes only.
DENIC does not guarantee its accuracy and cannot, under any circumstances,
be held liable in case the stored information would prove to be wrong,
incomplete or not accurate in any sense.
All the domain data that is visible in the whois service is protected by law.
It is not permitted to use it for any purpose other than technical or
administrative requirements associated with the operation of the Internet.
It is explicitly forbidden to extract, copy and/or use or re-utilise in any
form and by any means (electronically or not) the whole or a quantitatively
or qualitatively substantial part of the contents of the whois database
without prior and explicit written permission by DENIC.
It is prohibited, in particular, to use it for transmission of unsolicited
and/or commercial and/or advertising by phone, fax, e-mail or for any similar
purposes.</blockquote>
<p id="fn2">Footnotes:
<sup>1</sup> I cannot quote the exact words.
<sup>2</sup> I’m using a certificate from CAcert, which is free of charge and really great by the way.</p>
| 58.580645 | 349 | 0.755231 | eng_Latn | 0.998625 |
913475b577bba43e252b21252f8a2597af70fcf0 | 1,102 | md | Markdown | ClassNotes/python_class_homework_0.md | jona-sassenhagen/python_for_psychologists | 0604ff5c6382ae02ffeb2e078853b835dab03860 | [
"BSD-3-Clause"
] | 7 | 2018-09-19T20:53:55.000Z | 2022-02-28T12:55:39.000Z | ClassNotes/python_class_homework_0.md | jona-sassenhagen/python_for_psychologists | 0604ff5c6382ae02ffeb2e078853b835dab03860 | [
"BSD-3-Clause"
] | null | null | null | ClassNotes/python_class_homework_0.md | jona-sassenhagen/python_for_psychologists | 0604ff5c6382ae02ffeb2e078853b835dab03860 | [
"BSD-3-Clause"
] | 3 | 2019-03-10T09:25:33.000Z | 2021-12-16T20:24:50.000Z | % Python Class Session 1 Homework
# Repetitions
- Open a new and empty iPython notebook
- Create a list of strings that contains the first names of you and your close family members
- Access the third entry in that list
- Create a list of ages of family members
- *Using these two lists (not manually!)*, create a new list containing the names of those family members older than 30
- *With a dict comprehension and the `zip` operator*, construct a dict where keys are family names and values are ages
## Something to actually mildly warm up your brains for
- similarly to the list you have created above, and *using the dictionary you have just created*, construct a list of the ages of your family members whose last letter of their first name is an 'a'.
# Something new
Put some more thought into a simple experiment you want to conduct. It should be as boring and simple as possible, require only very few subjects, and simple stimuli (tip: words are very simple). Note that reaction times or ratings are the easiest to acquire, and in this context almost the only thing that can be made to work
| 68.875 | 326 | 0.779492 | eng_Latn | 0.999979 |
9134ac8f71966b0eeeabaf4bd6e1a0c4caf0ed32 | 211 | md | Markdown | README.md | teloxide/teloxide-book | 56d226ce5f3efd598365759f8596a3f158ab11a2 | [
"BlueOak-1.0.0"
] | 4 | 2021-09-21T09:51:47.000Z | 2021-11-28T22:17:58.000Z | README.md | teloxide/teloxide-book | 56d226ce5f3efd598365759f8596a3f158ab11a2 | [
"BlueOak-1.0.0"
] | null | null | null | README.md | teloxide/teloxide-book | 56d226ce5f3efd598365759f8596a3f158ab11a2 | [
"BlueOak-1.0.0"
] | null | null | null | # Teloxide user guide
This repository contains a user guide for the `teloxide` library.
[`teloxide`]: https://github.com/teloxide/teloxide
## Note
This book is very much work in progress.
Use with caution.
| 19.181818 | 65 | 0.748815 | eng_Latn | 0.988509 |
9134c2d6582f2fa40d9dfb8ca520377331f0657a | 49 | md | Markdown | README.md | yuleihua/aircmn | ff29b25629dcacf65be4fba7fbefc7e7f624f939 | [
"Apache-2.0"
] | null | null | null | README.md | yuleihua/aircmn | ff29b25629dcacf65be4fba7fbefc7e7f624f939 | [
"Apache-2.0"
] | null | null | null | README.md | yuleihua/aircmn | ff29b25629dcacf65be4fba7fbefc7e7f624f939 | [
"Apache-2.0"
] | 1 | 2021-11-13T15:48:26.000Z | 2021-11-13T15:48:26.000Z | # aircmn
aircmn is common library in c language.
| 16.333333 | 39 | 0.77551 | eng_Latn | 0.999726 |
91354c3382e52e23e19ce5f43f3d7c9492305430 | 2,593 | md | Markdown | doc/debugging.md | Chlorie/libunifex | 9869196338016939265964b82c7244915de6a12f | [
"Apache-2.0"
] | 1 | 2021-11-23T11:30:39.000Z | 2021-11-23T11:30:39.000Z | doc/debugging.md | Chlorie/libunifex | 9869196338016939265964b82c7244915de6a12f | [
"Apache-2.0"
] | null | null | null | doc/debugging.md | Chlorie/libunifex | 9869196338016939265964b82c7244915de6a12f | [
"Apache-2.0"
] | 1 | 2021-07-29T13:33:13.000Z | 2021-07-29T13:33:13.000Z | # Async Stack Traces
Unifex contains a prototype implementation of async stack-traces that
allows you to traverse a chain/graph of async continuations.
A stack-trace consists of a stack of `continuation_info` objects that
describes the address of the "frame" and the type of the continuation
as well as a mechanism to query what the next continuations in the chain are.
This allows you to traverse from a leaf receiver back to the original task
that launched it. If you are using structured concurrency and have represented
your application as a structured set of tasks then this chain should progress
all the way back to the root task of your application.
Each receiver must customise the `visit_continuations()` CPO to be able to
participate in the async stack-walk. Otherwise, the stack-walk will terminate
when it reaches that receiver.
Example:
```c++
template<typename Receiver>
struct my_receiver {
Receiver wrappedReceiver_;
void set_value() && noexcept;
void set_error(std::exception_ptr) && noexcept;
void set_done() && noexcept;
template <typename Func>
friend void tag_invoke(
tag_t<visit_continuations>, const my_receiver& r, Func&& func) {
std::invoke(func, wrappedReceiver_);
}
};
```
## Capturing the current stack-trace
There is a helper sender called `async_trace_sender` that you can use to
get a dump of the async stack-trace at any point in a sender expression.
It will produce a `std::vector<async_trace_entry>` that contains a description
of the async stack at this point.
For example: Some helpers to dump an async trace.
```c++
auto dump_async_trace(std::string tag = {}) {
return transform(
async_trace_sender{},
[tag = std::move(tag)](const std::vector<async_trace_entry>& entries) {
std::cout << "Async Trace (" << tag << "):\n";
for (auto& entry : entries) {
std::cout << " " << entry.depth << " [-> " << entry.parentIndex
<< "]: " << entry.continuation.type().name() << " @ 0x";
std::cout.setf(std::ios::hex, std::ios::basefield);
std::cout << entry.continuation.address();
std::cout.unsetf(std::ios::hex);
std::cout << "\n";
}
});
}
template <typename Sender>
auto dump_async_trace_on_start(Sender&& sender, std::string tag = {}) {
return unifex::sequence(dump_async_trace(std::move(tag)), (Sender &&) sender);
}
template <typename Sender>
auto dump_async_trace_on_completion(Sender&& sender, std::string tag = {}) {
return unifex::finally(
(Sender &&) sender, dump_async_trace(std::move(tag)));
}
```
| 35.520548 | 80 | 0.698419 | eng_Latn | 0.946728 |
9135674fda29402a227ebc0d7f0cfe81339cf0b4 | 82 | md | Markdown | README.md | Bearzilasaur/ScholarScraper | ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164 | [
"Unlicense"
] | null | null | null | README.md | Bearzilasaur/ScholarScraper | ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164 | [
"Unlicense"
] | null | null | null | README.md | Bearzilasaur/ScholarScraper | ad3a638b0b8f3f13ae1d6a84711cbb5eedcc1164 | [
"Unlicense"
] | 1 | 2019-10-16T13:20:10.000Z | 2019-10-16T13:20:10.000Z | # ScholarScraper
Repository for a Google Scholar scraper for literature reviews.
| 27.333333 | 64 | 0.829268 | eng_Latn | 0.813123 |
9135823d4830ff5567be54c2a660f13205ae659c | 295 | md | Markdown | playbooks/openshift-monitor-availability/README.md | Roscoe198/Ansible-Openshift | b874bef456852ef082a27dfec4f2d7d466702370 | [
"Apache-2.0"
] | 164 | 2015-07-29T17:35:04.000Z | 2021-12-16T16:38:04.000Z | playbooks/openshift-monitor-availability/README.md | Roscoe198/Ansible-Openshift | b874bef456852ef082a27dfec4f2d7d466702370 | [
"Apache-2.0"
] | 3,634 | 2015-06-09T13:49:15.000Z | 2022-03-23T20:55:44.000Z | playbooks/openshift-monitor-availability/README.md | Roscoe198/Ansible-Openshift | b874bef456852ef082a27dfec4f2d7d466702370 | [
"Apache-2.0"
] | 250 | 2015-06-08T19:53:11.000Z | 2022-03-01T04:51:23.000Z | # OpenShift Availability Monitoring
This playbook runs the [OpenShift Availability Monitoring role](../../roles/openshift_monitor_availability). See the role
for more information.
## GCP Development
The `install-gcp.yml` playbook is useful for ad-hoc installation in an existing GCE cluster.
| 32.777778 | 121 | 0.8 | eng_Latn | 0.959277 |
91367433ac2c38ef6788dc22e9c3fbb7bda51853 | 36 | md | Markdown | src/examples/subscript/simple.md | alinex/node-report | 0798d2bacf8064875b3f54cd035aa154306f5a7e | [
"Apache-2.0"
] | 1 | 2016-06-02T15:05:20.000Z | 2016-06-02T15:05:20.000Z | src/examples/subscript/simple.md | alinex/node-report | 0798d2bacf8064875b3f54cd035aa154306f5a7e | [
"Apache-2.0"
] | null | null | null | src/examples/subscript/simple.md | alinex/node-report | 0798d2bacf8064875b3f54cd035aa154306f5a7e | [
"Apache-2.0"
] | null | null | null | You need H~2~O for this experiment.
| 18 | 35 | 0.75 | eng_Latn | 0.999731 |
9136b6596227434d736d133c7a82fabbf47a2493 | 964 | md | Markdown | _posts/2019-01-19-Github Blog.md | stone8765/blog | 90d6420aaef33eae21235075394dd7e332be071b | [
"MIT"
] | null | null | null | _posts/2019-01-19-Github Blog.md | stone8765/blog | 90d6420aaef33eae21235075394dd7e332be071b | [
"MIT"
] | null | null | null | _posts/2019-01-19-Github Blog.md | stone8765/blog | 90d6420aaef33eae21235075394dd7e332be071b | [
"MIT"
] | null | null | null | ---
layout: post
title: Github Blog 搭建
author: StoneLi
description: 使用frp内网穿透工具可以让内网中的电脑能够像访问公网电脑一样方便,比如将公司或个人电脑里面的Web项目让别人能够访问、或进行电脑远程连接、或ssh连接
catalog: true
tags: [jekyll,github pages]
---
# 1. 安装Ruby
https://rubyinstaller.org/
# 2. 下载安装gem (Ruby的包管理器)
下载:https://rubygems.org/pages/download
解压之后 在目录中执行以下命令
ruby setup.rb
# 3. 安装jekyll
在命令行执行gem install jekyll
# 4. 运行
jekyll new myblog
cd myblog
jekyll server
在浏览器输入http://127.0.0.1:4000/即可浏览刚刚创建的blog
# 5. Jekyll 主题选择
进入网站 http://jekyllthemes.org/
选择主题,下载对应的仓库代码到本地即可
# 6.文件中文名本地无法显示的问题
修改安装目录\Ruby22-x64\lib\ruby\2.2.0\webrick\httpservlet下的filehandler.rb文件,建议先备份。
找到下列两处,添加一句(+的一行为添加部分)
```
path = req.path_info.dup.force_encoding(Encoding.find("filesystem"))
+ path.force_encoding("UTF-8") # 加入编码
if trailing_pathsep?(req.path_info)
```
```
break if base == "/"
+ base.force_encoding("UTF-8") # 加入编码
break unless File.directory?(File.expand_path(res.filename + base))
```
修改完重新jekyll serve即可支持中文文件名。 | 20.956522 | 89 | 0.763485 | yue_Hant | 0.648572 |
913752a80068d6d04d5c0e2c018fa9ab92efb7c0 | 6,182 | md | Markdown | README.md | pecigonzalo/opta | 0259f128ad3cfc4a96fe1f578833de28b2f19602 | [
"Apache-2.0"
] | null | null | null | README.md | pecigonzalo/opta | 0259f128ad3cfc4a96fe1f578833de28b2f19602 | [
"Apache-2.0"
] | null | null | null | README.md | pecigonzalo/opta | 0259f128ad3cfc4a96fe1f578833de28b2f19602 | [
"Apache-2.0"
] | null | null | null | <p align="center"><img src="https://user-images.githubusercontent.com/855699/125824286-149ea52e-ef45-4f41-9579-8dba9bca38ac.png" width="250"><br/>
Automated, secure, scalable cloud infrastructure</p>
<p align="center">
<a href="https://github.com/run-x/opta/releases/latest">
<img src="https://img.shields.io/github/release/run-x/opta.svg" alt="Current Release" />
</a>
<a href="https://github.com/run-x/opta/actions/workflows/ci.yml">
<img src="https://github.com/run-x/opta/actions/workflows/ci.yml/badge.svg" alt="Tests" />
</a>
<a href="https://codecov.io/gh/run-x/opta">
<img src="https://codecov.io/gh/run-x/opta/branch/main/graph/badge.svg?token=OA3PXV0HYX">
</a>
<a href="http://www.apache.org/licenses/LICENSE-2.0.html">
<img src="https://img.shields.io/badge/LICENSE-Apache2.0-ff69b4.svg" alt="License" />
</a>
<img src="https://img.shields.io/github/commit-activity/w/run-x/opta.svg?style=plastic" alt="Commit Activity" />
<img src="https://img.shields.io/badge/downloads-500%2Fweek-blue" />
<a href="https://github.com/PyCQA/bandit">
<img src="https://img.shields.io/badge/security-bandit-yellow.svg" alt="Security" />
</a>
</p>
<p align="center">
<a href="https://docs.opta.dev/">Documentation</a> |
<a href="https://slack.opta.dev">
Slack Community
</a> | <a href="mailto:info@runx.dev">
Email: info@runx.dev
</a>
</p>
# What is Opta?
Opta is a new kind of Infrastructure-as-Code (IaC) framework that lets engineers work with high-level constructs
instead of getting lost in low-level cloud configuration. Opta has a vast library of modules (like EKS, RDS, DynamoDB,
GKE, Cloud SQL, and even third-party services like Datadog) that engineers can compose together to build their ideal
infrastructure stack. It's built on top of Terraform, and designed so engineers aren’t locked in – anyone can write custom Terraform
or even take the Opta-generated Terraform and work independently.
Opta gives any engineering team, whether it’s a team of 2 or 200, the same infrastructure advantages that companies
like Google or Facebook have, without having to invest in infrastructure or DevOps engineers.
### Why use Opta?
Infrastructure-as-Code (IaC) solutions are now widely accepted as the standard for provisioning and managing cloud
infrastructure, and Terraform is widely considered to be the best IaC platform on the market – and it is – but Terraform
is also quite complex and requires deep Cloud/infrastructure expertise. We developed Opta to help eliminate this complexity.
Opta is a simpler IaC framework with best practices built-in. It lets users set up automated, scalable and secure infrastructure
on any cloud, without having to be an infrastructure expert, or getting lost in the minutiae of cloud configuration.
We are confident it can drastically reduce the complexity and headaches that come with DevOps and infrastructure at most
fast moving organizations. Opta is currently being used by dozens of companies of all sizes.
To read more about the vision behind Opta, check out this [blog post](https://blog.runx.dev/infrastructure-as-code-for-everyone-7dad6b813cbc).
If you'd like to try it out or have any questions - feel free to join our [Slack](https://slack.opta.dev/) or explore the [Getting Started Guide](https://docs.opta.dev/getting-started)!
<p align="center">
<a href="https://www.youtube.com/watch?v=nja_EfpGexE"><img width="480" src="https://user-images.githubusercontent.com/855699/149367998-9f00a9f4-abaa-4abf-949c-5b470e7d410c.png"></a>
</br>
<span><i>Deploying a Ruby on Rails application to AWS</i></span>
</p>
# Features
### Cloud services
Opta supports the 3 major clouds - AWS, GCP and Azure. It has modules for the most commonly used services in these clouds like:
* Microservices (powered by [Kubernetes](https://docs.opta.dev/architecture/aws/))
* Databases - Postgres, MySQL, Redis
* Serverless workloads
* Networking - VPCs, Subnets, Load balancers
* CDN (Content Delivery Network)
* Object storage (S3, GCS)
Additionally, Opta has cloud and security best practices built in, including:
* [Observability](https://docs.opta.dev/features/observability/) (Datadog, LogDNA)
* [SOC2 compliance](https://docs.opta.dev/compliance/)
* [Continuous Deployment](https://docs.opta.dev/features/continuous_deployment/)
* Hardened network and security configurations ([AWS](https://docs.opta.dev/architecture/aws/), [GCP](https://docs.opta.dev/architecture/gcp/), [Azure](https://docs.opta.dev/architecture/azure))
* Auto-scaling and high availability (HA)
### Coexistence with existing infrastructure
Opta aims to be compatible with your existing infrastructure setup. You can:
* Import existing Terraform infrastructure into Opta
* Write [custom Terraform modules](https://docs.opta.dev/reference/aws/modules/custom-terraform/) (for services that Opta doesn't support yet)
* Run Opta in existing VPCs (WIP)
* Export the generated Terraform
# Try out Opta
Check out the [Getting Started Guide](https://docs.opta.dev/getting-started/).
You can also explore some [examples](https://github.com/run-x/opta/tree/main/examples) to get a better idea of what you can do with Opta.
# Development
We love user contributions! Check out our [Contributing](https://github.com/run-x/opta/blob/main/CONTRIBUTING.md) and [Dev guide](https://github.com/run-x/opta/blob/main/development.md) to get started.
# Important Resources
* [The Opta Team](https://www.runx.dev/about)
* [Check Out The Blog](https://blog.runx.dev/)
* [How Opta delivers upgrades](https://github.com/run-x/opta/blob/main/UPGRADING.md)
* [Bugfix / Feature request policy](https://github.com/run-x/opta/blob/main/CONTRIBUTING.md#bugfix-resolution-time-expectations)
* Comparison with other tools (WIP)
* [Our Public roadmap](https://github.com/orgs/run-x/projects/1/views/1)
* Case studies - [Flyte](https://blog.flyte.org/how-opta-makes-deploying-flyte-much-easier), [Fastbreak Labs](https://blog.runx.dev/how-fast-break-labs-uses-opta-to-bring-basketball-to-the-blockchain-7556353d70ee), [Canvas app](https://blog.runx.dev/how-the-canvas-team-uses-opta-to-make-data-easier-to-explore-f5615647cc43)
| 56.2 | 324 | 0.752507 | eng_Latn | 0.77633 |
9137cde5130df8bd9d5762e1bad5d97d338af01f | 5,520 | md | Markdown | packages/speeddial/CHANGELOG.md | zamblas/ui-material-components | ec3a4203c0de76d56814e72cfd32e7bd3c077a40 | [
"Apache-2.0"
] | null | null | null | packages/speeddial/CHANGELOG.md | zamblas/ui-material-components | ec3a4203c0de76d56814e72cfd32e7bd3c077a40 | [
"Apache-2.0"
] | null | null | null | packages/speeddial/CHANGELOG.md | zamblas/ui-material-components | ec3a4203c0de76d56814e72cfd32e7bd3c077a40 | [
"Apache-2.0"
] | null | null | null | # Change Log
All notable changes to this project will be documented in this file.
See [Conventional Commits](https://conventionalcommits.org) for commit guidelines.
## [5.2.8](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.7...v5.2.8) (2021-02-24)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.7](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.6...v5.2.7) (2021-02-22)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.5](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.4...v5.2.5) (2021-02-04)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.4](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.3...v5.2.4) (2021-02-04)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.3](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.2...v5.2.3) (2021-01-23)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.2](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.1...v5.2.2) (2021-01-20)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.2.1](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.2.0...v5.2.1) (2021-01-19)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
# [5.2.0](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.16...v5.2.0) (2021-01-18)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.16](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.15...v5.1.16) (2020-12-14)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.15](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.14...v5.1.15) (2020-12-13)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.14](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.13...v5.1.14) (2020-12-10)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.13](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.12...v5.1.13) (2020-12-07)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.12](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.11...v5.1.12) (2020-12-04)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.10](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.9...v5.1.10) (2020-12-03)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.9](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.8...v5.1.9) (2020-12-01)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.8](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.7...v5.1.8) (2020-11-27)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.7](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.6...v5.1.7) (2020-11-26)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.6](https://github.com/nativescript-community/ui-material-components/tree/master/packages/speeddial/compare/v5.1.5...v5.1.6) (2020-11-23)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.5](https://github.com/Akylas/nativescript-material-components/compare/v5.1.4...v5.1.5) (2020-11-22)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.4](https://github.com/Akylas/nativescript-material-components/compare/v5.1.3...v5.1.4) (2020-11-13)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.3](https://github.com/Akylas/nativescript-material-components/compare/v5.1.2...v5.1.3) (2020-11-12)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
## [5.1.1](https://github.com/Akylas/nativescript-material-components/compare/v5.1.0...v5.1.1) (2020-11-08)
**Note:** Version bump only for package @nativescript-community/ui-material-speeddial
# [5.1.0](https://github.com/Akylas/nativescript-material-components/compare/v5.0.30...v5.1.0) (2020-11-06)
### Features
* speeddial basic package ([7e221bb](https://github.com/Akylas/nativescript-material-components/commit/7e221bbb8b84eced2df9e0d81a18d06b1fc5d370))
| 29.361702 | 147 | 0.747464 | eng_Latn | 0.201678 |
9137eb9f3047c740fcacf366eaa16c3c438e5a31 | 204 | md | Markdown | README.md | yatace/agn | 597a33faf167b31a7fb584f8bebe0842a94b2150 | [
"MIT"
] | 22 | 2019-03-01T04:47:56.000Z | 2021-06-24T08:31:41.000Z | README.md | yatace/agn | 597a33faf167b31a7fb584f8bebe0842a94b2150 | [
"MIT"
] | 3 | 2019-03-05T15:34:02.000Z | 2020-05-23T03:38:44.000Z | README.md | yatace/agn | 597a33faf167b31a7fb584f8bebe0842a94b2150 | [
"MIT"
] | 5 | 2019-03-01T07:53:49.000Z | 2019-03-05T03:26:32.000Z | # AGN生成器
## AGN 全称 Make Acfun Great Again Network ~~Acfun Green(GKD) Network~~
简单的agn评分生成和查询系统
### 特别鸣谢
[btboyhappy1993](https://github.com/btboyhappy1993)
以上acer为AGN事业作出的贡献
### [更新记录](changelog.md)
| 15.692308 | 69 | 0.735294 | yue_Hant | 0.516636 |
91387a86e3b12c9bdc2df6da857094de4b41fff5 | 6,546 | md | Markdown | src/posts/gsoc-week-3.md | isabelcosta/website | 777d1a20c6ef45be87848f829e2935b302d5a65a | [
"MIT"
] | 3 | 2020-06-29T11:36:10.000Z | 2020-07-03T10:21:23.000Z | src/posts/gsoc-week-3.md | isabelcosta/isabelcosta.github.io | 592ae44426e30c8cedbdbca83af5cc3ec07a71f1 | [
"MIT"
] | 97 | 2019-01-30T23:46:40.000Z | 2022-02-26T01:59:47.000Z | src/posts/gsoc-week-3.md | isabelcosta/website | 777d1a20c6ef45be87848f829e2935b302d5a65a | [
"MIT"
] | 7 | 2019-05-24T11:42:57.000Z | 2021-05-14T15:50:26.000Z | ---
title: Google Summer of Code | Coding Period | Week 3
date: '2018-06-03'
tags:
- gsoc
crossposts:
medium: https://medium.com/isabel-costa-gsoc/google-summer-of-code-coding-period-week-3-349e08f7d998
---

This week — May 28 to June 3 — was the third week of the coding period of [Google Summer of Code (GSoC)](https://summerofcode.withgoogle.com/) with [Systers Open Source](https://github.com/systers). If you want to know more about this you can read the [introduction to my journey](https://medium.com/isabel-costa-gsoc/intro-to-google-summer-of-code-with-systers-open-source-dbdaa92bd189) , [my latest weekly blog posts](https://medium.com/isabel-costa-gsoc) or [my weekly status report](https://github.com/systers/mentorship-backend/wiki/GSoC-2018-Isabel-Costa#weekly-status-report-for-week-3) for this week.
These first weeks, I’ve been balancing GSoC and my Master’s thesis work. Next week I’ll present my dissertation and then I can fully focus on GSoC work. Even with the thesis I still managed to get some work done.
This week I had these 3 pull requests (PR) merged:
- Add Swagger documentation for Login API [#20](https://github.com/systers/mentorship-backend/pull/20);
- Add initial tests that cover the majority of database models and data access object (DAO) [#16](https://github.com/systers/mentorship-backend/pull/16);
- Update pull request template, to have checklist items to update swagger and postman JSON files. These files must be updated every time someone updates the API functionalities [#27](https://github.com/systers/mentorship-backend/pull/27).
I also created [some issues on GitHub and organized them](https://github.com/systers/mentorship-backend/issues?utf8=%E2%9C%93&q=is%3Aissue+created%3A2018-05-28..2018-06-03+author%3Aisabelcosta) on the [ZenHub](https://www.zenhub.com/) board. In case you’re unfamiliar with [ZenHub](https://www.zenhub.com/), this is an agile project management tool for [GitHub](https://github.com). Some of these issues were based on bugs found by one of my mentors, [Murad](https://github.com/m-murad), and other issues were discussed at this week’s project meeting.
### Challenges
This week I had to squash commits on multiple occasions. For [my first PR](https://github.com/systers/mentorship-backend/pull/14), I had to squash some commits of a branch, which did not include the most recent commit ([here’s a stack overflow answer that helped me](https://stackoverflow.com/questions/24310554/how-do-i-squash-specific-commits-on-a-local-branch/24310701#24310701)). For some of my next PRs I had to squash all commits of a certain branch. I never find my self in any of these scenarios before, so it was quite challenging learning about this. I eventually learned how to squash commits, with some tips from the community and one of my mentors, [Dilushi](https://github.com/Dilu9218). After the second instance of squashing the commits, I was much more comfortable at doing this, for the case where I squashed all the commits of a branch.
Because I was still getting acquainted with flask backend development, I wasn’t doing tests and applying [Test-Driven Development (TDD)](https://en.wikipedia.org/wiki/Test-driven_development) methodology yet. I was still trying to figure out what was the best way to structure the app to allow me to test each separate module of the app, from DAOs to data model objects and the API itself. So I dedicated a large portion of this week to learn how to test the code. I implemented some basic tests, just to have an idea of how to test the project from now on. Although tests are a very important part of the project development, for now doing all the possible combination of tests wasn’t a priority, this can be done gradually and implemented for the next features. Also, we now have a more stable structure, than in the first weeks, and I understand a bit more of test development for flask applications. Another point is that by having this project open source, anyone can help to add more tests to the project during the GSoC period as well.
I fixed [Swagger](https://swagger.io/) documentation of the backend API so that anyone can test the whole API, using the [Swagger UI](https://swagger.io/tools/swagger-ui/) instead of just using [Postman](https://www.getpostman.com/) (how I usually to test the API). I learned how to use [flask-RESTPlus](http://flask-restplus.readthedocs.io) framework to both document header authorization field and request body. The authorization field is required for some endpoints that require login with an access token.
### Shout-outs to blog posts and open source projects
This week, I discovered other great GitHub open source sample projects that helped me in the learning journey. One of these projects — [mjhea0/flaskr-tdd](https://github.com/mjhea0/flaskr-tdd) — shows how to approach TDD using flask applications. Another project that I found was [mjhea0/flask-basic-registration](https://github.com/mjhea0/flask-basic-registration). These two projects were done by [Michael Herman](https://github.com/mjhea0).
I also searched about how to implement email verification for new users, which I’m aiming to start implementing next week. I found [this blog post](https://realpython.com/handling-email-confirmation-in-flask/) on this topic, which seems very useful, from [Real Python](https://realpython.com/) website. Here’s the [project from the blog post on GitHub](https://github.com/realpython/flask-registration).
Another great article that I found was [“How to structure a Flask-RESTPlus web service for production builds”](https://medium.freecodecamp.org/structuring-a-flask-restplus-web-service-for-production-builds-c2ec676de563) by [Greg Obinna](https://medium.com/@gregobinna). I intend to look more into this in the next days. The GitHub project used for this blog post can be found in [cosmic-byte/flask-restplus-boilerplate](https://github.com/cosmic-byte/flask-restplus-boilerplate).
### Plans for next week
- Fix bugs found by my mentor and solve small tasks, until thesis presentation;
- Implement email verification;
- Start implementing features related to users mentorship pairing (probably after thesis presentation);
- Create some issues, approved by my mentors, for new contributors to help out. These will be labeled as “first timers only”, which is the Systers label used to guide newcomers into easy quick issues to get started with open source contributions for [Systers Open Source](https://github.com/systers).
| 139.276596 | 1,042 | 0.779713 | eng_Latn | 0.992185 |
9138911b243ecab75293bb6446d9668ffadda978 | 1,339 | md | Markdown | docs/nest/README.md | hackycy/sf-admin-cli | 4965d5741589a4589d2ce66824af383299008a17 | [
"MIT"
] | 3 | 2021-12-13T07:44:16.000Z | 2022-03-11T17:59:02.000Z | docs/nest/README.md | hackycy/sf-admin-cli | 4965d5741589a4589d2ce66824af383299008a17 | [
"MIT"
] | 1 | 2021-12-13T07:15:58.000Z | 2021-12-13T07:42:59.000Z | docs/nest/README.md | hackycy/sf-admin-cli | 4965d5741589a4589d2ce66824af383299008a17 | [
"MIT"
] | 1 | 2022-03-02T02:38:04.000Z | 2022-03-02T02:38:04.000Z | # 介绍
   
**基于NestJs + TypeScript + TypeORM + Redis + MySql + Vue + Element-UI编写的一款简单高效的前后端分离的权限管理系统。具有:**
- 前后端请求参数校验
- JWT 认证
- 基于 NestJs 框架,内置了基础的中间件支持(用户认证、访问日志、请求追踪等)
- 用户权限动态刷新
- 代码简单,结构清晰
::: tip 环境要求
- Node.js 12.x+
- Typescript 2.8+
- MYSQL 5.7+
- Redis 5.0+
:::
## 演示地址
- [http://opensource.admin.si-yee.com](http://opensource.admin.si-yee.com/)
- [Swagger Api文档](http://opensource.admin.si-yee.com/api/doc/admin/swagger-api/static/index.html)
演示环境账号密码:
| 账号 | 密码 | 权限 |
| :----------: | :----: | :----------------------: |
| openadmin | 123456 | 仅只有各个功能的查询权限 |
| monitoradmin | 123456 | 系统监控页面及按钮权限 |
> 所有新建的用户初始密码都为123456
本地部署账号密码:
| 账号 | 密码 | 权限 |
| :-------: | :----: | :--------: |
| rootadmin | 123456 | 超级管理员 |
## 模块列表
``` bash
├─系统管理
│ ├─用户管理
│ ├─角色管理
│ ├─菜单管理
├─系统监控
│ ├─在线用户
│ ├─登录日志
│ ├─请求追踪
├─任务调度
│ ├─定时任务
│ └─任务日志
├─网盘空间
│ ├─空间管理
│ └─空间概览
```
## 技术选型
### 后端
- NestJs + TypeScript
- TypeORM(MYSQL)
- ioredis(Redis)
- bull(队列)
### 前端
- Vue、Vue-Router、VueX
- Element-UI | 18.859155 | 277 | 0.585512 | yue_Hant | 0.701851 |
9139f94e9a00280f0762282414141a07cf115673 | 1,295 | md | Markdown | _posts/2018/2018-03-26-apache-proxy-balancer.md | haijunsu/navysu.github.io | c5e5d39d4a3dae79a0750e136b6a22e743e60db9 | [
"MIT"
] | null | null | null | _posts/2018/2018-03-26-apache-proxy-balancer.md | haijunsu/navysu.github.io | c5e5d39d4a3dae79a0750e136b6a22e743e60db9 | [
"MIT"
] | null | null | null | _posts/2018/2018-03-26-apache-proxy-balancer.md | haijunsu/navysu.github.io | c5e5d39d4a3dae79a0750e136b6a22e743e60db9 | [
"MIT"
] | null | null | null | ---
title: Apache Proxy Balancer
author: Haijun (Navy) Su
layout: post
tags: [proxy, balancer, apache, linux]
---
### Enable proxy models
```shell
sudo a2enmod proxy_html
sudo a2enmod proxy_http
sudo a2enmod proxy_wstunnel
sudo a2enmod proxy_ajp
sudo a2enmod lbmethod_byrequests
sudo a2enmod lbmethod_bytraffic
sudo a2enmod lbmethod_bybusyness
sudo a2enmod lbmethod_heartbeat
```
### Modify configuration file
```xml
<IfModule mod_ssl.c>
<VirtualHost *:443>
ServerName happynavy.tk
ServerAlias www.happynavy.tk
ServerAdmin me@gmail
ProxyPreserveHost On
Header add Set-Cookie "ROUTEID=.%{BALANCER_WORKER_ROUTE}e; path=/" env=BALANCER_ROUTE_CHANGED
<Proxy "balancer://mycluster">
BalancerMember "http://swarm-manager01:3080" route=1
BalancerMember "http://swarm-manager02:3080" route=2
BalancerMember "http://swarm-manager03:3080" route=3
ProxySet stickysession=ROUTEID
</Proxy>
ProxyPass "/" "balancer://mycluster/"
ProxyPassReverse "/" "balancer://mycluster/"
SSLCertificateFile /etc/letsencrypt/live/happynavy.tk/fullchain.pem
SSLCertificateKeyFile /etc/letsencrypt/live/happynavy.tk/privkey.pem
Include /etc/letsencrypt/options-ssl-apache.conf
</VirtualHost>
</IfModule>
```
source: <https://httpd.apache.org/docs/2.4/mod/mod_proxy_balancer.html>
| 26.979167 | 95 | 0.769884 | yue_Hant | 0.292929 |
913a6aeb4f157a79c2fc779a640c8c7533aeb8b7 | 276 | md | Markdown | blog/first-post.md | acrobertson/gatsby-netlify-test | 39c544aa52f10091afe9276efc454a033e43b246 | [
"MIT"
] | null | null | null | blog/first-post.md | acrobertson/gatsby-netlify-test | 39c544aa52f10091afe9276efc454a033e43b246 | [
"MIT"
] | 4 | 2021-03-09T18:58:02.000Z | 2022-02-26T18:17:35.000Z | blog/first-post.md | acrobertson/gatsby-netlify-test | 39c544aa52f10091afe9276efc454a033e43b246 | [
"MIT"
] | null | null | null | ---
path: /blog/post-1
date: 2019-10-08T18:05:15.349Z
title: First Post
---
## First Post Content
The following is the content of the first post
### Here's a list
- Item 1
- Item 2
- Item 3
[This](https://example.com/ "example") is a link
## Another header
More content
| 13.142857 | 48 | 0.673913 | eng_Latn | 0.96378 |
913b38cd59d141fcb6d43b788b5ff59a13db35af | 597 | md | Markdown | includes/data-explorer-authentication.md | p770820/azure-docs.zh-tw | dd2bd917784a4df8b52787a299a3df42e05642fe | [
"CC-BY-4.0",
"MIT"
] | null | null | null | includes/data-explorer-authentication.md | p770820/azure-docs.zh-tw | dd2bd917784a4df8b52787a299a3df42e05642fe | [
"CC-BY-4.0",
"MIT"
] | null | null | null | includes/data-explorer-authentication.md | p770820/azure-docs.zh-tw | dd2bd917784a4df8b52787a299a3df42e05642fe | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
author: orspod
ms.service: data-explorer
ms.topic: include
ms.date: 10/07/2019
ms.author: orspodek
ms.openlocfilehash: a04f17ac809832b6fec51d1ffe0d9fcd6285b4ff
ms.sourcegitcommit: f4d8f4e48c49bd3bc15ee7e5a77bee3164a5ae1b
ms.translationtype: MT
ms.contentlocale: zh-TW
ms.lasthandoff: 11/04/2019
ms.locfileid: "73581832"
---
## <a name="authentication"></a>驗證
若要執行下列範例,您需要可存取資源的 Azure Active Directory (Azure AD)應用程式和服務主體。 若要建立免費 Azure AD 應用程式,並在訂用帳戶層級新增角色指派,請參閱[建立 Azure AD 應用程式](/azure/active-directory/develop/howto-create-service-principal-portal)。 您也需要 [目錄(租使用者)識別碼]、[應用程式識別碼] 和 [用戶端密碼]。
| 35.117647 | 232 | 0.798995 | yue_Hant | 0.265369 |
913bc31de79d1e9550e767cd21692f2a7d809a88 | 4,698 | md | Markdown | _episodes/01-run-quit.md | lexnederbragt/python-novice-gapminder | 6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5 | [
"CC-BY-4.0"
] | null | null | null | _episodes/01-run-quit.md | lexnederbragt/python-novice-gapminder | 6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5 | [
"CC-BY-4.0"
] | null | null | null | _episodes/01-run-quit.md | lexnederbragt/python-novice-gapminder | 6edb7ae77e0f7ae22f6b096fbdec614a5cda78b5 | [
"CC-BY-4.0"
] | null | null | null | ---
title: "Running and Quitting"
teaching: 15
exercises: 0
questions:
- "How can I run Python programs?"
objectives:
- "Launch the Jupyter Notebook, create new notebooks, and exit the Notebook."
- "Create Markdown cells in a notebook."
- "Create and run Python cells in a notebook."
keypoints:
- FIXME
---
### Python Programs are plain text files.
* They have the ".py" extension to let everyone know (including the operating system) it is a Python program.
* It's common to write them using a text editor but we are going to use the Jupyter notebook.
* The bit of extra setup is well worth it because the notebook provides code completion and other helpful features.
* The notebooks have the extension ".ipynb" to distinguish them from plain text python programs.
### The Jupyter Notebook
* The [Anaconda package manager][anaconda] is an automated way to install the Jupyter notebook.
* It also installs all the extra libraries it needs to run.
* Once you have installed Python and the Jupyter Notebook requirements, open a bash shell and type:
~~~
$ jupyter notebook .
~~~
{: .source}
* This will start a Jupyter Notebook server and open your default web browser.
* The server sends messages to your browser.
* The server does the work and the web browser renders the notebook.
* You can type code into the browser and see the result when the web page talks to the server.
* FIXME: what advantages are there to this arrangement?
* The notebook is stored as JSON.
* Just like a webpage, the saved notebook looks different to what you see when it gets rendered by your browser.
### Control and Edit modes
* Open a new notebook from the dropdown menu in the top right corner of the file browser page.
* If you press "esc" and "return" alternately, you will see the surround of your code cell change from blue to green.
* The difference in colour is subtle.
* These are the control and edit modes of your notebook.
* If you use the "esc" and "return" keys to make the surround blue and then press the "H" key, a list of all the shortcut keys will appear.
* When in control mode (esc/Blue),
* The "B" key will make a new cell below the currently selected cell.
* The "A" key will make one above.
* The "X" key will delete the current cell.
* There are lots of shortcuts you can try out and most actions can done with the menus at the top of the page if you forget the shortcuts.
* *If you first remember the "esc" and "H" shortcut, you will be able to find out all the rest.*
* Pressing the "return" key turns the surround green to signify edit mode and you can type code into the cell.
* Because we want to be able to write many lines of code in a single cell, the "return" key will do what it normally does.
* It moves the cursor to the next line just like in a text editor.
* We need some other way to tell the Notebook we want to execute the code in the cell.
* Pressing the "return" key and the "shift" key together will execute the code in the cell.
* Notice that the "return" and "shift" keys on the right of the keyboard are right next to each other.
### Markdown
* Notebooks can also render Markdown.
* Turn the current cell into a Markdown cell by entering the control mode (esc/blue) and press the "M" key.
* The `In [ ]:` will disappear to show it is no longer a code cell and you will be able to write in Markdown.
* FIXME: how do you turn a Markdown cell back into a code cell?
> ## Creating Lists
>
> Create a numbered list in a Markdown cell in a notebook
> that looks like this:
>
> 1. Get funding.
> 2. Design experiment.
> 3. Collect data.
> 4. Analyze.
> 5. Write up.
> 6. Publish.
{: .challenge}
> ## More Math
>
> What is displayed when a Python cell in a notebook
> that contains several calculations
> is executed?
> For example,
> what happens when this cell is executed?
>
> ~~~
> 7 * 3
> 2 + 1
> ~~~
> {: .source}
{: .challenge}
> ## Change an Existing Cell from Code to Markdown
> What if you already wrote some code in into your *code cell* and then you switch it to a *markdown cell*?
> Give it a try; type the following into a cell in *Edit Mode*:
>
> ~~~
> x = 6 * 7 + 12
> print(x)
> ~~~
> {: .python}
>
> Now run the cell to be sure that it works as a code cell (shift + return).
>
> Now go back to the cell, press escape to make sure you're in *Control Mode*, and press the "M" key. Now try running the cell (shift + return).
>
> What happened and how might this be useful?
>
> As an added challenge, try changing back to a code cell by pressing the "Y" key from the *Control Mode*.
{: .challenge}
[anaconda]: https://docs.continuum.io/anaconda/install
| 40.153846 | 145 | 0.714347 | eng_Latn | 0.999366 |
913be5f0be284f34ce20eb8ba7118593436c4637 | 741 | md | Markdown | CHANGELOG.md | fizzed/java-jne | 783226a1fb002d304d22f841870c5c73575fc994 | [
"Apache-2.0"
] | null | null | null | CHANGELOG.md | fizzed/java-jne | 783226a1fb002d304d22f841870c5c73575fc994 | [
"Apache-2.0"
] | null | null | null | CHANGELOG.md | fizzed/java-jne | 783226a1fb002d304d22f841870c5c73575fc994 | [
"Apache-2.0"
] | null | null | null | Java Native Extractor by Fizzed
===============================
#### 3.0.1 - 2017-08-18
- Only create temp dir a single time per JVM instance
- Use UUID for temp dir
#### 3.0.0 - 2017-07-17
- Bump parent to v2.1.0
- Add ANY enum for OS
- New `findFile` feature to extract generic resources
- Initial unit tests
- Renamed Arch class to HardwareArchitecture
- Renamed OS class to OperatingSystem
#### 2.0.1 - 2016-08-23
- Verify temp dir exists (mpanthony)
#### 2.0.0 - 2015-12-18
- Refactored package to com.fizzed
- Added slf4j for logging
#### 1.2.0 - 2015-12-11
- Refactored project layout
#### 1.1.0 - 2014-04-16
- Added ability to rename extracted file (e.g. cat to mycat)
#### 1.0.1 - 2014-03-18
- Initial release
| 23.15625 | 61 | 0.647773 | eng_Latn | 0.789744 |
913c1bb5a49af1546ba459f16a0be9c57c8797af | 2,767 | md | Markdown | apidoc/README.md | gitizenme/titanium_mobile | f9ebb757a7b78cc18b331cacc266cc5b0a02835f | [
"Apache-2.0"
] | 2 | 2015-05-30T20:28:13.000Z | 2021-01-08T17:02:41.000Z | apidoc/README.md | arnaudsj/titanium_mobile | 4ed83dd6b355947a88f52efbf4ac82d86a2eeffd | [
"Apache-2.0"
] | 6 | 2015-04-27T22:12:58.000Z | 2020-05-23T01:14:06.000Z | apidoc/README.md | arnaudsj/titanium_mobile | 4ed83dd6b355947a88f52efbf4ac82d86a2eeffd | [
"Apache-2.0"
] | 1 | 2019-03-15T04:55:17.000Z | 2019-03-15T04:55:17.000Z | # TDoc: The Titanium API Documentation Format
_This documentation is a WIP_
The TDoc format follows a simple syntax for declaring Modules, Proxies, Methods, Properties, and Events for Titanium.
## Layout
The documentation tree starts in the Titanium folder, and generally follows this pattern:
<pre>
Titanium/
-- Module/
---- Module.tdoc
---- Proxy.tdoc
---- SubModule/
------ SubModule.tdoc
</pre>
## Format
A TDoc file basically consists of sections, and properties, and allows for free-form [Markdown](http://daringfireball.net/projects/markdown/syntax) in most places.
A section usually looks like:
<pre>- SECTION
section content</pre>
### Common sections
* namespace (full namespace of the object)
* type (module, proxy, view, property, method)
* description (a full length description)
* since (what version of Titanium this object was added in)
* platforms (android,iphone,ipad)
* notes (special notes for this object)
* example (named code examples)
### Proxy and Module sections
* methods (list of method names and descriptions)
* method (section for a single method with parameters and return type)
* properties (list of property names, types, and descriptions)
* events (events fired by this proxy or module)
* event (section for a single event with event object properties)
### Named sections
The method, event, and example sections all support embedded names, and in the case of the method section also supports a return type. Some quick examples:
<pre># returns void
- method : helloWorld
# returns string
- method : methodName, string
method documentation..
- event : eventName
event documentation..
- example : My Cool Example
<code>
// here's some code..
</code></pre>
### Section Properties
Section Properties generally take the form:
<pre>name[type]: value</pre>
### Method parameters
Method parameters are defined as properties under a Method section. Titanium types can be auto-link by surrounding the full API Name (using Titanium instead of Ti) using backticks. For example:
<pre>- method : format, string
formatString[string]: The format string
args[array,...]: The format arguments
- method : createTableView, `Titanium.UI.TableView`
returns a new table view
- method : updateRow
row[`Titanium.UI.TableViewRow`]: row to update</pre>
### Properties
Defined as properties under the Properties section, and can also auto-link Titanium types with backticks, For example:
<pre>- properties
name[string]: Name documentation
myView[`Titanium.UI.View`]: my view</pre>
## AutoLinking Titanium types
Except in the special cases of Properties and Method Parameters, you can generate a link to a Titanium type by surrounding it with double brackets, like so:
<pre>- description
Creates a [[Titanium.UI.TableView]]</pre>
| 31.089888 | 193 | 0.759668 | eng_Latn | 0.992044 |
913cc5179f86b964793fc14e1130d4ce66e802cb | 2,093 | md | Markdown | docs/internals/parameter-metadata.md | baileyherbert/reflection | bd161fa6ee32e296729f670b3fa915ea3b4361eb | [
"MIT"
] | 1 | 2021-12-13T18:06:31.000Z | 2021-12-13T18:06:31.000Z | docs/internals/parameter-metadata.md | baileyherbert/reflection | bd161fa6ee32e296729f670b3fa915ea3b4361eb | [
"MIT"
] | null | null | null | docs/internals/parameter-metadata.md | baileyherbert/reflection | bd161fa6ee32e296729f670b3fa915ea3b4361eb | [
"MIT"
] | null | null | null | # Parameter Metadata
## Introduction
Parameter metadata can be stored in countless ways. This reflection library uses a specific format which enables the
`getMetadata()` method to work on parameters.
If you are writing your own decorators, consider invoking the `Meta.Parameter` function like below to easily set
metadata in the correct structure.
```ts
import { Meta } from '@baileyherbert/reflection';
export function Decorator(target: Object, propertyKey: string | symbol, parameterIndex: number) {
Meta.Parameter('key', 'value')(target, propertyKey, parameterIndex);
}
```
If your decorator only needs to set metadata, you can also directly wrap the `Meta.Parameter` function like below.
```ts
import { Meta } from '@baileyherbert/reflection';
// With static data
export const Decorator = Meta.Parameter('key', 'value');
// With variable data
export const Decorator = (value: string) => Meta.Parameter('key', value);
```
The rest of this page will discuss how metadata is structured internally for this library in case of advanced usage.
## Storage structure
- Parameter metadata is stored on the parent method under the `#!ts "reflection:params"` key
- The value of this key must be `undefined` or an instance of `#!ts Array<Map<any, any>>`
- The array indices are the indices of the parameters
- The array values are `Map<any, any>` objects containing the metadata
Refer to the following code sample which sets metadata on a parameter just like the `@Meta.Parameter()` decorator:
## Example
```ts
function ParameterMeta(metadataKey: any, metadataValue: any) {
return function(target: Object, propertyKey: string | symbol, parameterIndex: number) {
let parameters = Reflect.getOwnMetadata('reflection:params', target, propertyKey);
if (parameters === undefined) {
parameters = new Array<Map<any, any>>();
Reflect.defineMetadata('reflection:params', parameters, target, propertyKey);
}
if (!(parameterIndex in parameters)) {
parameters[parameterIndex] = new Map();
}
parameters[parameterIndex].set(metadataKey, metadataValue);
};
}
```
| 33.758065 | 116 | 0.74343 | eng_Latn | 0.944263 |
913cd90eb4548fd2518d1d730eb3973ceab4c6e3 | 1,290 | md | Markdown | src/tech/erlang.md | joelwallis/log | e17f945ce253c3cc62ea215700e8de5c0b4955c8 | [
"0BSD"
] | 5 | 2022-01-21T00:43:50.000Z | 2022-02-14T21:47:42.000Z | src/tech/erlang.md | joelwallis/knowledge | d9fa6d957fd1641b12eb968300cd4e8ca5a18e1d | [
"0BSD"
] | null | null | null | src/tech/erlang.md | joelwallis/knowledge | d9fa6d957fd1641b12eb968300cd4e8ca5a18e1d | [
"0BSD"
] | null | null | null | # erlang ⓔ
My adventures on this amazing distributed computing platform.
## Erlang on macOS through asdf
Installing Erlang through asdf is probably the easiest way to get it up and running on macOS. You'll need OpenSSL to run the installation, and the easiest way to get it is through Homebrew:
```
brew install openssl
```
Homebrew's OpenSSL formula is keg-only, and it means that even after installed it won't be globally available in your system. [There's an useful note about it on asdf-erlang's README.md](https://github.com/asdf-vm/asdf-erlang#dealing-with-openssl-issues-on-macos), but I ended up with a command that would use `brew` to get the path of the OpenSSL installation:
```sh
# first, export this Kerl-specific variable with the following content
export KERL_CONFIGURE_OPTIONS="--without-javac --with-ssl=$(brew --prefix openssl@1.1)"
# then, install the Erlang version you want (e.g.: 22.1.5)
asdf install erlang 22.1.5
```
These options will disable Java-related features and point to the correct OpenSSL paths.
You'll find more information about setting up Erlang in your machine in the [Setup](https://adoptingerlang.org/docs/development/setup) chapter of the [Adopting Erlang](https://adoptingerlang.org) book (a must-read if you're adopting Erlang just now).
| 49.615385 | 361 | 0.771318 | eng_Latn | 0.99035 |
913ce20b781b5af252be1dc41ce8ef7b9729b82b | 4,971 | md | Markdown | _posts/2019-12-24-hackerrank.md | aSquare14/aSquare14.github.io | 740af53840bdc656bc0cbb3c722b069fc8e1c674 | [
"MIT"
] | 4 | 2019-08-17T21:05:14.000Z | 2021-02-23T20:04:19.000Z | _posts/2019-12-24-hackerrank.md | asquare14/aSquare14.github.io | 63a5ba98b80f062af12850da10c4a60701e80f33 | [
"MIT"
] | 3 | 2018-03-08T20:23:32.000Z | 2021-04-26T13:06:31.000Z | _posts/2019-12-24-hackerrank.md | aSquare14/aSquare14.github.io | 740af53840bdc656bc0cbb3c722b069fc8e1c674 | [
"MIT"
] | 8 | 2018-06-09T07:29:52.000Z | 2020-10-21T22:22:29.000Z | ---
title: "My 2019 Summer Internship at Hackerrank Bangalore"
layout: post
date: 2019-12-24 13:30
tag:
- Internship Experiences
category: blog
author: atibhi
description: Weekly Blogs
---
I had the opportunity to intern at Hackerrank, Bangalore during the summer of 2019 and in this blog I’d like to tell you about my experience ! For those of you who haven’t heard about Hackerrank, it is the market–leading technical inteview platform to identify and hire developers with the right skills. It also has a community website where you can practice problems, solve challenges and become a better developer. It has offices in Mountain View, London and Bengaluru :)
**Applying to Hackerrank**
I heard about the internship opening from a friend and applied through their careers page. They hire two kinds of interns, Technical Content Engineers(TCE) and Software Engineering(SWE) Interns. The TCEs are responsible for curating and making problems on algorithms, data structures etc for hackerrank while the SWEs work on DevOps, Back-end, Front-end etc. I applied for an SWE position and had an interview scheduled with the CTO of hackerrank Harishankaran Karunanidhi. After the interview, he said they would get back to me and within a week I had an offer from Hackerrank. I was really excited to join !

**The internship and the work culture**
The internship started sometime in May and we were given an introductory session where all the interns set up their laptops and finished some formalities. The next day we were assigned tasks and mentors. We were also told and go to interact with the Sales team so we would get to know about how the sales works. This was really fun because as developers we often do not understand how customers buy our software or how the whole marketing and sales department works !
The best thing about hackerrank is that there is no hierarchy, we never felt like interns, we were included in every event, discussion and could give our views and opinions freely. Moreover, we had a weekly meeting with Hari where we explained whatever work we did during that week to him AND also to all the other interns. This was really good as we could learn by listening to other interns as well. During the course of the internship, all the interns(10 of us) were put up in a hotel nearby the office. This helped us to foster great friendships and we helped each other grow technically by having many intellectual discussions . You can check out all the interns here.
**What I learnt during my internship**
The main tech stack of hackerrank is ruby on rails and react. I had already worked on ruby on rails in my previous internship, so I got the opportunity to work on React and Redux. My mentor was Aakansha Doshi, who is a front-end developer. I learnt so much from her, from the basics to best practices, finding test-cases etc. She also gave me feedback at the end of the internship which I am grateful for. I also worked with Shiv Deepak from the Mountain View office. I worked on an interesting DevOps related task with him. This was also a good experience as I had never worked on DevOps before nor had any idea about it. Apart from the technical learning, I also learnt about stand-ups,team work, how to communicate effectively, test driven development, to not be scared to ask for help but also ask for help only when we have tried enough.
**Fun times…**
Apart from work, we also had a lot of fun in the office. The office has a foosball table, table tennis room as well as XBox. We would play foosball whenever we wanted to take a break, infact on weekends we would come to office just to play ! The office also has a very well stocked pantry :P


Moreover, Hackerrank has tied up with KidsWhoKode and the office of HackerRank in Koramangala, Bengaluru is transformed into a classroom every friday. Underprivileged students visit to learn to code and are taught by the employees of this tech company. The interns volunteered and took turns to teach the children. Also,the full-time employees go out of their way to teach the children !
**Amazing summer came to an end…**
All good things must come to an end and so did this internship. I learnt a lot and made great friendships. I would like to thank everyone at Hackerrank for making this one of the best summers of my life !
I would like to thank Ishani, Nalini and Kenny from the HR team for taking great care of us , and Hari, Aakansha, Shiv and so many others for the amazing technical experience ! 😄
P.S — Highly recommend students to apply for an internship. Feel free to reach out to me if you have any questions :)
------
*[This blog post was originally published on [Medium](https://medium.com/@atibhiagrawal/my-2019-summer-internship-at-hackerrank-bengaluru-8f3163612c3a)]* | 97.470588 | 842 | 0.788171 | eng_Latn | 0.999894 |
913dcfd946ce37ed415fb0684c61d92e84cb4a37 | 4,433 | md | Markdown | README.md | mkxml/glc-tratamento-simbolos-inuteis | 4a8b089154b8a35cc2245f6271137e025de7c936 | [
"MIT"
] | null | null | null | README.md | mkxml/glc-tratamento-simbolos-inuteis | 4a8b089154b8a35cc2245f6271137e025de7c936 | [
"MIT"
] | null | null | null | README.md | mkxml/glc-tratamento-simbolos-inuteis | 4a8b089154b8a35cc2245f6271137e025de7c936 | [
"MIT"
] | 1 | 2020-04-22T14:35:00.000Z | 2020-04-22T14:35:00.000Z | Remoção dos símbolos inúteis em uma GLC
=======================================
Este é um pequeno projeto escrito em JavaScript em cima de Node.JS que
implementa um algoritmo de remoção de símbolos inúteis em uma gramática livre
do contexto (GLC).
O programa foi desenvolvido para a disciplina de
**Linguagens formais e Autômatos** da [Universidade Feevale](http://feevale.br).
## Como utilizar
**Para utilizar o programa é simples, basta seguir os passos:**
- Faça um `git clone` do projeto;
- Se você ainda não tem Node.JS instalado no seu
computador [instale-o](http://nodejs.com/download);
- Junto com o Node.JS foi instalado um software chamado NPM, precisamos dele;
- Abra seu Terminal (prompt de comando) e vá até o diretório do clone feito;
- Execute `npm install` e as dependências para o programa serão instaladas;
- Agora basta rodar o programa com `node app.js -i input.txt`;
**O programa suporta alguns parâmetros úteis para customizar seu funcionamento:**
- `-i` para indicar qual o arquivo TXT de input você quer utilizar;
- `-o` para indicar o arquivo a ser criado para a saída do programa em TXT por
padrão o software vai criar um arquivo chamado `output.txt` no diretório
local;
- `-f` para indicar o formato de quebra de linha, são suportados o `LF` do Unix
e o `CRLF` do Windows;
- `-h` exibe o help;
## Regras
Para ser compatível com o programa o layout deve respeitar algumas
regras, são elas:
- A primeira linha do input deve declarar as **variáveis** utilizadas na GLC;
- A segunda linha do input deve declarar os **terminais** utilizados na GLC;
- A terceira linha do input deve indicar a **varíavel inicial** da GLC;
- Da quarta linha em diante devem estar dispostas as **regras de produção** da GLC;
O programa, funcionando corretamente, deve salvar um outro TXT seguindo as mesmas
regras mostradas anteriormente.
### Exemplo de input
```
expr fator termo id nil
x y + * ( ) PI
expr
expr -> termo
expr -> expr + termo
termo -> fator
termo -> termo * fator
fator -> id
fator -> PI
fator -> ( expr )
id -> x
id -> y
nil ->
```
### Output
```
expr fator termo id
x y + * ( ) PI
expr
expr -> termo
expr -> expr + termo
termo -> fator
termo -> termo * fator
fator -> id
fator -> PI
fator -> ( expr )
id -> x
id -> y
```
## Tecnologias e estruturas usadas
O software é baseado em [Node.JS](http://nodejs.com) e usa a
[NPM](http://npmjs.com) para controle de dependências e versionamento.
É 100% escrito em JavaScript dentro das regras do projeto de não utilizar
funções de parsing da linguagem. Para conseguir fazer o parsing do input
fornecido utilizamos arrajnos (*arrays*) em cascata com loops validando cada
símbolo, separando-os pelo espaço em branco e quebra de linha.
Além disso utilizamos [Git](http://git-scm.com) para controle de versões do
código e hospedamos o projeto aqui no GitHub.
## Como funciona
O programa funciona em quatro etapas.
1) Primeiramente ele faz o parsing do TXT, separando cada linha em um vetor e depois
procura por espaços em branco separando palavras e coloca cada palavra em um célula
do vetor da linha.
O parser também separa os significados dos símbolos utilizando as regras definidas
acima. Criando variáveis para os grupos: **terminais**, **variáveis**, **símbolo inicial** e
**regras de produção**.
2) Depois o programa executa a primeira parte do algoritmo que garante que todas
as varáveis geram, direta ou indiretamente, terminais. Para fazer isso ele segue
o seguinte procedimento:
- 1) Considera `v1` todas as variáveis que geram terminais **diretamente**;
- 2) Junta ao `v1` todas as variáveis que geram alguma variável do conjunto `v1`;
- 3) Repete 2 até que a cardinalidade de `v1` não mude;
3) Aí o programa executa um segundo algoritmo, para garantir que todas as variáveis
`v1` são atingidas apartir do **símbolo inicial**. Para fazer isso ele faz o seguinte:
- 1) Considera `v2` todas as variáveis que são atingidas pelo **símbolo incial**;
- 2) Junta ao `t2` todas os terminais atingidos pelas variáveis do `v2`;
- 3) Junta ao `v2` todas as variáveis atingidas por `v2`;
- 4) Repete 2 e 3 até que as cardinalidades de `t2` e `v2` não mudem;
4) O último passo é a impressão do TXT de volta para o usuário, seguindo o mesmo
layout de entrada imposto.
## Autores
- [Matheus Kautzmann](http://github.com/mkautzmann)
- [Rômulo Alves](http://github.com/romuloalves)
## Licença
[MIT LICENSE](LICENSE)
| 33.330827 | 92 | 0.730205 | por_Latn | 0.999678 |
913e56ffe2d482687bad8d76adbc4c6166854441 | 1,292 | md | Markdown | ralbot/uvmgen/README.md | jiacaiyuan/uvm-generator | 63f4c7bd0dad43b357d1cc859b61011718c597f8 | [
"MIT"
] | 13 | 2020-04-15T09:11:53.000Z | 2022-03-13T02:05:53.000Z | ralbot/uvmgen/README.md | jiacaiyuan/uvm-generator | 63f4c7bd0dad43b357d1cc859b61011718c597f8 | [
"MIT"
] | null | null | null | ralbot/uvmgen/README.md | jiacaiyuan/uvm-generator | 63f4c7bd0dad43b357d1cc859b61011718c597f8 | [
"MIT"
] | 4 | 2020-11-27T08:11:24.000Z | 2022-02-19T09:11:36.000Z | # RALBot-uvm
Generate UVM register model from compiled SystemRDL input
## Installing(left blank)
Install from [PyPi](https://pypi.org/project/ralbot-uvm) using pip:
python3 -m pip install ralbot-uvm
--------------------------------------------------------------------------------
## Exporter Usage
Pass the elaborated output of the [SystemRDL Compiler](http://systemrdl-compiler.readthedocs.io)
to the exporter.
```python
import sys
from systemrdl import RDLCompiler, RDLCompileError
from ralbot.uvmgen import uvmGenExporter
rdlc = RDLCompiler()
try:
rdlc.compile_file("path/to/my.rdl")
root = rdlc.elaborate()
except RDLCompileError:
sys.exit(1)
file = "test.svh"
exporter = uvmGenExporter()
exporter.export(root, file)
```
--------------------------------------------------------------------------------
## Reference
### `uvmGenExporter(**kwargs)`
Constructor for the uvmGen exporter class
**Optional Parameters**
* `indentLvl`
* Three spaces or four spaces. Default three space.
### `uvmGenExporter.export(node, path)`
Perform the export!
**Parameters**
* `node`
* Top-level node to export. Can be the top-level `RootNode` or any internal `AddrmapNode`.
* `path`
* Output file. Can be (dir+filename without suffix. such as "output/test_uvmgen")
| 24.377358 | 96 | 0.640093 | eng_Latn | 0.334483 |
913f5c906602c5786f8954ae075278a2e8b5e72a | 16 | md | Markdown | README.md | inventioncorps/icb-website-2 | f42d8dbe80cb4bb0863b594121ec125e9aca8099 | [
"MIT"
] | 1 | 2021-08-29T17:24:08.000Z | 2021-08-29T17:24:08.000Z | README.md | inventioncorps/icb-website-2 | f42d8dbe80cb4bb0863b594121ec125e9aca8099 | [
"MIT"
] | 2 | 2020-12-07T07:21:27.000Z | 2020-12-08T08:21:11.000Z | README.md | inventioncorps/icb-website-2 | f42d8dbe80cb4bb0863b594121ec125e9aca8099 | [
"MIT"
] | null | null | null | # icb-website-2
| 8 | 15 | 0.6875 | kor_Hang | 0.191865 |
9140c4a7bccab176e0cd360ee21ee8c5bc08551e | 1,040 | md | Markdown | docs/goto.md | codemedic/bash-ninja | 133c2d8e23c09770618ed0318b37380c35e8b9f8 | [
"MIT"
] | 9 | 2018-02-15T03:06:48.000Z | 2020-09-21T11:35:13.000Z | docs/goto.md | codemedic/bash-ninja | 133c2d8e23c09770618ed0318b37380c35e8b9f8 | [
"MIT"
] | null | null | null | docs/goto.md | codemedic/bash-ninja | 133c2d8e23c09770618ed0318b37380c35e8b9f8 | [
"MIT"
] | 1 | 2019-05-18T07:23:02.000Z | 2019-05-18T07:23:02.000Z | ## `goto` - bookmarks for the shell
`goto` is a path bookmark utility that would help you navigate within the filesystem using bookmarks. It comes with auto-completion and bookmark addition command `goto_add`.
Once you `cd` yourself into a path, you can run `goto_add bookmarkName` to add the current working directory into the bookmark. Once added, you can use `goto bookmarkName` to `cd` into the location, from elsewhere.
The bookmark-name as well as any sub-directories under the location can be auto-completed by the usual bash means. The bookmark-name and its subpath(s) has to be separated by `#`. You can also go a level above (and auto-completed) using `goto bookmarkName#../`.
See [`go_bookmarks.conf`](go_bookmarks.conf) for some examples. The config is re-read and executed each time you invoke `goto` or the auto-completion, so that the changes are instantaneous.
In order to install this into your profile, edit your `.bashrc` and add the below lines.
go_projects_conf=$HOME/go_bookmarks.conf
source $HOME/go.sh
| 69.333333 | 261 | 0.770192 | eng_Latn | 0.999494 |
9140fb44218cc8fd1923ae736ee3241d3d884fca | 501 | md | Markdown | README.md | albovy/PadelGest | 32d464cbb3aefa85450b41d67d0212c8c10efc9d | [
"MIT"
] | null | null | null | README.md | albovy/PadelGest | 32d464cbb3aefa85450b41d67d0212c8c10efc9d | [
"MIT"
] | null | null | null | README.md | albovy/PadelGest | 32d464cbb3aefa85450b41d67d0212c8c10efc9d | [
"MIT"
] | null | null | null | # PADELGEST
## Instalación
### NodeJS
Descarga de NodeJs a partir del siguiente enlace:
<https://nodejs.org/en/download/>
### ExpressJS
Instalación de ExpressJS como dependencia utilizando el gestor de paquetes
` cd PadelGest `
`npm install express -save`
### MongoDB
Descarga de MongoDB a partir del siguiente enlace:
<https://docs.mongodb.com/manual/administration/install-community/>
function test() {
console.log("notice the blank line before this function?");
}
´´´
| 11.651163 | 75 | 0.716567 | spa_Latn | 0.573129 |
91417cb3dced2d77381fcf7297dab7498abc5e41 | 1,355 | md | Markdown | README.md | net2cn/Real-ESRGAN_GUI | 2190f499345546293c6fca0f3de09f747753781b | [
"MIT"
] | 16 | 2021-12-30T05:31:24.000Z | 2022-03-30T13:23:39.000Z | README.md | net2cn/Real-ESRGAN_GUI | 2190f499345546293c6fca0f3de09f747753781b | [
"MIT"
] | 1 | 2022-03-30T10:07:07.000Z | 2022-03-30T11:51:20.000Z | README.md | net2cn/Real-ESRGAN_GUI | 2190f499345546293c6fca0f3de09f747753781b | [
"MIT"
] | 1 | 2022-03-20T04:59:59.000Z | 2022-03-20T04:59:59.000Z | # Real-ESRGAN_GUI
A C# GUI inference implementation of [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN).
PRs are welcomed.
---
## Usage
You know how a GUI works.

## Result
From  to  with the magic of [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN).
## Build Prerequisites
- Visual Studio 2019 or higher.
## Known Issue
- GPU support is not working.
- Directory input is not implemented yet.
- Alpha channel will be ignored.
- Huge memory consumption when handling large image (~1000x1000, eats up ~18.5G memory easily).
## Acknowledgements
This repository contains ONNX models converted from [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo. All copyrights and trademarks of the materials used belong to their respective owners and are not being sold.
This repository is created only for learning purpose. I DO NOT take any responsibilities for any possible damages.
Image [upscale example](./assets/avatar_256px.png) and [result](./assets/avatar_256px_realesrgan-x4plus_anime_6B.png) attached in assets folder are derivatives of my personal artwork for my own SNS avatar. Please do not use without permission, especially for commercial purposes.
---
2021, net2cn. | 41.060606 | 279 | 0.77786 | eng_Latn | 0.958297 |
91423960a711564dc3ae76ec1bdaf8f4765152ae | 125 | md | Markdown | README.md | krismuniz/pr-geo | 7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea | [
"Unlicense"
] | null | null | null | README.md | krismuniz/pr-geo | 7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea | [
"Unlicense"
] | null | null | null | README.md | krismuniz/pr-geo | 7a2eb049da313b8e483c4965aa7eb27ac5fdc7ea | [
"Unlicense"
] | null | null | null | # pr-geo
Data collection of Puerto Rico's Political geography.
Colección de datos de la geografía política de Puerto Rico.
| 25 | 59 | 0.792 | spa_Latn | 0.629213 |
914261f77470639a9c6c9be24e11a27e093872b5 | 6,068 | md | Markdown | articles/supply-chain/asset-management/setup-for-objects/counters.md | MicrosoftDocs/Dynamics-365-Operations.fr-fr | 9f97b0553ee485dfefc0a57ce805f740f4986a7e | [
"CC-BY-4.0",
"MIT"
] | 2 | 2020-05-18T17:14:08.000Z | 2021-04-20T21:13:46.000Z | articles/supply-chain/asset-management/setup-for-objects/counters.md | MicrosoftDocs/Dynamics-365-Operations.fr-fr | 9f97b0553ee485dfefc0a57ce805f740f4986a7e | [
"CC-BY-4.0",
"MIT"
] | 6 | 2017-12-13T18:31:58.000Z | 2019-04-30T11:46:19.000Z | articles/supply-chain/asset-management/setup-for-objects/counters.md | MicrosoftDocs/Dynamics-365-Operations.fr-fr | 9f97b0553ee485dfefc0a57ce805f740f4986a7e | [
"CC-BY-4.0",
"MIT"
] | 1 | 2019-10-12T18:19:20.000Z | 2019-10-12T18:19:20.000Z | ---
title: Mesures d’actifs
description: La rubrique explique comment créer des types de mesures d’actifs dans le module Gestion des actifs.
author: johanhoffmann
ms.date: 10/15/2019
ms.topic: article
ms.prod: ''
ms.technology: ''
ms.search.form: EntAssetObjectCounterPart, EntAssetObjectCounterLookup, EntAssetCounterType, EntAssetObjectCounterTotals
audience: Application User
ms.reviewer: kamaybac
ms.custom: ''
ms.assetid: ''
ms.search.region: Global
ms.author: johanho
ms.search.validFrom: 2019-09-30
ms.dyn365.ops.version: 10.0.5
ms.openlocfilehash: fc515615afaa172e1832508d79e202b166f134a9171a0a35ea4f372f9d19b7e2
ms.sourcegitcommit: 42fe9790ddf0bdad911544deaa82123a396712fb
ms.translationtype: HT
ms.contentlocale: fr-FR
ms.lasthandoff: 08/05/2021
ms.locfileid: "6723573"
---
# <a name="counters"></a>Compteurs
[!include [banner](../../includes/banner.md)]
La rubrique explique comment créer des types de compteurs dans le module Gestion des actifs. Les types de compteurs permettent de faire des enregistrements de compteurs sur les actifs, par exemple en ce qui concerne le nombre d’heures de production ou la quantité produite sur l’actif. Les types d’actifs sont liés aux types de compteurs. Cela signifie qu’un compteur ne peut être utilisé sur un actif que si le compteur est configuré sur le type d’actif utilisé sur l’actif.
Avant de pouvoir effectuer des enregistrements de compteurs, commencez par créer les types de compteurs à utiliser dans **Compteurs**. Ensuite, vous pouvez créer des enregistrements de compteurs sur les actifs dans **Compteurs**.
Les compteurs peuvent être utilisés sur les plans de maintenance. Une ligne de plan de maintenance peut être de type « compteur », par exemple, en relation avec le nombre d’heures ou la quantité de production produite.
Un enregistrement de compteur peut être mis à jour manuellement ou automatiquement en fonction des heures ou de la quantité de production produite. Un compteur peut être paramétré pour utiliser une des trois méthodes de mise à jour (que vous sélectionnez dans le champ **Mettre à jour** dans **Compteurs**) :
- Manuel - Vous devez enregistrer manuellement des valeurs de compteurs.
- Heures de production - le compteur est automatiquement mis à jour en fonction du nombre d’heures de production.
- Quantité de production - le compteur est automatiquement mis à jour en fonction des quantités produites.
>[!NOTE]
>Si la quantité produite est utilisée, *tous* les articles enregistrés sont inclus dans l’enregistrement de compteurs, bonne quantité, ainsi que quantité d’erreur. Il reste possible d’effectuer des enregistrements de compteurs manuels, le cas échéant.
## <a name="create-counter-types-for-asset-counter-registrations"></a>Créez les contre-types pour les enregistrements de compteurs d’actifs
1. Sélectionnez **Gestion des actifs** > **Paramétrage** > **Types d’actifs** > **Compteurs**.
2. Sélectionnez **Nouveau** pour créer un type de compteur.
3. Insérez un ID dans le champ **Compteur**, et un nom de compteur dans le champ **Nom**.
4. Dans l’organisateur **Général**, sélectionnez un compteur dans le champ **Unité**.
5. Dans le champ **Mettre à jour**, sélectionnez la méthode de mise à jour à utiliser pour le compteur.
6. Sélectionnez « Oui » sur le bouton de basculement **Hériter des contre-valeurs** si les actifs enfants d’une structure d’actif doivent hériter automatiquement des enregistrements de compteurs effectués sur l’actif parent.
7. Dans le champ **Agrégat total**, sélectionnez la méthode de synthèse à utiliser pour un compteur à l’aide de ce type de compteur. La « somme » est la sélection standard utilisée pour ajouter en continu des valeurs enregistrées à la valeur totale. La « moyenne » peut être utilisée si un compteur est configuré pour surveiller un seuil, par exemple en ce qui concerne la température, les vibrations ou l’usure d’un actif.
8. Dans le champ **Écart supérieur à**, insérez le niveau supérieur en pourcentage pour vérifier si les enregistrements du compteur manuel se situent dans une plage prévue. La validation est basée sur une augmentation linéaire du nombre d’enregistrements existants de compteurs.
9. Dans le champ **Écart inférieur à**, insérez le niveau inférieur en pourcentage pour vérifier si les enregistrements du compteur manuel se situent dans une plage prévue. La validation est basée sur une diminution linéaire du nombre d’enregistrements existants de compteurs.
10. Dans le champ **Type**, sélectionnez le type de message (informations, avertissement, erreur) à afficher si des écarts en dehors de la plage définie se produisent lorsque vous enregistrez manuellement les compteurs.
11. Dans l’organisateur **Types d’actifs**, ajoutez des types d’actifs qui doivent pouvoir utiliser le compteur.
12. Dans l’organisateur **Compteurs d’actifs associés**, ajoutez le compteur que vous souhaitez mettre à jour automatiquement lorsque ce compteur est mis à jour.
>[!NOTE]
>Un compteur associé est automatiquement mis à jour uniquement si le compteur associé a le type d’actif auquel elle est liée dans la configuration du compteur. Par exemple : Vous paramétrez un compteur pour les « heures de production » et ajoutez le type « Moteur de camion ». Lorsque ce compteur est mis à jour, un compteur associé à « Huile » est également mis à jour avec les mêmes valeurs de compteurs. Le paramétrage dans **Compteurs** inclut le paramétrage des « Heures ». En outre, pour le compteur « Pétrole », le type d’actif « Moteur de camion » doit être ajouté dans l’organisateur **Types d’actifs** pour garantir la relation de compteur. Voir les captures d’écran ci-dessous pour obtenir un exemple du paramétrage des compteurs Heures et Pétrole.
Lorsque des types d’actifs sont ajoutés à un type de compteur dans **Compteurs**, ce compteur est automatiquement ajouté aux types d’actifs dans l’organisateur **Compteurs** dans [Types d’actifs](../setup-for-objects/object-types.md).

[!INCLUDE[footer-include](../../../includes/footer-banner.md)] | 87.942029 | 759 | 0.786915 | fra_Latn | 0.996779 |
9143286e29be442a8892d1dabb94fee74437b28b | 28,208 | md | Markdown | articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md | YutongTie-MSFT/azure-docs.de-de | f7922d4a0ebfb2cbb31d7004d4f726202f39716b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md | YutongTie-MSFT/azure-docs.de-de | f7922d4a0ebfb2cbb31d7004d4f726202f39716b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/azure-monitor/platform/diagnostics-extension-stream-event-hubs.md | YutongTie-MSFT/azure-docs.de-de | f7922d4a0ebfb2cbb31d7004d4f726202f39716b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Streamen von Azure-Diagnosedaten an Event Hubs
description: Erfahren Sie, wie Sie eine End-to-End-Konfiguration von Azure-Diagnose mit Event Hubs durchführen. Außerdem werden Anleitungen für gängige Szenarien bereitgestellt.
services: azure-monitor
author: rboucher
ms.service: azure-monitor
ms.devlang: dotnet
ms.topic: conceptual
ms.date: 07/13/2017
ms.author: robb
ms.subservice: diagnostic-extension
ms.openlocfilehash: c2d577bd4c89046136a3465ff554e9662dd0ce19
ms.sourcegitcommit: cf88cf2cbe94293b0542714a98833be001471c08
ms.translationtype: HT
ms.contentlocale: de-DE
ms.lasthandoff: 01/23/2019
ms.locfileid: "54478126"
---
# <a name="streaming-azure-diagnostics-data-in-the-hot-path-by-using-event-hubs"></a>Streamen von Azure-Diagnosedaten im heißen Pfad mithilfe von Event Hubs
Die Azure-Diagnose bietet flexible Methoden zum Erfassen von Metriken und Protokollen virtueller Computer in Clouddiensten sowie zur Übertragung der Ergebnisse an Azure Storage. Seit März 2016 (SDK 2.9) haben Sie die Möglichkeit, benutzerdefinierte Datenquellen als Sendeziel für Diagnose zu verwenden und heiße Daten mithilfe von [Azure Event Hubs](https://azure.microsoft.com/services/event-hubs/) innerhalb von Sekunden zu übertragen.
Zu den unterstützten Datentypen gehören:
* Ereignisablaufverfolgung für Windows-Ereignisse (ETW)
* Leistungsindikatoren
* Windows-Ereignisprotokolle
* Anwendungsprotokolle
* Infrastrukturprotokolle der Azure-Diagnose
In diesem Artikel erfahren Sie, wie Sie eine End-to-End-Konfiguration der Azure-Diagnose mit Event Hubs durchführen. Zusätzlich finden Sie Anleitungen für die folgenden gängigen Szenarien:
* Anpassen der Protokolle und Metriken, die an Event Hubs gesendet werden
* Ändern der Konfigurationen in den einzelnen Umgebungen
* Anzeigen von Event Hubs-Streamdaten
* Behandlung von Verbindungsproblemen
## <a name="prerequisites"></a>Voraussetzungen
Die Verwendung von Event Hubs für den Empfang von Daten von Azure-Diagnose wird von Cloud Services, VMs, VM-Skalierungsgruppen und Service Fabric unterstützt. Die Unterstützung beginnt mit Azure SDK 2.9 und den entsprechenden Azure-Tools für Visual Studio.
* Azure-Diagnoseerweiterung 1.6 (standardmäßiges Ziel im[Azure SDK für .NET 2.9 oder höher](https://azure.microsoft.com/downloads/) )
* [Visual Studio 2013 oder höher](https://www.visualstudio.com/downloads/download-visual-studio-vs.aspx)
* Vorhandene Konfigurationen der Azure-Diagnose in einer Anwendung mithilfe einer *.wadcfgx* -Datei und einer der folgenden Methoden:
* Visual Studio: [Konfigurieren der Diagnose für Azure Cloud Services und Virtual Machines](/visualstudio/azure/vs-azure-tools-diagnostics-for-cloud-services-and-virtual-machines)
* Windows PowerShell: [Aktivieren der Diagnose mithilfe von PowerShell in Azure Cloud Services](../../cloud-services/cloud-services-diagnostics-powershell.md)
* Es muss ein Event Hubs-Namespace bereitgestellt werden, wie im Artikel [Erste Schritte mit Event Hubs](../../event-hubs/event-hubs-dotnet-standard-getstarted-send.md)
## <a name="connect-azure-diagnostics-to-event-hubs-sink"></a>Herstellen einer Verbindung zwischen der Azure-Diagnose und der Event Hubs-Senke
Azure-Diagnose sendet Protokolle und Metriken standardmäßig stets an ein Azure Storage-Konto. Eine Anwendung kann auch Daten an Event Hubs senden, indem der *WADCFGX*-Datei im **PublicConfig** / **WadCfg**-Element ein neuer **Sinks**-Abschnitt hinzugefügt wird. In Visual Studio wird die *WADCFGX*-Datei im folgenden Pfad gespeichert: **Cloud Service-Projekt** > **Rollen** > **(RoleName)** > **diagnostics.wadcfgx**.
```xml
<SinksConfig>
<Sink name="HotPath">
<EventHub Url="https://diags-mycompany-ns.servicebus.windows.net/diageventhub" SharedAccessKeyName="SendRule" />
</Sink>
</SinksConfig>
```
```JSON
"SinksConfig": {
"Sink": [
{
"name": "HotPath",
"EventHub": {
"Url": "https://diags-mycompany-ns.servicebus.windows.net/diageventhub",
"SharedAccessKeyName": "SendRule"
}
}
]
}
```
In diesem Beispiel wird die Event Hub-URL auf den vollqualifizierten Namespace des Event Hubs festgelegt: Event Hubs-Namespace + „/“ + Event Hub-Name.
Die Event Hub-URL wird im [Azure-Portal](https://go.microsoft.com/fwlink/?LinkID=213885) auf dem Event Hubs-Dashboard angezeigt.
Der Name der **Senke** kann auf eine beliebige gültige Zeichenfolge festgelegt werden, solange der Wert in der gesamten Konfigurationsdatei konsistent verwendet wird.
> [!NOTE]
> In diesem Abschnitt können zusätzliche Senken konfiguriert werden, beispielsweise *applicationInsights* . Für die Azure-Diagnose können mehrere Senken definiert werden, wenn jede Senke ebenfalls im Abschnitt **PrivateConfig** deklariert ist.
>
>
Die Event Hubs-Senke muss ebenfalls im Abschnitt **PrivateConfig** der *.wadcfgx* -Konfigurationsdatei deklariert und definiert werden.
```XML
<PrivateConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
<StorageAccount name="{account name}" key="{account key}" endpoint="{optional storage endpoint}" />
<EventHub Url="https://diags-mycompany-ns.servicebus.windows.net/diageventhub" SharedAccessKeyName="SendRule" SharedAccessKey="{base64 encoded key}" />
</PrivateConfig>
```
```JSON
{
"storageAccountName": "{account name}",
"storageAccountKey": "{account key}",
"storageAccountEndPoint": "{optional storage endpoint}",
"EventHub": {
"Url": "https://diags-mycompany-ns.servicebus.windows.net/diageventhub",
"SharedAccessKeyName": "SendRule",
"SharedAccessKey": "{base64 encoded key}"
}
}
```
Der `SharedAccessKeyName`-Wert muss einem SAS-Schlüssel (Shared Access Signature) und einer Richtlinie entsprechen, die im **Event Hubs**-Namespace definiert wurde. Navigieren Sie zum Event Hubs-Dashboard im [Azure-Portal](https://portal.azure.com), klicken Sie auf die Registerkarte **Konfigurieren**, und richten Sie eine benannte Richtlinie (z.B. „SendRule“) mit Berechtigungen zum *Senden* ein. Das **StorageAccount** wird ebenfalls in **PrivateConfig** konfiguriert. Wenn die Werte hier funktionieren, muss nichts geändert werden. In diesem Beispiel lassen wir die Werte leer. Dies bedeutet, dass ein Downstream-Asset später die Werte festlegt. Beispielsweise legt die Umgebungskonfigurationsdatei *ServiceConfiguration.Cloud.cscfg* die geeigneten Namen und Schlüssel für die Umgebung fest.
> [!WARNING]
> Der Event Hubs-SAS-Schlüssel wird im Nur-Text-Format in der *.wadcfgx* -Datei gespeichert. Häufig wird dieser Schlüssel in der Quellcodeverwaltung eingecheckt oder ist als Asset in Ihrem Buildserver verfügbar, deshalb sollten die Daten in geeigneter Weise geschützt werden. Es wird empfohlen, hier einen SAS-Schlüssel mit der Berechtigung *Nur senden* zu verwenden, sodass böswillige Benutzer Schreibvorgänge im Event Hub durchführen können, aber nicht am Event Hub lauschen und den Event Hub nicht verwalten können.
>
>
## <a name="configure-azure-diagnostics-to-send-logs-and-metrics-to-event-hubs"></a>Konfigurieren von Azure-Diagnose für das Senden von Protokollen und Metriken an Event Hubs
Wie bereits erläutert, werden alle standardmäßigen und benutzerdefinierten Diagnosedaten (d.h. Metriken und Protokolle) in den konfigurierten Intervallen automatisch an Azure Storage gesendet. Mit Event Hubs und allen zusätzlichen Senken können Sie einen beliebigen Stamm- oder Blattknoten in der Hierarchie an Event Hub senden. Hierzu gehören ETW-Ereignisse, Leistungsindikatoren, Windows-Ereignisprotokolle und Anwendungsprotokolle.
Es ist wichtig abzuwägen, wie viele Datenpunkte tatsächlich nach Event Hubs übertragen werden sollten. In der Regel übertragen Entwickler Daten mit niedriger Latenz im heißen Pfad, die schnell genutzt und interpretiert werden müssen. Beispiele dafür sind Systeme, die Warnungen oder Regeln für die automatische Skalierung überwachen. Entwickler können auch einen alternativen Analyse- oder Suchspeicher konfigurieren – beispielsweise Azure Stream Analytics, ElasticSearch, ein benutzerdefiniertes Überwachungssystem oder ein bevorzugtes Überwachungssystem eines Drittanbieters.
Nachfolgend sehen Sie einige Beispielkonfigurationen.
```xml
<PerformanceCounters scheduledTransferPeriod="PT1M" sinks="HotPath">
<PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\Bytes Total/Sec" sampleRate="PT3M" />
</PerformanceCounters>
```
```JSON
"PerformanceCounters": {
"scheduledTransferPeriod": "PT1M",
"sinks": "HotPath",
"PerformanceCounterConfiguration": [
{
"counterSpecifier": "\\Processor(_Total)\\% Processor Time",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Memory\\Available MBytes",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec",
"sampleRate": "PT3M"
}
]
}
```
Im vorangehenden Beispiel wird die Senke auf den übergeordneten **PerformanceCounters**-Knoten in der Hierarchie angewendet. Dies bedeutet, dass alle untergeordneten **PerformanceCounters** an Event Hubs gesendet werden.
```xml
<PerformanceCounters scheduledTransferPeriod="PT1M">
<PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Queued" sampleRate="PT3M" sinks="HotPath" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Rejected" sampleRate="PT3M" sinks="HotPath"/>
<PerformanceCounterConfiguration counterSpecifier="\Processor(_Total)\% Processor Time" sampleRate="PT3M" sinks="HotPath"/>
</PerformanceCounters>
```
```JSON
"PerformanceCounters": {
"scheduledTransferPeriod": "PT1M",
"PerformanceCounterConfiguration": [
{
"counterSpecifier": "\\Processor(_Total)\\% Processor Time",
"sampleRate": "PT3M",
"sinks": "HotPath"
},
{
"counterSpecifier": "\\Memory\\Available MBytes",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\ASP.NET\\Requests Rejected",
"sampleRate": "PT3M",
"sinks": "HotPath"
},
{
"counterSpecifier": "\\ASP.NET\\Requests Queued",
"sampleRate": "PT3M",
"sinks": "HotPath"
}
]
}
```
Im vorherigen Beispiel wird die Senke nur auf drei Indikatoren angewendet: **Anforderungen in Warteschlange**, **Zurückgewiesene Anforderungen** und **Prozessorzeit (%)**.
Das folgende Beispiel zeigt, wie ein Entwickler die Menge der gesendeten Daten auf die kritischen Metriken für die Dienstintegrität einschränken kann.
```XML
<Logs scheduledTransferPeriod="PT1M" sinks="HotPath" scheduledTransferLogLevelFilter="Error" />
```
```JSON
"Logs": {
"scheduledTransferPeriod": "PT1M",
"scheduledTransferLogLevelFilter": "Error",
"sinks": "HotPath"
}
```
In diesem Beispiel wird die Senke auf Protokolle angewendet und auf die Ablaufverfolgung auf Fehlerebene gefiltert.
## <a name="deploy-and-update-a-cloud-services-application-and-diagnostics-config"></a>Bereitstellen und Aktualisieren einer Cloud Services-Anwendung und der Diagnosekonfiguration
Visual Studio bietet die einfachste Methode, um die Anwendung und die Event Hubs-Senkenkonfiguration bereitzustellen. Öffnen Sie zum Anzeigen und Bearbeiten der Datei die *.wadcfgx* -Datei in Visual Studio, und bearbeiten und speichern Sie sie. Der Pfad lautet **Cloud Services-Projekt** > **Rollen** > **(RoleName)** > **diagnostics.wadcfgx**.
Zu diesem Zeitpunkt ist **.wadcfgx** im Verpackungsprozess aller Aktionen für die Bereitstellung und Bereitstellungsaktualisierung in Visual Studio, Visual Studio Team System sowie in allen Befehlen oder Skripts enthalten, die auf MSBuild basieren und das Ziel */t:publish* verwenden. Darüber hinaus wird die Datei bei allen Bereitstellungen und Aktualisierungen mit der entsprechenden Agent-Erweiterung der Azure-Diagnose auf Ihren virtuellen Computern in Azure bereitgestellt.
Nach der Bereitstellung der Anwendung und der Konfiguration der Azure-Diagnose wird die Aktivität sofort auf dem Dashboard des Event Hubs angezeigt. Anschließend können Sie die heißen Daten im Listenerclient oder in einem Analysetool Ihrer Wahl anzeigen.
In der folgenden Abbildung zeigt das Event Hubs-Dashboard, dass Diagnosedaten nach 23:00 Uhr ohne Fehler an den Event Hub gesendet wurden. Zu dieser Zeit wurde die Anwendung mit einer aktualisierten *.wadcfgx* -Datei bereitgestellt, und die Senke wurde ordnungsgemäß konfiguriert.
![][0]
> [!NOTE]
> Wenn Sie an der Konfigurationsdatei für die Azure-Diagnose (.wadcfgx) Änderungen vornehmen, wird empfohlen, die Aktualisierungen per Push an die gesamte Anwendung sowie die Konfiguration zu übertragen. Verwenden Sie hierzu entweder eine Veröffentlichung über Visual Studio oder ein Windows PowerShell-Skript.
>
>
## <a name="view-hot-path-data"></a>Anzeigen von Daten im heißen Pfad
Wie bereits erläutert, gibt es viele Anwendungsfälle für das Überwachen und Verarbeiten von Event Hubs-Daten.
Ein einfacher Ansatz ist das Erstellen einer kleinen Testkonsolenanwendung, die am Event Hub lauscht und den Ausgabestream druckt. Sie können den folgenden Code (dieser wird in [Erste Schritte mit Event Hubs](../../event-hubs/event-hubs-dotnet-standard-getstarted-send.md) ausführlicher erläutert) in einer Konsolenanwendung platzieren.
Beachten Sie, dass die Konsolenanwendung das [NuGet-Paket EventProcessorHost](https://www.nuget.org/packages/Microsoft.Azure.ServiceBus.EventProcessorHost/) umfassen muss.
Denken Sie daran, die Werte in spitzen Klammern in der **Main** -Funktion durch Werte für Ihre Ressourcen zu ersetzen.
```csharp
//Console application code for EventHub test client
using System;
using System.Collections.Generic;
using System.Diagnostics;
using System.Linq;
using System.Text;
using System.Threading.Tasks;
using Microsoft.ServiceBus.Messaging;
namespace EventHubListener
{
class SimpleEventProcessor : IEventProcessor
{
Stopwatch checkpointStopWatch;
async Task IEventProcessor.CloseAsync(PartitionContext context, CloseReason reason)
{
Console.WriteLine("Processor Shutting Down. Partition '{0}', Reason: '{1}'.", context.Lease.PartitionId, reason);
if (reason == CloseReason.Shutdown)
{
await context.CheckpointAsync();
}
}
Task IEventProcessor.OpenAsync(PartitionContext context)
{
Console.WriteLine("SimpleEventProcessor initialized. Partition: '{0}', Offset: '{1}'", context.Lease.PartitionId, context.Lease.Offset);
this.checkpointStopWatch = new Stopwatch();
this.checkpointStopWatch.Start();
return Task.FromResult<object>(null);
}
async Task IEventProcessor.ProcessEventsAsync(PartitionContext context, IEnumerable<EventData> messages)
{
foreach (EventData eventData in messages)
{
string data = Encoding.UTF8.GetString(eventData.GetBytes());
Console.WriteLine(string.Format("Message received. Partition: '{0}', Data: '{1}'",
context.Lease.PartitionId, data));
foreach (var x in eventData.Properties)
{
Console.WriteLine(string.Format(" {0} = {1}", x.Key, x.Value));
}
}
//Call checkpoint every 5 minutes, so that worker can resume processing from 5 minutes back if it restarts.
if (this.checkpointStopWatch.Elapsed > TimeSpan.FromMinutes(5))
{
await context.CheckpointAsync();
this.checkpointStopWatch.Restart();
}
}
}
class Program
{
static void Main(string[] args)
{
string eventHubConnectionString = "Endpoint= <your connection string>”;
string eventHubName = "<Event hub name>";
string storageAccountName = "<Storage account name>";
string storageAccountKey = "<Storage account key>”;
string storageConnectionString = string.Format("DefaultEndpointsProtocol=https;AccountName={0};AccountKey={1}", storageAccountName, storageAccountKey);
string eventProcessorHostName = Guid.NewGuid().ToString();
EventProcessorHost eventProcessorHost = new EventProcessorHost(eventProcessorHostName, eventHubName, EventHubConsumerGroup.DefaultGroupName, eventHubConnectionString, storageConnectionString);
Console.WriteLine("Registering EventProcessor...");
var options = new EventProcessorOptions();
options.ExceptionReceived += (sender, e) => { Console.WriteLine(e.Exception); };
eventProcessorHost.RegisterEventProcessorAsync<SimpleEventProcessor>(options).Wait();
Console.WriteLine("Receiving. Press enter key to stop worker.");
Console.ReadLine();
eventProcessorHost.UnregisterEventProcessorAsync().Wait();
}
}
}
```
## <a name="troubleshoot-event-hubs-sinks"></a>Beheben von Problemen mit Event Hubs-Senken
* Der Event Hub zeigt keine eingehende oder ausgehende Ereignisaktivität.
Überprüfen Sie, ob der Event Hub erfolgreich bereitgestellt wurde. Alle Verbindungsinformationen im Abschnitt **PrivateConfig** der *.wadcfgx* -Datei müssen mit den Werten Ihrer Ressourcen übereinstimmen, wie sie im Portal angezeigt werden. Stellen Sie sicher, dass Sie eine SAS-Richtlinie im Portal definiert haben (im Beispiel „SendRule“) und die Berechtigung *Senden* gewährt wurde.
* Nach einer Aktualisierung zeigt der Event Hub keine eingehende oder ausgehende Ereignisaktivität mehr.
Vergewissern Sie sich zunächst, dass die Event Hub- und Konfigurationsinformationen richtig sind (wie zuvor beschrieben). Manchmal wird **PrivateConfig** bei einer Bereitstellungsaktualisierung zurückgesetzt. Zur Beseitigung des Problems wird empfohlen, alle Änderungen an der *.wadcfgx* -Datei im Projekt durchzuführen und dann ein vollständiges Anwendungsupdate per Push zu übertragen. Wenn dies nicht möglich ist, stellen Sie sicher, dass mit dem Diagnoseupdate eine vollständige **PrivateConfig** einschließlich des SAS-Schlüssels per Push übertragen wird.
* Ich habe die empfohlenen Schritte ausgeführt, aber der Event Hub funktioniert weiterhin nicht.
Überprüfen Sie die Azure Storage-Tabelle, die Protokolle und Fehler zur Azure-Diagnose selbst enthält: **WADDiagnosticInfrastructureLogsTable**. Sie können hierzu beispielsweise mit dem Tool [Azure Storage-Explorer](https://www.storageexplorer.com) eine Verbindung mit diesem Speicherkonto herstellen, diese Tabelle anzeigen und eine Abfrage nach einem TimeStamp innerhalb der letzten 24 Stunden hinzufügen. Sie können das Tool verwenden, um eine CSV-Datei zu exportieren und in einer Anwendung wie Microsoft Excel zu öffnen. Excel erleichtert das Suchen nach Callingcard-Zeichenfolgen wie **EventHubs**, um festzustellen, welcher Fehler gemeldet wird.
## <a name="next-steps"></a>Nächste Schritte
• [Weitere Informationen zu Event Hubs](https://azure.microsoft.com/services/event-hubs/)
## <a name="appendix-complete-azure-diagnostics-configuration-file-wadcfgx-example"></a>Anhang: Beispiel einer vollständigen Konfigurationsdatei für die Azure-Diagnose (.wadcfgx)
```xml
<?xml version="1.0" encoding="utf-8"?>
<DiagnosticsConfiguration xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
<PublicConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
<WadCfg>
<DiagnosticMonitorConfiguration overallQuotaInMB="4096" sinks="applicationInsights.errors">
<DiagnosticInfrastructureLogs scheduledTransferLogLevelFilter="Error" />
<Directories scheduledTransferPeriod="PT1M">
<IISLogs containerName="wad-iis-logfiles" />
<FailedRequestLogs containerName="wad-failedrequestlogs" />
</Directories>
<PerformanceCounters scheduledTransferPeriod="PT1M" sinks="HotPath">
<PerformanceCounterConfiguration counterSpecifier="\Memory\Available MBytes" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\ISAPI Extension Requests/sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Web Service(_Total)\Bytes Total/Sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET Applications(__Total__)\Requests/Sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET Applications(__Total__)\Errors Total/Sec" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Queued" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\ASP.NET\Requests Rejected" sampleRate="PT3M" />
<PerformanceCounterConfiguration counterSpecifier="\Processor(_Total)\% Processor Time" sampleRate="PT3M" />
</PerformanceCounters>
<WindowsEventLog scheduledTransferPeriod="PT1M">
<DataSource name="Application!*" />
</WindowsEventLog>
<CrashDumps>
<CrashDumpConfiguration processName="WaIISHost.exe" />
<CrashDumpConfiguration processName="WaWorkerHost.exe" />
<CrashDumpConfiguration processName="w3wp.exe" />
</CrashDumps>
<Logs scheduledTransferPeriod="PT1M" sinks="HotPath" scheduledTransferLogLevelFilter="Error" />
</DiagnosticMonitorConfiguration>
<SinksConfig>
<Sink name="HotPath">
<EventHub Url="https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py" SharedAccessKeyName="SendRule" />
</Sink>
<Sink name="applicationInsights">
<ApplicationInsights />
<Channels>
<Channel logLevel="Error" name="errors" />
</Channels>
</Sink>
</SinksConfig>
</WadCfg>
<StorageAccount>ACCOUNT_NAME</StorageAccount>
</PublicConfig>
<PrivateConfig xmlns="http://schemas.microsoft.com/ServiceHosting/2010/10/DiagnosticsConfiguration">
<StorageAccount name="{account name}" key="{account key}" endpoint="{storage endpoint}" />
<EventHub Url="https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py" SharedAccessKeyName="SendRule" SharedAccessKey="YOUR_KEY_HERE" />
</PrivateConfig>
<IsEnabled>true</IsEnabled>
</DiagnosticsConfiguration>
```
Die ergänzende Datei *ServiceConfiguration.Cloud.cscfg* für dieses Beispiel sieht folgendermaßen aus.
```xml
<?xml version="1.0" encoding="utf-8"?>
<ServiceConfiguration serviceName="MyFixItCloudService" xmlns="http://schemas.microsoft.com/ServiceHosting/2008/10/ServiceConfiguration" osFamily="3" osVersion="*" schemaVersion="2015-04.2.6">
<Role name="MyFixIt.WorkerRole">
<Instances count="1" />
<ConfigurationSettings>
<Setting name="Microsoft.WindowsAzure.Plugins.Diagnostics.ConnectionString" value="YOUR_CONNECTION_STRING" />
</ConfigurationSettings>
</Role>
</ServiceConfiguration>
```
Entsprechende JSON-Einstellungen für virtuelle Computer lauten wie folgt:
Öffentliche Einstellungen:
```JSON
{
"WadCfg": {
"DiagnosticMonitorConfiguration": {
"overallQuotaInMB": 4096,
"sinks": "applicationInsights.errors",
"DiagnosticInfrastructureLogs": {
"scheduledTransferLogLevelFilter": "Error"
},
"Directories": {
"scheduledTransferPeriod": "PT1M",
"IISLogs": {
"containerName": "wad-iis-logfiles"
},
"FailedRequestLogs": {
"containerName": "wad-failedrequestlogs"
}
},
"PerformanceCounters": {
"scheduledTransferPeriod": "PT1M",
"sinks": "HotPath",
"PerformanceCounterConfiguration": [
{
"counterSpecifier": "\\Memory\\Available MBytes",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Web Service(_Total)\\ISAPI Extension Requests/sec",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Web Service(_Total)\\Bytes Total/Sec",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\ASP.NET Applications(__Total__)\\Requests/Sec",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\ASP.NET Applications(__Total__)\\Errors Total/Sec",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\ASP.NET\\Requests Queued",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\ASP.NET\\Requests Rejected",
"sampleRate": "PT3M"
},
{
"counterSpecifier": "\\Processor(_Total)\\% Processor Time",
"sampleRate": "PT3M"
}
]
},
"WindowsEventLog": {
"scheduledTransferPeriod": "PT1M",
"DataSource": [
{
"name": "Application!*"
}
]
},
"Logs": {
"scheduledTransferPeriod": "PT1M",
"scheduledTransferLogLevelFilter": "Error",
"sinks": "HotPath"
}
},
"SinksConfig": {
"Sink": [
{
"name": "HotPath",
"EventHub": {
"Url": "https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py",
"SharedAccessKeyName": "SendRule"
}
},
{
"name": "applicationInsights",
"ApplicationInsights": "",
"Channels": {
"Channel": [
{
"logLevel": "Error",
"name": "errors"
}
]
}
}
]
}
},
"StorageAccount": "{account name}"
}
```
Geschützte Einstellungen:
```JSON
{
"storageAccountName": "{account name}",
"storageAccountKey": "{account key}",
"storageAccountEndPoint": "{storage endpoint}",
"EventHub": {
"Url": "https://diageventhub-py-ns.servicebus.windows.net/diageventhub-py",
"SharedAccessKeyName": "SendRule",
"SharedAccessKey": "YOUR_KEY_HERE"
}
}
```
## <a name="next-steps"></a>Nächste Schritte
Weitere Informationen zu Event Hubs finden Sie unter den folgenden Links:
* [Übersicht über Event Hubs](../../event-hubs/event-hubs-about.md)
* [Erstellen eines Event Hubs](../../event-hubs/event-hubs-create.md)
* [Event Hubs – häufig gestellte Fragen](../../event-hubs/event-hubs-faq.md)
<!-- Images. -->
[0]: ../../event-hubs/media/event-hubs-streaming-azure-diags-data/dashboard.png
| 54.455598 | 797 | 0.694448 | deu_Latn | 0.875963 |
9143d7326c3d92b7dcb88b1ccad7e8fb0ffbcdce | 13,797 | md | Markdown | treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md | emmettstr/docs | 2d0376d6e07f3ffa828f6152d12cf260a530c64d | [
"Apache-2.0"
] | null | null | null | treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md | emmettstr/docs | 2d0376d6e07f3ffa828f6152d12cf260a530c64d | [
"Apache-2.0"
] | null | null | null | treebanks/ru_syntagrus/ru_syntagrus-dep-fixed.md | emmettstr/docs | 2d0376d6e07f3ffa828f6152d12cf260a530c64d | [
"Apache-2.0"
] | null | null | null | ---
layout: base
title: 'Statistics of fixed in UD_Russian-SynTagRus'
udver: '2'
---
## Treebank Statistics: UD_Russian-SynTagRus: Relations: `fixed`
This relation is universal.
9959 nodes (1%) are attached to their parents as `fixed`.
9959 instances of `fixed` (100%) are left-to-right (parent precedes child).
Average distance between parent and child is 1.21889747966663.
The following 83 pairs of parts of speech are connected with `fixed`: <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (3357; 34% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (950; 10% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (622; 6% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (533; 5% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (440; 4% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (369; 4% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (358; 4% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (339; 3% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt> (271; 3% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (253; 3% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (250; 3% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (196; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (190; 2% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (188; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (187; 2% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (175; 2% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (156; 2% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (114; 1% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (108; 1% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (95; 1% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (88; 1% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (86; 1% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (84; 1% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (52; 1% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (45; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (41; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (36; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (36; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (34; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (28; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt> (28; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (18; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (15; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (15; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (13; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (12; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-AUX.html">AUX</a></tt> (11; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (10; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (9; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (8; 0% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-DET.html">DET</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (7; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (6; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt> (6; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-AUX.html">AUX</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (5; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (3; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (3; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-INTJ.html">INTJ</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-X.html">X</a></tt>-<tt><a href="ru_syntagrus-pos-X.html">X</a></tt> (2; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADP.html">ADP</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NOUN.html">NOUN</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-NUM.html">NUM</a></tt>-<tt><a href="ru_syntagrus-pos-SYM.html">SYM</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-ADJ.html">ADJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PART.html">PART</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt>-<tt><a href="ru_syntagrus-pos-CCONJ.html">CCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-PRON.html">PRON</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-SCONJ.html">SCONJ</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt>-<tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-ADV.html">ADV</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-VERB.html">VERB</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances), <tt><a href="ru_syntagrus-pos-X.html">X</a></tt>-<tt><a href="ru_syntagrus-pos-PROPN.html">PROPN</a></tt> (1; 0% instances).
~~~ conllu
# visual-style 3 bgColor:blue
# visual-style 3 fgColor:white
# visual-style 1 bgColor:blue
# visual-style 1 fgColor:white
# visual-style 1 3 fixed color:blue
1 До до ADP _ _ 4 advmod 4:advmod _
2 сих сей DET _ Case=Gen|Number=Plur 1 fixed 1:fixed _
3 пор пора NOUN _ Animacy=Inan|Case=Gen|Gender=Fem|Number=Plur 1 fixed 1:fixed _
4 идут идти VERB _ Aspect=Imp|Mood=Ind|Number=Plur|Person=3|Tense=Pres|VerbForm=Fin|Voice=Act 0 root 0:root _
5 споры спор NOUN _ Animacy=Inan|Case=Nom|Gender=Masc|Number=Plur 4 nsubj 4:nsubj _
6 о о ADP _ _ 7 case 7:case _
7 том то PRON _ Animacy=Inan|Case=Loc|Gender=Neut|Number=Sing 5 nmod 5:nmod SpaceAfter=No
8 , , PUNCT _ _ 9 punct 9:punct _
9 достигли достичь VERB _ Aspect=Perf|Mood=Ind|Number=Plur|Tense=Past|VerbForm=Fin|Voice=Act 7 acl 7:acl _
10 они они PRON _ Case=Nom|Number=Plur|Person=3 9 nsubj 9:nsubj _
11 вершины вершина NOUN _ Animacy=Inan|Case=Gen|Gender=Fem|Number=Sing 9 obl 9:obl _
12 или или CCONJ _ _ 13 cc 13.1:cc _
13 нет нет PART _ _ 9 conj 13.1:discourse SpaceAfter=No
~~~
~~~ conllu
# visual-style 2 bgColor:blue
# visual-style 2 fgColor:white
# visual-style 1 bgColor:blue
# visual-style 1 fgColor:white
# visual-style 1 2 fixed color:blue
1 Так так ADV _ Degree=Pos 4 mark 4:mark _
2 что что SCONJ _ _ 1 fixed 1:fixed _
3 - - PUNCT _ _ 1 punct 1:punct _
4 спасибо спасибо NOUN _ Animacy=Inan|Case=Nom|Gender=Neut|Number=Sing 0 root 0:root SpaceAfter=No
5 , , PUNCT _ _ 6 punct 6:punct _
6 власть власть NOUN _ Animacy=Inan|Case=Nom|Gender=Fem|Number=Sing 4 parataxis 4:parataxis SpaceAfter=No
7 , , PUNCT _ _ 8 punct 8:punct _
8 спасибо спасибо NOUN _ Animacy=Inan|Case=Nom|Gender=Neut|Number=Sing 4 conj 4:conj SpaceAfter=No
9 , , PUNCT _ _ 10 punct 10:punct _
10 Путин Путин PROPN _ Animacy=Anim|Case=Nom|Gender=Masc|Number=Sing 8 parataxis 8:parataxis SpaceAfter=No
11 . . PUNCT _ _ 4 punct 4:punct _
~~~
~~~ conllu
# visual-style 6 bgColor:blue
# visual-style 6 fgColor:white
# visual-style 5 bgColor:blue
# visual-style 5 fgColor:white
# visual-style 5 6 fixed color:blue
1 Разумеется разумеется ADV _ Degree=Pos 7 parataxis 7:parataxis SpaceAfter=No
2 , , PUNCT _ _ 1 punct 1:punct _
3 эти этот DET _ Case=Acc|Number=Plur 4 det 4:det _
4 объяснения объяснение NOUN _ Animacy=Inan|Case=Acc|Gender=Neut|Number=Plur 8 obj 8:obj _
5 вряд вряд ADV _ Degree=Pos 7 advmod 7:advmod _
6 ли ли PART _ _ 5 fixed 5:fixed _
7 можно можно ADV _ Degree=Pos 0 root 0:root _
8 счесть счесть VERB _ Aspect=Perf|VerbForm=Inf|Voice=Act 7 csubj 7:csubj _
9 убедительными убедительный ADJ _ Case=Ins|Degree=Pos|Number=Plur 8 obl 8:obl SpaceAfter=No
10 . . PUNCT _ _ 7 punct 7:punct _
~~~
| 166.228916 | 10,758 | 0.679278 | yue_Hant | 0.670034 |
914405e6e10ea424254abb3a4b037e754c09afb0 | 6,509 | md | Markdown | README.md | by46/WhaleFS | 20029ad9a9b59089a3fdd681699ddfb3a5624ebc | [
"MIT"
] | 1 | 2018-06-10T08:54:54.000Z | 2018-06-10T08:54:54.000Z | README.md | by46/whalefs | 20029ad9a9b59089a3fdd681699ddfb3a5624ebc | [
"MIT"
] | null | null | null | README.md | by46/whalefs | 20029ad9a9b59089a3fdd681699ddfb3a5624ebc | [
"MIT"
] | null | null | null | # whalefs
## seaweedfs
```bash
weed master -port=9001
weed master -port=9002 -peers="localhost:9001"
weed volume -port=9081 -mserver="localhost:9001" -dir="data"
/opt/weed/weed master -mdir="/opt/dfs/master"
/opt/weed/weed volume -ip=192.168.1.9 -port=18081 -mserver="localhost:9333" -dir="/opt/dfs/data1"
/opt/weed/weed volume -ip=192.168.1.9 -port=18082 -mserver="localhost:9333" -dir="/opt/dfs/data2"
/opt/weed/weed volume -ip=192.168.1.9 -port=18083 -mserver="localhost:9333" -dir="/opt/dfs/data3"
```
### buckets
system.buckets
```json
{
"buckets": [
"system.bucket.benjamin"
]
}
```
### bucket
system.bucket.benjamin
```json
{
"name": "benjamin",
"memo":"mo bucket",
"basis": {
"alias": "pdt",
"collection": "",
"replication": "100",
"expires": 20,
"prepare_thumbnail_min_width": 1024,
"prepare_thumbnail": ""
},
"expires": 20,
"extends":[{"key":"keepdate", "value":"21"}],
"limit": {
"min_size": null,
"max_size": 102400,
"width": null,
"height": null,
"mime_types": ["image/png", "image/jpeg", "image/png"]
},
"last_edit_date": 123143,
"last_edit_user":"by46",
"overlays": [
{"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8}
],
"sizes": [
{"name": "p200", "width":200, "height": 150, "mode": "stretch"},
{"name": "p60", "width":60, "height": 45, "mode": "fit"},
{"name": "p160", "width":160, "height": 120, "mode": "thumbnail"}
]
}
```
system.bucket.package
```json
{
"name": "package",
"memo":"mo bucket",
"basis": {
"collection": "",
"replication": "100",
"expires": 20,
"prepare_thumbnail_min_width": 1024,
"prepare_thumbnail": ""
},
"expires": 20,
"extends":[{"key":"keepdate", "value":"21"}],
"limit": {
"min_size": null,
"max_size": 102400,
"width": null,
"height": null,
"mime_types": ["image/png", "image/jpeg", "image/png"]
},
"last_edit_date": 123143,
"last_edit_user":"by46",
"overlays": [
{"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8}
],
"sizes": [
{"name": "p200", "width":200, "height": 150, "mode": "stretch"},
{"name": "p60", "width":60, "height": 45, "mode": "fit"},
{"name": "p160", "width":160, "height": 120, "mode": "thumbnail"}
]
}
```
system.bucket.product
```json
{
"name": "product",
"memo":"mo bucket",
"basis": {
"alias": "pdt",
"collection": "",
"replication": "100",
"expires": 20,
"prepare_thumbnail_min_width": 1024,
"prepare_thumbnail": ""
},
"expires": 20,
"extends":[{"key":"keepdate", "value":"21"}],
"limit": {
"min_size": null,
"max_size": 102400,
"width": null,
"height": null,
"mime_types": ["image/png", "image/jpeg", "image/png"]
},
"last_edit_date": 123143,
"last_edit_user":"by46",
"overlays": [
{"name": "demo1", "default": true, "position": "TopLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo2", "default": false, "position": "TopRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo3", "default": false, "position": "BottomLeft", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo4", "default": false, "position": "BottomRight", "image": "7,15154f3ef7", "opacity": 0.8},
{"name": "demo5", "default": false, "position": "{\"top\":null, \"right\":0,\"bottom\":0, \"left\":0}", "image": "7,15154f3ef7", "opacity": 0.8}
],
"sizes": [
{"name": "p200", "width":200, "height": 150, "mode": "stretch"},
{"name": "p60", "width":60, "height": 45, "mode": "fit"},
{"name": "p160", "width":160, "height": 120, "mode": "thumbnail"}
]
}
```
```
package: github.com/by46/whalefs
homepage: https://github.com/by46/whalefs
license: MIT
owners:
- name: benjamin.c.yan
email: ycs_ctbu_2010@126.com
import:
- package: github.com/spf13/viper
version: ^1.0.2
- package: github.com/sirupsen/logrus
version: ^1.0.5
- package: github.com/couchbase/go-couchbase
- package: github.com/couchbase/gomemcached
- package: github.com/couchbase/goutils
- package: github.com/mholt/binding
version: ^0.3.0
- package: github.com/spf13/cobra
version: 0.0.2
- package: github.com/pkg/errors
version: ^0.8.0
```
### 普罗米修斯
http://172.16.0.158:9090/graph
### golang 编程规范
### 接口 大文件上传
#### 初始化上传
POST /benjamin/demo/hello.jpg?uploads
{
"upload_id":"uuid1"
}
#### 上传chunk
PUT /benjamin/demo/hello.jpg?uploadId=uuid1&partNumber=partNumber1
<multipart-form>
</multipart-form>
#### 完成上传
POST /benjamin/demo/hello.jpg?uploadId=uuid1
[{
"part_number": "part number 1",
"etag": "etag1"
}]
whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="banner,banner_01"
whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="bond,bond_01,contract,contract_01,eInvoice,eInvoice_01"
whalefs.exe migrate --location="D:\application\ImageServer" --target="192.168.1.9:8000" --includes="ibuild,ibuild_01"
192.168.110.81:5672,192.168.110.84:5672
spring.rabbitmq.virtual-host = /blockchain
spring.rabbitmq.username = ifs_admin
spring.rabbitmq.password = ifs@123
```bash
rsync -av -e ssh /Users/benjamin.c.yan/golang/pkg/mod user@remotehost:~/mod
``` | 29.586364 | 156 | 0.583039 | yue_Hant | 0.093307 |
91442347ac9791c6716df4e68ad870c0da73ab48 | 1,906 | md | Markdown | docs/database/tometek/README.md | friendly-router/friendly-router | 5020871d222b912d280eb90c1f9fc1bde71fad51 | [
"0BSD"
] | 3 | 2021-04-01T12:41:55.000Z | 2021-04-05T13:07:08.000Z | docs/database/tometek/README.md | friendly-router/friendly-router | 5020871d222b912d280eb90c1f9fc1bde71fad51 | [
"0BSD"
] | null | null | null | docs/database/tometek/README.md | friendly-router/friendly-router | 5020871d222b912d280eb90c1f9fc1bde71fad51 | [
"0BSD"
] | null | null | null | ---
lang: en-US
title: Tometek
sidebar: auto
draft: true
prev: ../
meta:
- name: "twitter:card"
value: "Friendly Router Project"
- name: "twitter:site"
value: "https://friendly-router.org/database/tometek"
- name: "twitter:title"
value: "Database | Tometek"
- name: "description"
value: |
Tometek manufacturer devices and hardware.
- name: "twitter:creator"
value: "@FriendlyRouter"
- name: "twitter:image"
value: "https://friendly-router.org/images/switchy-200x200.png"
- property: "og:site_name"
content: "Friendly Router Project"
- property: "og:title"
content: "Database | Tometek"
- property: "og:type"
content: "website"
- property: "og:url"
content: "https://friendly-router.org/database/tometek"
- property: "og:image"
content: "https://friendly-router.org/images/switchy-200x200.png"
- property: "og:description"
content: |
Tometek manufacturer devices and hardware.
---
> TOM was founded in 2004 by three co-founders who have working
> experience in world-class enterprise in Low-voltage Electric and
> New energy & Computing application & Telecommunication field while
> combined 10+ industry experience in China. In October 2013 TOM and
> his partner provided power facilities and IT solution to power the
> 27th SEA Games in 4 cities in Myanmar.
>
> -- [Official Website](http://www.tometek.com/about.asp?id=207)
## Products
| Name | Doc | Test | Bench | Hack | Note |
|---------------------------|-----|------|-------|------|------|
| [max-sppc](max-sppc/)
| [max-tsppc](max-tsppc/)
| [max-tsssc](max-tsssc/)
| [max-tts](max-tts/)
| [pro-sccc](pro-sccc/)
| [pro-sscc](pro-sscc/)
| [pro-sssc](pro-sssc/)
| [pro-sc](pro-sc/)
| [lite-cc](lite-cc/)
| [lite-cccc](lite-cccc/)
## Archives
## References and Resources
* [http://www.tometek.com](http://www.tometek.com)
| 27.623188 | 69 | 0.641658 | eng_Latn | 0.584988 |
9144861c3ea492cc1437a3a22b1e8c3a60e62d72 | 1,500 | md | Markdown | README.md | Szlavicsek/Quiz | 7e70d0e32db0a1da5c1cbee36dd60a435c49a522 | [
"MIT"
] | null | null | null | README.md | Szlavicsek/Quiz | 7e70d0e32db0a1da5c1cbee36dd60a435c49a522 | [
"MIT"
] | null | null | null | README.md | Szlavicsek/Quiz | 7e70d0e32db0a1da5c1cbee36dd60a435c49a522 | [
"MIT"
] | null | null | null | # Quizzit app
A quiz game project using the [Open Trivia API](https://opentdb.com/api_config.php)
## Getting Started
To get started, clone the repo to your local machine and install the dependencies as listed below. Please note that for the Google login api to work properly, you will also need:
* to configure the game as a new project on Google (you can do so [here](https://developers.google.com/identity/sign-in/web/sign-in))
* a development url to register within the Google dev console for your project (which is other than your local host's url, else it won't work either)
After registering, replace the google client id in the "google-signin-client_id" meta tag with yours.
For testing purposes, registering a new project is not mandatory, skipping the process above will only affect the corresponding functionality of the app.
### Installing prerequisites
The project files are bundled with webpack. To install it, run the following code in your terminal:
```
npm install --save-dev webpack-cli
```
To manage audio files properly on all devices, you will also need the [howlerjs](https://github.com/goldfire/howler.js) library:
```
npm install howler
```
Once the dependencies are installed, you can open the project with live server and run `npm run dev` which will watch and bundle the files automatically during development.
To build the production version, you can run `npm run build`.
## Authors
Pataki Péter
## License
This project is licensed under the MIT License.
| 35.714286 | 178 | 0.772 | eng_Latn | 0.99651 |
9144e74aef19e60a189a4f3a507b250bbec82a04 | 39 | md | Markdown | readme.md | eladzlot/minno-sequencer | 0c34902ce359b5e4850089c5ad341f0cd4f58d4e | [
"Apache-2.0"
] | 1 | 2021-04-13T05:03:38.000Z | 2021-04-13T05:03:38.000Z | readme.md | eladzlot/minno-sequencer | 0c34902ce359b5e4850089c5ad341f0cd4f58d4e | [
"Apache-2.0"
] | 5 | 2019-11-20T17:12:55.000Z | 2022-03-02T05:05:16.000Z | readme.md | eladzlot/minno-sequencer | 0c34902ce359b5e4850089c5ad341f0cd4f58d4e | [
"Apache-2.0"
] | null | null | null | # Minno sequencer
The minno sequencer
| 9.75 | 19 | 0.794872 | eng_Latn | 0.951134 |
9145812f2f6f975b3760b2daec93decd6b6e84aa | 2,025 | md | Markdown | docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md | turibbio/docs.it-it | 2212390575baa937d6ecea44d8a02e045bd9427c | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md | turibbio/docs.it-it | 2212390575baa937d6ecea44d8a02e045bd9427c | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/winforms/controls/ways-to-select-a-windows-forms-button-control.md | turibbio/docs.it-it | 2212390575baa937d6ecea44d8a02e045bd9427c | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Modalità di selezione di un controllo Button
ms.date: 03/30/2017
helpviewer_keywords:
- Button control [Windows Forms], selecting
ms.assetid: fe2fc058-5118-4f70-b264-6147d64a7a8d
ms.openlocfilehash: 145166d182f1ec51068ab3e0c23c12b471b69231
ms.sourcegitcommit: de17a7a0a37042f0d4406f5ae5393531caeb25ba
ms.translationtype: MT
ms.contentlocale: it-IT
ms.lasthandoff: 01/24/2020
ms.locfileid: "76740011"
---
# <a name="ways-to-select-a-windows-forms-button-control"></a>Modalità di selezione di un controllo Button Windows Form
È possibile selezionare un pulsante Windows Forms nei modi seguenti:
- Utilizzare un mouse per fare clic sul pulsante.
- Richiama l'evento <xref:System.Windows.Forms.Control.Click> del pulsante nel codice.
- Spostare lo stato attivo sul pulsante premendo il tasto TAB, quindi scegliere il pulsante premendo la barra SPAZIAtrice o invio.
- Premere il tasto di accesso (ALT + la lettera sottolineata) per il pulsante. Per altre informazioni sulle chiavi di accesso, vedere [procedura: creare chiavi di accesso per controlli Windows Forms](how-to-create-access-keys-for-windows-forms-controls.md).
- Se il pulsante è il pulsante "Accetto" del modulo, premendo invio viene scelto il pulsante, anche se un altro controllo ha lo stato attivo, ad eccezione del fatto che un altro controllo è un altro pulsante, una casella di testo a più righe o un controllo personalizzato che intrappola il tasto INVIO.
- Se il pulsante è il pulsante "Annulla" del modulo, premendo ESC viene scelto il pulsante anche se un altro controllo ha lo stato attivo.
- Chiamare il metodo <xref:System.Windows.Forms.Button.PerformClick%2A> per selezionare il pulsante a livello di codice.
## <a name="see-also"></a>Vedere anche
- [Panoramica sul controllo Button](button-control-overview-windows-forms.md)
- [Procedura: Rispondere alla selezione dei pulsanti di Windows Form](how-to-respond-to-windows-forms-button-clicks.md)
- [Controllo Button](button-control-windows-forms.md)
| 56.25 | 304 | 0.784691 | ita_Latn | 0.994376 |
9145d227d380c093e845def87ac0a8db15ec55d1 | 6,400 | md | Markdown | _posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md | Anja-Allende/Anja-Allende | 4acf09e3f38033a4abc7f31f37c778359d8e1493 | [
"MIT"
] | 2 | 2019-02-28T03:47:33.000Z | 2020-04-06T07:49:53.000Z | _posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md | Anja-Allende/Anja-Allende | 4acf09e3f38033a4abc7f31f37c778359d8e1493 | [
"MIT"
] | null | null | null | _posts/2018-11-28-Download-miss-rosie-apos-s-spice-of-life-quilts-leisure-arts.md | Anja-Allende/Anja-Allende | 4acf09e3f38033a4abc7f31f37c778359d8e1493 | [
"MIT"
] | null | null | null | ---
layout: post
comments: true
categories: Other
---
## Download Miss rosie apos s spice of life quilts leisure arts book
"I can't imagine whole cities burning. could see the silver drops pooling on his tongue before he swallowed. He wouldn't need the bottle any more, and he quickly slipped inside. D and Micky at the kitchen gardening supplies? " as possible to miss rosie apos s spice of life quilts leisure arts guests, leaving her employed she sits on miss rosie apos s spice of life quilts leisure arts part of the skin and stretches out the other The pending storm had broken, he found in Zedd the welcome reassurance that it was all right to be you had to do with a good hard-packed clay miss rosie apos s spice of life quilts leisure arts was sweep it and now and then sprinkle it to keep couldn't see through you. I'm not sure that God, it's mostly lies, O my masters, Sarytschev says that at St, he had taken a preventive dose of paregoric, Sarytschev says that at St. He is unable to work up "Not. Nevertheless the resemblance is so strong that he must be a how he might ever again trust anyone sufficiently to take the wedding Without breaking stride, but those were daylight. How strange. When Micky heard miss rosie apos s spice of life quilts leisure arts pet name, so acidic that his gums burned from contact with tortured beyond endurance, but a would-be stage magician eventually needed a mentor to campsites with power-and-water hookups to motor homes and travel trailers. with someone headed for a more populous area that will provide even better Throughout the morning, when he was out from under the tower. "When you're a star," she said once, macaroni and cheese, Geneva gone. tip?" "Like most of us these days, she'd she must have noticed it. A less self-controlled man might have seized a nearby bronze vase-fashioned to resemble miss rosie apos s spice of life quilts leisure arts stool-and stuffed her into it or vice versa? They broke a law that must not be broken! Nearing the house, _although the safety notes. " had eons ago circled too close to the treacherous bogs that swallowed them by no louder than before, an obituary should be published one day or the At the sofabed again. Anselmo's Orphanage, more important, wise men, my caseload being the new Adam or with a yearning for fatherhood, more efficient and more thoroughly reasoned, more relaxed, for those who hate because they have no recognition of the Leilani? A shudder, when mourners gone. A police officer. I can now inform you that, the king summoned his vizier and bade him tell the story of the king who lost kingdom and wife and wealth, but in a way that slipped subliminally into the mind of the beholder and gripped it by its deepest roots, There was some strength in him or with him. Barty came out of the house with the library copy of Podkayne Of Mary, i. " might enable the magic to repeat. " shapes, I'd crawl in there again, regardless of what if she miss rosie apos s spice of life quilts leisure arts been merely frightened. Micky clawed in miss rosie apos s spice of life quilts leisure arts, Hal. "You behave yourself tonight," Celestina said. Where and when?" Hanlon said. Now he had no hope of resisting Gelluk in any way. He'd read it in a novel, as if she were a high-school girl and he were her with utmost consideration, little sticky spots, but at last tore it off, which would explain the stonecarver's accelerated service, clatters as he climbs across it. He himself is not a hunter, was there tap water?" of the crystal rended reflections of the flames into red-orange-yellow-green- with morning drinking or perhaps with drinking binges at any hour, and Without the advantage of surprise, and he believes that he could head of a monster, to cover her discomfort at how close Micky had come to the truth, Curtis watches the receding figure until it's clear the man won't attempt to sneak back, ii. Nolan had shrugged, opening herself to the anger and hatred with which she had regarded this child in the operating room, but not by the Circle of Friends. This always proved to be a had of bold strategy, the mind had a thermostat of its own, called, sprints on, but I'm pissed. social organisation. There were people of the Hand there whom he trusted. TETGALES, 1741. Gelluk was powerful, they used them on pickles on the side, she had reached him even though he didn't want to become involved, he backed out. Or the kid might have been placed for adoption through some baby brokerage in it strictly for the money. Just forget the busload of nuns smashed on the tracks, you were depressed and feeling hopeless. You don't have to worry about me. What was Simon's fee?" They're illegal but not hard to find. " She looked at me through her lashes. The sail struck the water, with the old wizard, Chapter 71 anyone but herself, to cover her discomfort at how close Micky had come to the truth. Along with the bird now described there occur, the, I take grasp. "Do I have a cloud inside me now?" breast of darkness. thought he had no room to plant the seeds for any more? Shaw, 'Whoso stumbleth on a stone and returneth thereto, Curtis retreats from the dog and from the motor home, i. They're most likely fast approaching from the other side of the vehicle. of cold was the same; that is to say, he saw them spread out the gold before them and play with it and heard one of them say, he looked at her as he spoke. Working with patching kits and lasers to cut the tough material, and I am very much at home. In its "More vanilla Coke, taking with them a few bewildered secretaries and office workers that they had bumped into on the way. "She was a good cow, of course? It was clear by this time that the Organizer had no intention of settling for a smaller package, which had all been partially crushed by that pewter candlestick. " He shrugged. " struck numb and mute by the conflict between yearning and inexperience. in a _baydar_ and kill them all but two. together as if with fine-draw stitches. the beginning of the seventeenth century reproduced below, Mrs. "What I hope I found there was direction, by rolling onto my back. were buying Grandma's locally famous black bean-and-corn salsa. 9 -1. "Good thing we weren't playing Russian roulette," Leilani said. It was not until afterwards that we of aspirin! I'm writing to you about an In addition to the firearm, one Maria. | 711.111111 | 6,275 | 0.78625 | eng_Latn | 0.999962 |
91460c75c66d3563f4add5912fbe94e1f77ccf30 | 15,441 | md | Markdown | articles/virtual-machines/windows/tutorial-create-vmss.md | changeworld/azure-docs.it- | 34f70ff6964ec4f6f1a08527526e214fdefbe12a | [
"CC-BY-4.0",
"MIT"
] | 1 | 2017-06-06T22:50:05.000Z | 2017-06-06T22:50:05.000Z | articles/virtual-machines/windows/tutorial-create-vmss.md | changeworld/azure-docs.it- | 34f70ff6964ec4f6f1a08527526e214fdefbe12a | [
"CC-BY-4.0",
"MIT"
] | 41 | 2016-11-21T14:37:50.000Z | 2017-06-14T20:46:01.000Z | articles/virtual-machines/windows/tutorial-create-vmss.md | changeworld/azure-docs.it- | 34f70ff6964ec4f6f1a08527526e214fdefbe12a | [
"CC-BY-4.0",
"MIT"
] | 7 | 2016-11-16T18:13:16.000Z | 2017-06-26T10:37:55.000Z | ---
title: 'Esercitazione: Creare un set di scalabilità di macchine virtuali Windows'
description: Informazioni su come usare Azure PowerShell per creare e distribuire un'applicazione a disponibilità elevata nelle VM Windows usando un set di scalabilità di macchine virtuali
author: ju-shim
ms.author: jushiman
ms.topic: tutorial
ms.service: virtual-machine-scale-sets
ms.subservice: windows
ms.date: 11/30/2018
ms.reviewer: mimckitt
ms.custom: mimckitt, devx-track-azurepowershell
ms.openlocfilehash: b3853ddc71d1a9be26b2492764a9b341446e0eeb
ms.sourcegitcommit: f28ebb95ae9aaaff3f87d8388a09b41e0b3445b5
ms.translationtype: MT
ms.contentlocale: it-IT
ms.lasthandoff: 03/29/2021
ms.locfileid: "89078742"
---
# <a name="tutorial-create-a-virtual-machine-scale-set-and-deploy-a-highly-available-app-on-windows-with-azure-powershell"></a>Esercitazione: Creare un set di scalabilità di macchine virtuali e distribuire un'app a disponibilità elevata in Windows con Azure PowerShell
Un set di scalabilità di macchine virtuali consente di distribuire e gestire un set di macchine virtuali identiche con scalabilità automatica. È possibile adattare manualmente il numero di macchine virtuali nel set di scalabilità, nonché definire regole di scalabilità automatica in base all'utilizzo della CPU, alla richiesta di memoria o al traffico di rete. In questa esercitazione viene distribuito un set di scalabilità di macchine virtuali in Azure e viene spiegato come:
> [!div class="checklist"]
> * Usare l'estensione dello script personalizzata per definire un sito IIS da ridimensionare
> * Creare un bilanciamento del carico per il set di scalabilità
> * Creare un set di scalabilità di macchine virtuali
> * Aumentare o diminuire il numero di istanze in un set di scalabilità
> * Creare regole di scalabilità automatica
## <a name="launch-azure-cloud-shell"></a>Avviare Azure Cloud Shell
Azure Cloud Shell è una shell interattiva gratuita che può essere usata per eseguire la procedura di questo articolo. Include strumenti comuni di Azure preinstallati e configurati per l'uso con l'account.
Per aprire Cloud Shell, basta selezionare **Prova** nell'angolo superiore destro di un blocco di codice. È anche possibile avviare Cloud Shell in una scheda separata del browser visitando [https://shell.azure.com/powershell](https://shell.azure.com/powershell). Selezionare **Copia** per copiare i blocchi di codice, incollarli in Cloud Shell e premere INVIO per eseguirli.
## <a name="scale-set-overview"></a>Informazioni generali sui set di scalabilità
Un set di scalabilità di macchine virtuali consente di distribuire e gestire un set di macchine virtuali identiche con scalabilità automatica. Le macchine virtuali di un set di scalabilità vengono distribuite in domini logici di errore e di aggiornamento in uno o più *gruppi di posizionamento*. I gruppi di posizionamento sono gruppi di macchine virtuali configurate in modo simile, analoghi ai [set di disponibilità](tutorial-availability-sets.md).
Le VM vengono create in base alle esigenze in un set di scalabilità. È possibile definire regole di scalabilità automatica per controllare le modalità e i tempi di aggiunta e rimozione delle VM dal set di scalabilità. Queste regole possono essere attivate in base a determinate metriche, ad esempio il carico della CPU, l'utilizzo della memoria o il traffico di rete.
I set di scalabilità supportano fino a 1000 macchine virtuali quando si usa un'immagine della piattaforma Azure. Per i carichi di lavoro con requisiti significativi di installazione o personalizzazione di VM, si consiglia di [creare un'immagine di VM personalizzata](tutorial-custom-images.md). È possibile creare fino a 600 macchine virtuali in un set di scalabilità quando si usa un'immagine personalizzata.
## <a name="create-a-scale-set"></a>Creare un set di scalabilità
Creare un set di scalabilità di macchine virtuali con [New-AzVmss](/powershell/module/az.compute/new-azvmss). L'esempio seguente crea un set di scalabilità denominato *myScaleSet* che usa l'immagine della piattaforma *Windows Server 2016 Datacenter*. Vengono create automaticamente le risorse di rete di Azure per la rete virtuale, l'indirizzo IP pubblico e il bilanciamento del carico. Quando richiesto, è possibile impostare le credenziali amministrative per le istanze di macchina virtuale nel set di scalabilità:
```azurepowershell-interactive
New-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-Location "EastUS" `
-VMScaleSetName "myScaleSet" `
-VirtualNetworkName "myVnet" `
-SubnetName "mySubnet" `
-PublicIpAddressName "myPublicIPAddress" `
-LoadBalancerName "myLoadBalancer" `
-UpgradePolicyMode "Automatic"
```
La creazione e la configurazione di tutte le macchine virtuali e risorse del set di scalabilità richiedono alcuni minuti.
## <a name="deploy-sample-application"></a>Distribuire l'applicazione di esempio
Per testare il set di scalabilità, installare un'applicazione Web di base. Usare l'estensione script personalizzati di Azure per scaricare ed eseguire uno script che installa IIS nelle istanze di macchina virtuale. Questa estensione è utile per la configurazione post-distribuzione, l'installazione di software o qualsiasi altra attività di configurazione o gestione. Per altre informazioni, vedere [Panoramica dell'estensione script personalizzata](../extensions/custom-script-windows.md).
Usare l'estensione script personalizzati per installare un server Web IIS di base. Applicare l'estensione di script personalizzati che installa IIS nel modo seguente:
```azurepowershell-interactive
# Define the script for your Custom Script Extension to run
$publicSettings = @{
"fileUris" = (,"https://raw.githubusercontent.com/Azure-Samples/compute-automation-configurations/master/automate-iis.ps1");
"commandToExecute" = "powershell -ExecutionPolicy Unrestricted -File automate-iis.ps1"
}
# Get information about the scale set
$vmss = Get-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet"
# Use Custom Script Extension to install IIS and configure basic website
Add-AzVmssExtension -VirtualMachineScaleSet $vmss `
-Name "customScript" `
-Publisher "Microsoft.Compute" `
-Type "CustomScriptExtension" `
-TypeHandlerVersion 1.8 `
-Setting $publicSettings
# Update the scale set and apply the Custom Script Extension to the VM instances
Update-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-Name "myScaleSet" `
-VirtualMachineScaleSet $vmss
```
## <a name="allow-traffic-to-application"></a>Consentire il traffico verso l'applicazione
Per consentire l'accesso all'applicazione Web di base, creare un gruppo di sicurezza di rete con [New-AzNetworkSecurityRuleConfig](/powershell/module/az.network/new-aznetworksecurityruleconfig) e [New-AzNetworkSecurityGroup](/powershell/module/az.network/new-aznetworksecuritygroup). Per altre informazioni, vedere [Rete per i set di scalabilità di macchine virtuali](../../virtual-machine-scale-sets/virtual-machine-scale-sets-networking.md).
```azurepowershell-interactive
# Get information about the scale set
$vmss = Get-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet"
#Create a rule to allow traffic over port 80
$nsgFrontendRule = New-AzNetworkSecurityRuleConfig `
-Name myFrontendNSGRule `
-Protocol Tcp `
-Direction Inbound `
-Priority 200 `
-SourceAddressPrefix * `
-SourcePortRange * `
-DestinationAddressPrefix * `
-DestinationPortRange 80 `
-Access Allow
#Create a network security group and associate it with the rule
$nsgFrontend = New-AzNetworkSecurityGroup `
-ResourceGroupName "myResourceGroupScaleSet" `
-Location EastUS `
-Name myFrontendNSG `
-SecurityRules $nsgFrontendRule
$vnet = Get-AzVirtualNetwork `
-ResourceGroupName "myResourceGroupScaleSet" `
-Name myVnet
$frontendSubnet = $vnet.Subnets[0]
$frontendSubnetConfig = Set-AzVirtualNetworkSubnetConfig `
-VirtualNetwork $vnet `
-Name mySubnet `
-AddressPrefix $frontendSubnet.AddressPrefix `
-NetworkSecurityGroup $nsgFrontend
Set-AzVirtualNetwork -VirtualNetwork $vnet
# Update the scale set and apply the Custom Script Extension to the VM instances
Update-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-Name "myScaleSet" `
-VirtualMachineScaleSet $vmss
```
## <a name="test-your-scale-set"></a>Testare il set di scalabilità
Per verificare il funzionamento del set di scalabilità, ottenere l'indirizzo IP pubblico del servizio di bilanciamento del carico con [Get-AzPublicIPAddress](/powershell/module/az.network/get-azpublicipaddress). L'esempio seguente mostra l'indirizzo IP per *myPublicIP* creato come parte del set di scalabilità:
```azurepowershell-interactive
Get-AzPublicIPAddress `
-ResourceGroupName "myResourceGroupScaleSet" `
-Name "myPublicIPAddress" | select IpAddress
```
Immettere l'indirizzo IP pubblico in un Web browser. Verrà visualizzata l'app Web, con il nome host della macchina virtuale a cui il servizio di bilanciamento del carico ha distribuito il traffico:

Per verificare il funzionamento del set di scalabilità, è possibile imporre l'aggiornamento del Web browser per visualizzare la distribuzione del traffico da parte del bilanciamento del carico tra tutte le macchine virtuali che eseguono l'app.
## <a name="management-tasks"></a>Attività di gestione
Nel ciclo di vita del set di scalabilità, potrebbe essere necessario eseguire una o più attività di gestione. Si potrebbe anche voler creare script per automatizzare le attività di ciclo di vita. Azure PowerShell offre un modo rapido per eseguire queste operazioni. Di seguito vengono illustrate alcune attività comuni.
### <a name="view-vms-in-a-scale-set"></a>Visualizzare le macchine virtuali in un set di scalabilità
Per visualizzare un elenco delle istanze di VM in un set di scalabilità, usare [Get-AzVmssVM](/powershell/module/az.compute/get-azvmssvm) come illustrato di seguito:
```azurepowershell-interactive
Get-AzVmssVM `
-ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet"
```
L'output di esempio seguente mostra due istanze di VM nel set di scalabilità:
```powershell
ResourceGroupName Name Location Sku InstanceID ProvisioningState
----------------- ---- -------- --- ---------- -----------------
MYRESOURCEGROUPSCALESET myScaleSet_0 eastus Standard_DS1_v2 0 Succeeded
MYRESOURCEGROUPSCALESET myScaleSet_1 eastus Standard_DS1_v2 1 Succeeded
```
Per visualizzare altre informazioni su un'istanza di VM specifica, aggiungere il parametro `-InstanceId` a [Get-AzVmssVM](/powershell/module/az.compute/get-azvmssvm). L'esempio seguente consente di visualizzare informazioni sull'istanza di macchina virtuale *1*:
```azurepowershell-interactive
Get-AzVmssVM `
-ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet" `
-InstanceId "1"
```
### <a name="increase-or-decrease-vm-instances"></a>Aumentare o diminuire le istanze delle macchine virtuali
Per visualizzare il numero di istanze attualmente presente in un set di scalabilità, usare [Get-AzVmss](/powershell/module/az.compute/get-azvmss) ed eseguire una query su *sku.capacity*:
```azurepowershell-interactive
Get-AzVmss -ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet" | `
Select -ExpandProperty Sku
```
È quindi possibile aumentare o ridurre manualmente il numero di macchine virtuali nel set di scalabilità con [Update-AzVmss](/powershell/module/az.compute/update-azvmss). L'esempio seguente imposta il numero di macchine virtuali del set di scalabilità su *3*:
```azurepowershell-interactive
# Get current scale set
$scaleset = Get-AzVmss `
-ResourceGroupName "myResourceGroupScaleSet" `
-VMScaleSetName "myScaleSet"
# Set and update the capacity of your scale set
$scaleset.sku.capacity = 3
Update-AzVmss -ResourceGroupName "myResourceGroupScaleSet" `
-Name "myScaleSet" `
-VirtualMachineScaleSet $scaleset
```
Sono necessari alcuni minuti per aggiornare il numero specificato di istanze del set di scalabilità.
### <a name="configure-autoscale-rules"></a>Configurare le regole di scalabilità automatica
Anziché scalare manualmente il numero di istanze del set di scalabilità, si definiscono regole di scalabilità automatica. Queste regole monitorano le istanze nel set di scalabilità e rispondono di conseguenza in base alle metriche e alle soglie definite. L'esempio seguente aumenta il numero di istanze di uno quando il carico della CPU medio è maggiore del 60% per un periodo di 5 minuti. Se il carico della CPU medio scende poi al di sotto del 30% per un periodo di 5 minuti, le istanze vengono ridotte di una istanza:
```azurepowershell-interactive
# Define your scale set information
$mySubscriptionId = (Get-AzSubscription)[0].Id
$myResourceGroup = "myResourceGroupScaleSet"
$myScaleSet = "myScaleSet"
$myLocation = "East US"
$myScaleSetId = (Get-AzVmss -ResourceGroupName $myResourceGroup -VMScaleSetName $myScaleSet).Id
# Create a scale up rule to increase the number instances after 60% average CPU usage exceeded for a 5-minute period
$myRuleScaleUp = New-AzAutoscaleRule `
-MetricName "Percentage CPU" `
-MetricResourceId $myScaleSetId `
-Operator GreaterThan `
-MetricStatistic Average `
-Threshold 60 `
-TimeGrain 00:01:00 `
-TimeWindow 00:05:00 `
-ScaleActionCooldown 00:05:00 `
-ScaleActionDirection Increase `
-ScaleActionValue 1
# Create a scale down rule to decrease the number of instances after 30% average CPU usage over a 5-minute period
$myRuleScaleDown = New-AzAutoscaleRule `
-MetricName "Percentage CPU" `
-MetricResourceId $myScaleSetId `
-Operator LessThan `
-MetricStatistic Average `
-Threshold 30 `
-TimeGrain 00:01:00 `
-TimeWindow 00:05:00 `
-ScaleActionCooldown 00:05:00 `
-ScaleActionDirection Decrease `
-ScaleActionValue 1
# Create a scale profile with your scale up and scale down rules
$myScaleProfile = New-AzAutoscaleProfile `
-DefaultCapacity 2 `
-MaximumCapacity 10 `
-MinimumCapacity 2 `
-Rule $myRuleScaleUp,$myRuleScaleDown `
-Name "autoprofile"
# Apply the autoscale rules
Add-AzAutoscaleSetting `
-Location $myLocation `
-Name "autosetting" `
-ResourceGroup $myResourceGroup `
-TargetResourceId $myScaleSetId `
-AutoscaleProfile $myScaleProfile
```
Per altre informazioni di progettazione sull'uso della scalabilità automatica, vedere [Procedure consigliate per la scalabilità automatica](/azure/architecture/best-practices/auto-scaling).
## <a name="next-steps"></a>Passaggi successivi
In questa esercitazione è stato creato un set di scalabilità di macchine virtuali. Si è appreso come:
> [!div class="checklist"]
> * Usare l'estensione dello script personalizzata per definire un sito IIS da ridimensionare
> * Creare un bilanciamento del carico per il set di scalabilità
> * Creare un set di scalabilità di macchine virtuali
> * Aumentare o diminuire il numero di istanze in un set di scalabilità
> * Creare regole di scalabilità automatica
Passare all'esercitazione successiva per maggiori informazioni sui concetti di bilanciamento del carico per le macchine virtuali.
> [!div class="nextstepaction"]
> [Bilanciare il carico di macchine virtuali](tutorial-load-balancer.md)
| 53.614583 | 520 | 0.786348 | ita_Latn | 0.968289 |
9146fad553db1c08e8ff90baad6c2a72579e035f | 11,600 | md | Markdown | src/pages/posts/2021-05-30T12:00:04-post.md | evanmacbride/reddit-digest | 47659c6b52d9b7d74025c517931e107cb2f6be94 | [
"MIT"
] | 1 | 2020-02-03T02:35:55.000Z | 2020-02-03T02:35:55.000Z | src/pages/posts/2021-05-30T12:00:04-post.md | evanmacbride/reddit-snapshots | 00dcad012a949243e7399a45dd9a37720cfe6576 | [
"MIT"
] | null | null | null | src/pages/posts/2021-05-30T12:00:04-post.md | evanmacbride/reddit-snapshots | 00dcad012a949243e7399a45dd9a37720cfe6576 | [
"MIT"
] | null | null | null | ---
title: '05/30/21 12:00PM UTC Snapshot'
date: '2021-05-30T12:00:04'
---
<ul>
<h2>Sci/Tech</h2>
<li><a href='https://i.redd.it/41rq1hape5271.jpg'><img src='https://b.thumbs.redditmedia.com/UeqXq9JRz54nw-Zlwbk3VeD9mpMV79ZVO3_eXACaRCs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/41rq1hape5271.jpg'>After 4 nights of clouds, I finally captured an incredibly rare atmospheric phenomenon in my astrophoto (OC)</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/MDieterich'>MDieterich</a> in <a href='https://www.reddit.com/r/space'>space</a> 40917 points & 639 <a href='https://www.reddit.com/r/space/comments/nnzc96/after_4_nights_of_clouds_i_finally_captured_an/'>comments</a></div></li>
<li><a href='https://www.businessinsider.com/cybersecurity-colonial-pipeline-companies-resemble-preschoolers-running-with-pointy-scissors-2021-5'><img src='https://a.thumbs.redditmedia.com/mKOBTn2htsI14JECVfdBSWrN8ORk6fFwdmzbcmOBMc8.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.businessinsider.com/cybersecurity-colonial-pipeline-companies-resemble-preschoolers-running-with-pointy-scissors-2021-5'>American companies resemble a bunch of preschoolers running with pointy scissors when it comes to cybersecurity</a></div>(businessinsider.com) posted by <a href='https://www.reddit.com/user/wewewawa'>wewewawa</a> in <a href='https://www.reddit.com/r/technology'>technology</a> 8775 points & 363 <a href='https://www.reddit.com/r/technology/comments/nnoofg/american_companies_resemble_a_bunch_of/'>comments</a></div></li>
<li><a href='https://www.businessinsider.com/unredacted-google-lawsuit-docs-detail-efforts-to-collect-user-location-2021-5'><svg version='1.1' viewBox='-34 -14 104 64' preserveAspectRatio='xMidYMid meet' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'>
<title>link thumbnail</title>
<path d='M32,4H4A2,2,0,0,0,2,6V30a2,2,0,0,0,2,2H32a2,2,0,0,0,2-2V6A2,2,0,0,0,32,4ZM4,30V6H32V30Z'></path>
<path d='M8.92,14a3,3,0,1,0-3-3A3,3,0,0,0,8.92,14Zm0-4.6A1.6,1.6,0,1,1,7.33,11,1.6,1.6,0,0,1,8.92,9.41Z'></path>
<path d='M22.78,15.37l-5.4,5.4-4-4a1,1,0,0,0-1.41,0L5.92,22.9v2.83l6.79-6.79L16,22.18l-3.75,3.75H15l8.45-8.45L30,24V21.18l-5.81-5.81A1,1,0,0,0,22.78,15.37Z'></path>
</svg></a><div><div class='linkTitle'><a href='https://www.businessinsider.com/unredacted-google-lawsuit-docs-detail-efforts-to-collect-user-location-2021-5'>'Apple is eating our lunch': Google employees admit in lawsuit that the company made it nearly impossible for users to keep their location private</a></div>(businessinsider.com) posted by <a href='https://www.reddit.com/user/giantyetifeet'>giantyetifeet</a> in <a href='https://www.reddit.com/r/tech'>tech</a> 2863 points & 185 <a href='https://www.reddit.com/r/tech/comments/nnvejg/apple_is_eating_our_lunch_google_employees_admit/'>comments</a></div></li>
<li><a href='https://www.reddit.com/gallery/nnomas'><img src='https://a.thumbs.redditmedia.com/gFC7bCfyKhf3yL2gZD7Ke28JAR0hE5lJRSuDx59hcf0.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nnomas'>No longer thought to be the smallest-known dinosaur, this bird-like lizard skull, trapped in Burmese amber nearly 100 million years ago, has been subject to a great deal of controversy because of its unusual combination of features, and it continues to be a very interesting specimen to science.</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/KimCureAll'>KimCureAll</a> in <a href='https://www.reddit.com/r/Naturewasmetal'>Naturewasmetal</a> 2769 points & 63 <a href='https://www.reddit.com/r/Naturewasmetal/comments/nnomas/no_longer_thought_to_be_the_smallestknown/'>comments</a></div></li>
<li><a href='https://i.redd.it/a5it7v4mc2271.jpg'><img src='https://b.thumbs.redditmedia.com/S0eEZ4cvYHUVwmCGrfPdh6kLcEU84Ujc4-u5H0Xi_HI.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/a5it7v4mc2271.jpg'>Whose THICC legs are these?</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/Onnisciente'>Onnisciente</a> in <a href='https://www.reddit.com/r/whatsthisbug'>whatsthisbug</a> 2119 points & 53 <a href='https://www.reddit.com/r/whatsthisbug/comments/nnnqjj/whose_thicc_legs_are_these/'>comments</a></div></li>
<li><a href='https://www.theverge.com/2021/5/29/22460070/google-difficult-android-privacy-settings-arizona'><img src='https://b.thumbs.redditmedia.com/z9bKr-WSlEEqJ1kkjh7u47Df4HcaX4Mh0zXKWqdqqKo.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.theverge.com/2021/5/29/22460070/google-difficult-android-privacy-settings-arizona'>Google reportedly made it difficult for smartphone users to find privacy settings</a></div>(theverge.com) posted by <a href='https://www.reddit.com/user/kry_some_more'>kry_some_more</a> in <a href='https://www.reddit.com/r/technology'>technology</a> 5863 points & 233 <a href='https://www.reddit.com/r/technology/comments/no0ruf/google_reportedly_made_it_difficult_for/'>comments</a></div></li>
<h2>Maker</h2>
<li><a href='https://www.reddit.com/gallery/nns4st'><img src='https://b.thumbs.redditmedia.com/Y_OotEvmZ7ko0zvxay56aUaY9YGKgKN64miD1FsfhQs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nns4st'>No it’s wasn’t photoshopped and yes it’s a Bust! Happy now?</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/evilmaul'>evilmaul</a> in <a href='https://www.reddit.com/r/3Dprinting'>3Dprinting</a> 5483 points & 317 <a href='https://www.reddit.com/r/3Dprinting/comments/nns4st/no_its_wasnt_photoshopped_and_yes_its_a_bust/'>comments</a></div></li>
<li><a href='https://i.imgur.com/Uyr5bHL.jpg'><img src='https://b.thumbs.redditmedia.com/_WBNm4FJVs_x6b7ZaKqA7IttEEKgwm5KO74Pv8VMkIs.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.imgur.com/Uyr5bHL.jpg'>Array methods in JavaScript. Original author unknown.</a></div>(i.imgur.com) posted by <a href='https://www.reddit.com/user/imAmarok'>imAmarok</a> in <a href='https://www.reddit.com/r/webdev'>webdev</a> 4170 points & 115 <a href='https://www.reddit.com/r/webdev/comments/nnwtf6/array_methods_in_javascript_original_author/'>comments</a></div></li>
<li><a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'><svg version='1.1' viewBox='-34 -12 104 64' preserveAspectRatio='xMidYMid slice' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'>
<title>text link thumbnail</title>
<path d='M12.19,8.84a1.45,1.45,0,0,0-1.4-1h-.12a1.46,1.46,0,0,0-1.42,1L1.14,26.56a1.29,1.29,0,0,0-.14.59,1,1,0,0,0,1,1,1.12,1.12,0,0,0,1.08-.77l2.08-4.65h11l2.08,4.59a1.24,1.24,0,0,0,1.12.83,1.08,1.08,0,0,0,1.08-1.08,1.64,1.64,0,0,0-.14-.57ZM6.08,20.71l4.59-10.22,4.6,10.22Z'>
</path>
<path d='M32.24,14.78A6.35,6.35,0,0,0,27.6,13.2a11.36,11.36,0,0,0-4.7,1,1,1,0,0,0-.58.89,1,1,0,0,0,.94.92,1.23,1.23,0,0,0,.39-.08,8.87,8.87,0,0,1,3.72-.81c2.7,0,4.28,1.33,4.28,3.92v.5a15.29,15.29,0,0,0-4.42-.61c-3.64,0-6.14,1.61-6.14,4.64v.05c0,2.95,2.7,4.48,5.37,4.48a6.29,6.29,0,0,0,5.19-2.48V26.9a1,1,0,0,0,1,1,1,1,0,0,0,1-1.06V19A5.71,5.71,0,0,0,32.24,14.78Zm-.56,7.7c0,2.28-2.17,3.89-4.81,3.89-1.94,0-3.61-1.06-3.61-2.86v-.06c0-1.8,1.5-3,4.2-3a15.2,15.2,0,0,1,4.22.61Z'>
</path>
</svg></a><div><div class='linkTitle'><a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'>RTX 3090 but still I have micro stutters</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/furytayx'>furytayx</a> in <a href='https://www.reddit.com/r/buildapc'>buildapc</a> 2176 points & 332 <a href='https://www.reddit.com/r/buildapc/comments/nnqbew/rtx_3090_but_still_i_have_micro_stutters/'>comments</a></div></li>
<li><a href='https://learnxinyminutes.com/'><svg version='1.1' viewBox='-34 -14 104 64' preserveAspectRatio='xMidYMid meet' xmlns='http://www.w3.org/2000/svg' xmlns:xlink='http://www.w3.org/1999/xlink'>
<title>link thumbnail</title>
<path d='M32,4H4A2,2,0,0,0,2,6V30a2,2,0,0,0,2,2H32a2,2,0,0,0,2-2V6A2,2,0,0,0,32,4ZM4,30V6H32V30Z'></path>
<path d='M8.92,14a3,3,0,1,0-3-3A3,3,0,0,0,8.92,14Zm0-4.6A1.6,1.6,0,1,1,7.33,11,1.6,1.6,0,0,1,8.92,9.41Z'></path>
<path d='M22.78,15.37l-5.4,5.4-4-4a1,1,0,0,0-1.41,0L5.92,22.9v2.83l6.79-6.79L16,22.18l-3.75,3.75H15l8.45-8.45L30,24V21.18l-5.81-5.81A1,1,0,0,0,22.78,15.37Z'></path>
</svg></a><div><div class='linkTitle'><a href='https://learnxinyminutes.com/'>Programming in many languages, and switching back and forth, I was looking for something with syntax and basics to make a working snippet like example file for myself. Found this gem.</a></div>(learnxinyminutes.com) posted by <a href='https://www.reddit.com/user/megaman5'>megaman5</a> in <a href='https://www.reddit.com/r/programming'>programming</a> 471 points & 21 <a href='https://www.reddit.com/r/programming/comments/nnooum/programming_in_many_languages_and_switching_back/'>comments</a></div></li>
<h2>Etcetera</h2>
<li><a href='https://i.redd.it/46hb188im2271.gif'><img src='https://b.thumbs.redditmedia.com/3_SUfqfXMu4XK5etgZprQao98f3UADXDXDDlCrKrf9U.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/46hb188im2271.gif'>A new bar for the game!</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/Ihor_Hontarenko'>Ihor_Hontarenko</a> in <a href='https://www.reddit.com/r/PixelArt'>PixelArt</a> 1845 points & 53 <a href='https://www.reddit.com/r/PixelArt/comments/nnoso4/a_new_bar_for_the_game/'>comments</a></div></li>
<li><a href='https://i.redd.it/s15lquq5o2271.jpg'><img src='https://b.thumbs.redditmedia.com/sAhLwplykSb6ubXDqnhOk38gc4BQAorUmgY09LFK4jk.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/s15lquq5o2271.jpg'>Sci-fi diagnostic equipment? Nah, barcode scanner (Star Trek: Into Darkness, 2013)</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/HalDimond'>HalDimond</a> in <a href='https://www.reddit.com/r/Thatsabooklight'>Thatsabooklight</a> 1200 points & 26 <a href='https://www.reddit.com/r/Thatsabooklight/comments/nnoym2/scifi_diagnostic_equipment_nah_barcode_scanner/'>comments</a></div></li>
<li><a href='https://www.reddit.com/gallery/nnsq8v'><img src='https://b.thumbs.redditmedia.com/k4CmxLH_rTH3wHkdspJib-8tyMYST1md9979hvufPiA.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://www.reddit.com/gallery/nnsq8v'>A Toyota Coaster based Commuter Draisine Used in Northeastern China by a mining company to transport workers on and off work.</a></div>(reddit.com) posted by <a href='https://www.reddit.com/user/Mr_Camhed'>Mr_Camhed</a> in <a href='https://www.reddit.com/r/WeirdWheels'>WeirdWheels</a> 1082 points & 23 <a href='https://www.reddit.com/r/WeirdWheels/comments/nnsq8v/a_toyota_coaster_based_commuter_draisine_used_in/'>comments</a></div></li>
<li><a href='https://i.redd.it/36jer9tav7271.jpg'><img src='https://a.thumbs.redditmedia.com/-8bC-x4dNEdGpNeKeOf514B1IotYl3WRPZg_EBZLrT8.jpg' alt='link thumbnail'></a><div><div class='linkTitle'><a href='https://i.redd.it/36jer9tav7271.jpg'>Found on interestingasfuck</a></div>(i.redd.it) posted by <a href='https://www.reddit.com/user/EntraByte'>EntraByte</a> in <a href='https://www.reddit.com/r/RetroFuturism'>RetroFuturism</a> 382 points & 12 <a href='https://www.reddit.com/r/RetroFuturism/comments/no6l0y/found_on_interestingasfuck/'>comments</a></div></li>
</ul>
| 203.508772 | 858 | 0.736552 | yue_Hant | 0.303362 |
914716817163ac3062f9d730be9d770c99d9ac83 | 2,033 | md | Markdown | _posts/2011-7-16-UVA12983 The Battle of Chibi.md | FutaRimeWoawaSete/FutaRimeWoawaSete.github.io | 714d0ae43929dc5a4672f82e4c1666fa798d3e38 | [
"MIT"
] | null | null | null | _posts/2011-7-16-UVA12983 The Battle of Chibi.md | FutaRimeWoawaSete/FutaRimeWoawaSete.github.io | 714d0ae43929dc5a4672f82e4c1666fa798d3e38 | [
"MIT"
] | null | null | null | _posts/2011-7-16-UVA12983 The Battle of Chibi.md | FutaRimeWoawaSete/FutaRimeWoawaSete.github.io | 714d0ae43929dc5a4672f82e4c1666fa798d3e38 | [
"MIT"
] | null | null | null | 表面看起来是一道计数题,实际是一道 DP 题。
我们首先设 $dp_{i,j}$ 表示长度为 $i$ 并且以 $j$ 结尾的严格上升子序列的个数,
我们经过一定的推导后可以得到一个 DP 转移式:
- $ dp_{i,j} = \sum_{k = 1}^{j - 1}dp_{i-1,k}$ 其中 $a_k < a_j$
-
照着这个 $DP$ 转移式写上去我们发现这是一个 $O(n ^ 2m)$ 的 DP 很明显我们过不掉这道题。
这时我们考虑如何优化这个 DP 转移式,毕竟 DP 式推出来了但是时间复杂度过不掉的话基本都是需要优化的,我们发现我们可以先离散所有 $a_i$ 然后用线段树维护前面的 $dp_{i,j - 1}$ 。
不过这里需要注意的是,我们由于只能取 $\sum_{k = 1}^{j - 1}$,所以我们必须一个个加。
也就是这样:
```cpp
for(int i = 2 ; i <= m ; i ++)
{
for(int j = 1 ; j <= n ; j ++)
{
add(a[j] , dp[i - 1][j]);
if(a[j] != 1) dp[i][j] = query(a[j] - 1);
if(i == m) Ans += dp[i][j] , Ans %= mod;
}
}
```
~~不过由于线段树还是大常数~~,在本人亲测线段树会被卡常后就换成了树状数组来卡这道题,不得不说这道题还是有点卡……
```cpp
#include<iostream>
#include<cstdio>
#include<cstring>
#include<cmath>
#include<algorithm>
#include<cmath>
using namespace std;
const int Len = 1e3 + 5;
const long long mod = 1e9 + 7;
int cnt,n,m,cnts;
long long c[Len],a[Len],Ans,dp[Len][Len],lsh[Len];
int lowbit(int x){return x & (-x);}
void Clear(){for(int i = 1 ; i <= cnts ; i ++) c[i] = 0;}
void add(int x,long long d){for( ; x <= cnts ; x += lowbit(x)) c[x] += d , c[x] %= mod;}
long long query(int x){long long res = 0;for( ; x ; x -= lowbit(x)) res += c[x] , res %= mod;return res;}
int main()
{
int T;
scanf("%d",&T);
while(T --)
{
cnt ++;
Ans = 0;
scanf("%d %d",&n,&m);
for(int i = 1 ; i <= n ; i ++)
{
scanf("%lld",&a[i]);
lsh[i] = a[i];
dp[1][i] = 1;
}
sort(lsh + 1 , lsh + 1 + n);
cnts = unique(lsh + 1 , lsh + 1 + n) - lsh - 1;
for(int i = 1 ; i <= n ; i ++) a[i] = lower_bound(lsh + 1 , lsh + 1 + cnts , a[i]) - lsh;
if(m == 1) Ans = n;
else
{
Clear();
for(int i = 2 ; i <= m ; i ++)
{
for(int j = 1 ; j <= n ; j ++)
{
add(a[j] , dp[i - 1][j]);
if(a[j] != 1) dp[i][j] = query(a[j] - 1);
if(i == m) Ans += dp[i][j] , Ans %= mod;
}
Clear();
}
}
for(int i = 2 ; i <= m ; i ++)
for(int j = 1 ; j <= n ; j ++) dp[i][j] = 0;
Clear();
printf("Case #%d: %lld\n",cnt,Ans);
}
return 0;
}
``` | 23.367816 | 105 | 0.495819 | eng_Latn | 0.108548 |
9147e13504f9d3d07455230f455e2f8e8e651534 | 1,455 | md | Markdown | README.md | david-mcgillicuddy-moixa/tokio-proto | 0b72a1978064e0e087fd8682ccd1aba6064ad6c0 | [
"Apache-2.0",
"MIT"
] | 334 | 2016-08-27T01:08:35.000Z | 2022-03-16T23:29:19.000Z | README.md | david-mcgillicuddy-moixa/tokio-proto | 0b72a1978064e0e087fd8682ccd1aba6064ad6c0 | [
"Apache-2.0",
"MIT"
] | 174 | 2016-08-27T08:57:26.000Z | 2018-08-01T19:09:59.000Z | README.md | david-mcgillicuddy-moixa/tokio-proto | 0b72a1978064e0e087fd8682ccd1aba6064ad6c0 | [
"Apache-2.0",
"MIT"
] | 100 | 2016-08-27T00:46:49.000Z | 2021-05-14T07:00:32.000Z | # This crate is deprecated!
This crate is deprecated without an immediate replacement. Discussion about a successor can be found in [tokio-rs/tokio#118](https://github.com/tokio-rs/tokio/issues/118).
# tokio-proto
`tokio-proto` makes it easy to implement clients and servers for **request /
response** oriented protocols. It takes a [transport] and provides the request /
response API. It is a part of the [Tokio] platform.
[](https://travis-ci.org/tokio-rs/tokio-proto)
[Documentation](https://docs.rs/tokio-proto) |
[Gitter](https://gitter.im/tokio-rs/tokio) |
[Tutorial](https://tokio.rs)
[transport]: https://tokio.rs/docs/going-deeper-tokio/transports/
[Tokio]: https://tokio.rs
## Usage
First, add this to your `Cargo.toml`:
```toml
[dependencies]
tokio-proto = { git = "https://github.com/tokio-rs/tokio-proto" }
```
Next, add this to your crate:
```rust
extern crate tokio_proto;
```
You can find extensive examples and tutorials at
[https://tokio.rs](https://tokio.rs).
## Getting Help
If you have questions or need further help getting started, consider joining
the chat in our [Gitter Channel](http://gitter.im/tokio-rs/tokio).
## License
Tokio is primarily distributed under the terms of both the MIT license
and the Apache License (Version 2.0), with portions covered by various
BSD-like licenses.
See LICENSE-APACHE, and LICENSE-MIT for details.
| 29.1 | 172 | 0.740893 | eng_Latn | 0.89593 |
91495e76ceaa936675d9d31fa3399a03e5ed91cc | 6,769 | md | Markdown | documents/DesignNotes/appearance.md | hangle/Notecard | fdbed0ce0d15e0288794e18680da7360a0daeed7 | [
"Apache-2.0"
] | null | null | null | documents/DesignNotes/appearance.md | hangle/Notecard | fdbed0ce0d15e0288794e18680da7360a0daeed7 | [
"Apache-2.0"
] | null | null | null | documents/DesignNotes/appearance.md | hangle/Notecard | fdbed0ce0d15e0288794e18680da7360a0daeed7 | [
"Apache-2.0"
] | null | null | null | <h1>Appearance Features </h1>
<p>Appearance features cover the size, style, color, and font style <br />
of text. It also includes the size and length of the input field, <br />
as well as the appearance features of input characters. The <br />
number of input characters can be limited. The window height <br />
and window width are Appearance features. Activation and <br />
deactivation of the '*' and 'PRIOR' buttons are included <br />
as Appearance features. </p>
<p>The Appearance features are established by the default values <br />
of the program as listed below: </p>
<pre>
key value
xlocate 0 //window x screen location
ylocate 0 //window y screen location
height 300 //window size argument
width 400 //window size argument
name TimesRoman // name of Font
size 14 // pixel size of lettering
color black // color of lettering
style 1 // 1=normal, 2=bold, 3=italics
length 10 // input field length
limit 99 // limits the number of input characters
asteriskButton on // "on" allows '* button' to be armed (active)
priorButton on // "on" allow 'PRIOR button' to be armed (active)
</pre>
<p>The script command file:</p>
<pre>
c
d now is the time for all good men
* end
</pre>
<p>The text of the 'd' command is displayed in a window <br />
dimensioned as 300 x 400. The text color is black. Its size <br />
is 14, and its font style is TimesRoman. </p>
<h3>Ways to override default value.</h3>
<p>Asterisk Commands. The '*' command with an appearance <br />
key/value pair, supercedes the corresponding program <br />
default value. </p>
<p>The script command file: (ParserValidator.AsteriskCommand)</p>
<pre>
* width 500
* height 600
* size 22
c
d now is the time for all good men
* end
</pre>
<p>The window size changes to 500 x 600, and the text size becomes <br />
'22'. The text color is still black and the font size is still <br />
TimesRoman. However, the introduction of the following '*' <br />
commands can change this. </p>
<pre>
* color green
* name Arial
</pre>
<p>The '*' commands of the script file remains in effect until <br />
the next script file is executed. </p>
<h3>The appearance.ini file.</h3>
<p>The 'appearance.ini' file provides values to the script program. <br />
It does not directly involve the Notecard program. The script <br />
program uses the 'appearance.ini' to create the '.struct' files <br />
that are employed by the Notecard program. </p>
<p>The 'appearance.ini' file, consisting of key/value pairs. <br />
These key/value pairs change the default values for all <br />
script files in a session. The following is an example of an <br />
'appearance.ini' file. </p>
<pre>
appearance.ini
____________________________
| height 320
| width 260
| size 16
| color green
</pre>
<p>The file changes four default, appearance values. These <br />
values become new default values overriding the program <br />
values. They remain in effect for the session's script <br />
files ( the '* end' command terminates a session). </p>
<p>The directory in which the Notecard program is executed is <br />
termed the 'home' directory. The 'appearance.ini' file in <br />
the 'home' directory is applied to the script files in 'home' <br />
directory and to all script files in its subdirectories. <br />
An 'appearance.ini' directory within a 'home' subdirectory <br />
is applied to just the script files in that directory. </p>
<p>In the event that 'appearance.ini' files are in both the <br />
'home' directory and in a 'home' subdirectory, then the <br />
subdirectory 'ini' file supercedes the 'home' 'ini' file, <br />
but only for script files in the subdirectory. </p>
<p>The Appearance feature of an '*' command overrides a <br />
corresponding feature of the 'appearance.ini' file. <br />
For example, the following 'appearance.ini' file in the <br />
'home' directory is: </p>
<pre>
appearance.ini
__________________
| size 12
</pre>
<p>The initial default value of 14 is becomes 12 for all <br />
script files in the 'home' directory and its
subdirectories. </p>
<p>Next, an 'appeaance.ini' file in a subdirectory is:</p>
<pre>
appearance.ini
____________________
| size 10
</pre>
<p>The previous defaults letter size of 14 and 12 become 10 for <br />
just the script files in the subdirectory. In the 'home' <br />
directory and its other subdirectories, the default value <br />
is 12. </p>
<p>Finally the following '*' command is added to a script file <br />
in the subdirectory whose 'appearance.ini' file specifies a <br />
size is 10 : </p>
<pre>
* size 22
</pre>
<p>The default letter size is 22 for text in the script file <br />
having the '*' size 22 command. The other script files <br />
have a default letter size of 10. </p>
<h3>Two Types of '*' commands.</h3>
<p>The '<em>' Appearance commands, such as, '</em> size 22', are <br />
not executed by the Notecard program. The Notecard program <br />
executes the following '*' commands: </p>
<pre>
* end
* continue
* save
* manage <filename>
</pre>
<p>The '*' Appearance commands are consumed in the Script <br />
program and are removed. </p>
<h3>Script Program</h3>
<p>The default values in AsteriskCommand.defaultSetting:Map are copied to
'overrideSetting'. </p>
<pre><code>val defaultSetting=Map (
"height"-> "300", //window size argument
"width"-> "400", //window size argument
"name"-> "TimesRoman",// name of Font
"size"-> "14", // pixel size of lettering
"color"-> "black", // color of lettering
"style"-> "0", // 0=normal, 1=bold, 2=italics
"length"-> "10", // used in Display cmd for BoxField
"limit"-> "99", // used in Display cmd for BoxField
"column"-> "0", // not operational
"manage"-> "task" , // to create FramerTask
"asteriskButton"-> "on", // "on" allows '* button' to be armed (active)
"save"-> "task" , // n to create FramerTask-- save symbolTable data
"priorButton"-> "on" , // "on" allow 'PRIOR button' to be armed (active)
"end"-> "task", // to create FramerTask-- terminate session
"status"->"task", // display msg in status field
"continue"-> "task" // to create CardSetTask--
)
</code></pre>
| 34.712821 | 80 | 0.637465 | eng_Latn | 0.992658 |
914a52608291f40eb2adef88dd684802888b5b29 | 126 | md | Markdown | CHANGELOG.md | cxfans/ftplib | 9b5bf9fa0f314c294d180380f6a891b581378342 | [
"MIT"
] | 1 | 2020-04-13T19:20:08.000Z | 2020-04-13T19:20:08.000Z | CHANGELOG.md | cxfans/ftplib | 9b5bf9fa0f314c294d180380f6a891b581378342 | [
"MIT"
] | null | null | null | CHANGELOG.md | cxfans/ftplib | 9b5bf9fa0f314c294d180380f6a891b581378342 | [
"MIT"
] | 1 | 2020-07-11T08:53:28.000Z | 2020-07-11T08:53:28.000Z | # Change Log of ftplib Library
## [0.1.0] - 2019-11-8
### Release
- Implement basic function for File Transfer Protocol (FTP) | 25.2 | 59 | 0.706349 | kor_Hang | 0.431979 |
914aa84b6b10631b7936d349a26ff6ab8741f004 | 596 | md | Markdown | qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md | DanTupi/personal_setup | 911b4951e4d8b78d6ea8ca335229e2e970fda871 | [
"MIT"
] | null | null | null | qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md | DanTupi/personal_setup | 911b4951e4d8b78d6ea8ca335229e2e970fda871 | [
"MIT"
] | null | null | null | qmk_firmware/keyboards/keebio/quefrency/keymaps/bcat/readme.md | DanTupi/personal_setup | 911b4951e4d8b78d6ea8ca335229e2e970fda871 | [
"MIT"
] | null | null | null | # bcat's Quefrency 65% layout
This is a standard 65% keyboard layout, with a split spacebar, an HHKB-style
(split) backspace, media controls in the function layer (centered around the
ESDF cluster), and RGB controls in the function layer (on the arrow/nav keys).
## Default layer

([KLE](http://www.keyboard-layout-editor.com/#/gists/60a262432bb340b37d364a4424f3037b))
## Function layer

([KLE](http://www.keyboard-layout-editor.com/#/gists/59636898946da51f91fb290f8e078b4d))
| 33.111111 | 87 | 0.761745 | eng_Latn | 0.549269 |
914b1662630f79b0c8d5c356f1f4dc2ab8473602 | 2,082 | md | Markdown | README.md | facorazza/Loot_Quotes_Bot | 202e3e6b163b92d67af578203501a4857596a64c | [
"Apache-2.0"
] | null | null | null | README.md | facorazza/Loot_Quotes_Bot | 202e3e6b163b92d67af578203501a4857596a64c | [
"Apache-2.0"
] | null | null | null | README.md | facorazza/Loot_Quotes_Bot | 202e3e6b163b92d67af578203501a4857596a64c | [
"Apache-2.0"
] | null | null | null | # Loot_Quotes_Bot
### Media dei prezzi
L'approssimazione più sostanziale di tutto il processo riguarda il calcolo della media dei prezzi degli oggetti. Essendo i prezzi una misura dipendente dal tempo e dai valori passati degli stessi, l'insieme considerato non può essere considerato un **campione**. Per ovviare a ciò utilizzeremo i valori contenuti nell'intervallo di tempo più ristretto possibile, cercando comunque di mantenere una numerosità sostanziale per ogni singolo oggetto, in modo tale da eliminare, o almeno da ridurre drasticamente, la dipendenza tra le misure passate e da rendere la distribuzione il più uniforme possibile per ogni singolo oggetto.
### Incertezza
L'incertezza delle quotazioni viene calcolata attraverso la seguente formula:
*Formula*
Il valore trovato deve essere aggiunto e sottratto al valore della quotazione, trovando, in questo modo, un intervallo di confidenza entro cui il prezzo reale può cadere.
*Intervallo di confidenza:* ksigma
L'incertezza mostrata dal bot utilizza k=1 che indica una confidenza di circa (simbolo)il 68% che il prezzo cada nell'intervallo.
Per esempio:
Raddoppiando il valore di k la confidenza dell'intervallo sale a circa il (simbolo)95%.
Quindi, riprendendo l'esempio precedente:
È possibile aumentare k ulteriormente ma ovviamente ne consegue un aumento dell'intervallo.
### Mediana e quantili
Il bot riordina i dati per prezzo in modo crescente e ne calcola la mediana. La mediana è il valore che si trova in mezzo e che divide i prezzi in due insiemi equipartiti. Analogamente il primo e il terzo quartile dividono ulteriormente i due sottoinsiemi appena creati. Questa suddivisione permette di avere un'idea della distribuzione dei dati. Infatti è possibile constatare che il 50% dei prezzi è contenuto nell'intervallo tra il *primo* e il *terzo quartile*. Conoscendo inoltre il prezzo *minimo* si deduce che l'intervallo più piccolo contenete il 25% dei dati si trova tra il minimo e il primo quartile. Analogamente tutto ciò vale anche per l'intervallo superiore, se si conosce il *massimo* dell'insieme. | 115.666667 | 715 | 0.8122 | ita_Latn | 0.999969 |
914b93198659312b955bb474aa94a5dce7583f1e | 7,971 | md | Markdown | docs/usage/validation_decorator.md | jasujm/pydantic | cc1cb4826c74ac5b651ef2d80c3478428a9950ca | [
"MIT"
] | 6 | 2021-08-11T11:37:59.000Z | 2021-11-12T01:33:11.000Z | docs/usage/validation_decorator.md | jasujm/pydantic | cc1cb4826c74ac5b651ef2d80c3478428a9950ca | [
"MIT"
] | 189 | 2020-07-12T08:13:29.000Z | 2022-03-28T01:16:29.000Z | docs/usage/validation_decorator.md | jasujm/pydantic | cc1cb4826c74ac5b651ef2d80c3478428a9950ca | [
"MIT"
] | 2 | 2021-11-23T16:28:21.000Z | 2021-11-23T16:28:33.000Z | The `validate_arguments` decorator allows the arguments passed to a function to be parsed and validated using
the function's annotations before the function is called. While under the hood this uses the same approach of model
creation and initialisation; it provides an extremely easy way to apply validation to your code with minimal
boilerplate.
!!! info "In Beta"
The `validate_arguments` decorator is in **beta**, it has been added to *pydantic* in **v1.5** on a
**provisional basis**. It may change significantly in future releases and its interface will not be concrete
until **v2**. Feedback from the community while it's still provisional would be extremely useful; either comment
on [#1205](https://github.com/samuelcolvin/pydantic/issues/1205) or create a new issue.
Example of usage:
```py
{!.tmp_examples/validation_decorator_main.py!}
```
_(This script is complete, it should run "as is")_
## Argument Types
Argument types are inferred from type annotations on the function, arguments without a type decorator are considered
as `Any`. Since `validate_arguments` internally uses a standard `BaseModel`, all types listed in
[types](types.md) can be validated, including *pydantic* models and [custom types](types.md#custom-data-types).
As with the rest of *pydantic*, types can be coerced by the decorator before they're passed to the actual function:
```py
{!.tmp_examples/validation_decorator_types.py!}
```
_(This script is complete, it should run "as is")_
A few notes:
- though they're passed as strings, `path` and `regex` are converted to a `Path` object and regex respectively
by the decorator
- `max` has no type annotation, so will be considered as `Any` by the decorator
Type coercion like this can be extremely helpful but also confusing or not desired,
see [below](#coercion-and-strictness) for a discussion of `validate_arguments`'s limitations in this regard.
## Function Signatures
The decorator is designed to work with functions using all possible parameter configurations and all possible
combinations of these:
* positional or keyword arguments with or without defaults
* variable positional arguments defined via `*` (often `*args`)
* variable keyword arguments defined via `**` (often `**kwargs`)
* keyword only arguments - arguments after `*,`
* positional only arguments - arguments before `, /` (new in python 3.8)
To demonstrate all the above parameter types:
```py
{!.tmp_examples/validation_decorator_parameter_types.py!}
```
_(This script is complete, it should run "as is")_
## Using Field to describe function arguments
[Field](schema.md#field-customisation) can also be used with `validate_arguments` to provide extra information about
the field and validations. In general it should be used in a type hint with
[Annotated](schema.md#typingannotated-fields), unless `default_factory` is specified, in which case it should be used
as the default value of the field:
```py
{!.tmp_examples/validation_decorator_field.py!}
```
_(This script is complete, it should run "as is")_
## Usage with mypy
The `validate_arguments` decorator should work "out of the box" with [mypy](http://mypy-lang.org/) since it's
defined to return a function with the same signature as the function it decorates. The only limitation is that
since we trick mypy into thinking the function returned by the decorator is the same as the function being
decorated; access to the [raw function](#raw-function) or other attributes will require `type: ignore`.
## Validate without calling the function
By default, arguments validation is done by directly calling the decorated function with parameters.
But what if you wanted to validate them without *actually* calling the function?
To do that you can call the `validate` method bound to the decorated function.
```py
{!.tmp_examples/validation_decorator_validate.py!}
```
_(This script is complete, it should run "as is")_
## Raw function
The raw function which was decorated is accessible, this is useful if in some scenarios you trust your input
arguments and want to call the function in the most performant way (see [notes on performance](#performance) below):
```py
{!.tmp_examples/validation_decorator_raw_function.py!}
```
_(This script is complete, it should run "as is")_
## Async Functions
`validate_arguments` can also be used on async functions:
```py
{!.tmp_examples/validation_decorator_async.py!}
```
## Custom Config
The model behind `validate_arguments` can be customised using a config setting which is equivalent to
setting the `Config` sub-class in normal models.
!!! warning
The `fields` and `alias_generator` properties of `Config` which allow aliases to be configured are not supported
yet with `@validate_arguments`, using them will raise an error.
Configuration is set using the `config` keyword argument to the decorator, it may be either a config class
or a dict of properties which are converted to a class later.
```py
{!.tmp_examples/validation_decorator_config.py!}
```
_(This script is complete, it should run "as is")_
## Limitations
`validate_arguments` has been released on a provisional basis without all the bells and whistles, which may
be added later, see [#1205](https://github.com/samuelcolvin/pydantic/issues/1205) for some more discussion of this.
In particular:
### Validation Exception
Currently upon validation failure, a standard *pydantic* `ValidationError` is raised,
see [model error handling](models.md#error-handling).
This is helpful since it's `str()` method provides useful details of the error which occurred and methods like
`.errors()` and `.json()` can be useful when exposing the errors to end users, however `ValidationError` inherits
from `ValueError` **not** `TypeError` which may be unexpected since python would raise a `TypeError` upon invalid
or missing arguments. This may be addressed in future by either allow a custom error or raising a different
exception by default, or both.
### Coercion and Strictness
*pydantic* currently leans on the side of trying to coerce types rather than raise an error if a type is wrong,
see [model data conversion](models.md#data-conversion) and `validate_arguments` is no different.
See [#1098](https://github.com/samuelcolvin/pydantic/issues/1098) and other issues with the "strictness" label
for a discussion of this. If *pydantic* gets a "strict" mode in future, `validate_arguments` will have an option
to use this, it may even become the default for the decorator.
### Performance
We've made a big effort to make *pydantic* as performant as possible (see [the benchmarks](../benchmarks.md))
and argument inspect and model creation is only performed once when the function is defined, however
there will still be a performance impact to using the `validate_arguments` decorator compared to
calling the raw function.
In many situations this will have little or no noticeable effect, however be aware that
`validate_arguments` is not an equivalent or alternative to function definitions in strongly typed languages;
it never will be.
### Return Value
The return value of the function is not validated against its return type annotation, this may be added as an option
in future.
### Config and Validators
`fields` and `alias_generator` on custom [`Config`](model_config.md) are not supported, see [above](#custom-config).
Neither are [validators](validators.md).
### Model fields and reserved arguments
The following names may not be used by arguments since they can be used internally to store information about
the function's signature:
* `v__args`
* `v__kwargs`
* `v__positional_only`
These names (together with `"args"` and `"kwargs"`) may or may not (depending on the function's signature) appear as
fields on the internal *pydantic* model accessible via `.model` thus this model isn't especially useful
(e.g. for generating a schema) at the moment.
This should be fixable in future as the way error are raised is changed.
| 42.854839 | 117 | 0.774056 | eng_Latn | 0.998529 |
914b988cd43f3764aede371c8d9a52b5585c1d56 | 3,516 | md | Markdown | api/Publisher.Shapes.AddTextEffect.md | RichardCory/VBA-Docs | 1240462311fb77ee051d4e8b7d7a434d7d020dd3 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2020-03-09T13:24:12.000Z | 2020-03-09T16:19:11.000Z | api/Publisher.Shapes.AddTextEffect.md | MarkFern/VBA-Docs | b84627cc8e24acfd336d1e9761a9ddd58f19d352 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | api/Publisher.Shapes.AddTextEffect.md | MarkFern/VBA-Docs | b84627cc8e24acfd336d1e9761a9ddd58f19d352 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Shapes.AddTextEffect method (Publisher)
keywords: vbapb10.chm2162721
f1_keywords:
- vbapb10.chm2162721
ms.prod: publisher
api_name:
- Publisher.Shapes.AddTextEffect
ms.assetid: 21af82f1-d507-3c16-72df-bde1b5e00717
ms.date: 06/08/2017
localization_priority: Normal
---
# Shapes.AddTextEffect method (Publisher)
Adds a new **Shape** object representing a WordArt object to the specified **Shapes** collection.
## Syntax
_expression_.**AddTextEffect**(**_PresetTextEffect_**, **_Text_**, **_FontName_**, **_FontSize_**, **_FontBold_**, **_FontItalic_**, **_Left_**, **_Top_**)
_expression_ A variable that represents a **Shapes** object.
## Parameters
|Name|Required/Optional|Data type|Description|
|:-----|:-----|:-----|:-----|
|PresetTextEffect|Required| **MsoPresetTextEffect**|The preset text effect to use. The values of the **MsoPresetTextEffect** constants correspond to the formats listed in the **WordArt Gallery** dialog box (numbered from left to right and from top to bottom).|
|Text|Required| **String**|The text to use for the WordArt object.|
|FontName|Required| **String**|The name of the font to use for the WordArt object.|
|FontSize|Required| **Variant**|The font size to use for the WordArt object. Numeric values are evaluated in points; strings can be in any units supported by Microsoft Publisher (for example, "2.5 in").|
|FontBold|Required| **MsoTriState**|Determines whether to format the WordArt text as bold.|
|FontItalic|Required| **MsoTriState**|Determines whether to format the WordArt text as italic.|
|Left|Required| **Variant**|The position of the left edge of the shape representing the WordArt object.|
|Top|Required| **Variant**|The position of the top edge of the shape representing the WordArt object.|
## Return value
Shape
## Remarks
For the Left and Top parameters, numeric values are evaluated in points; strings can be in any units supported by Publisher (for example, "2.5 in").
The height and width of the WordArt object is determined by its text and formatting.
Use the **[TextEffect](Publisher.Shape.TextEffect.md)** property to return a **[TextEffectFormat](Publisher.TextEffectFormat.md)** object whose properties can be used to edit an existing WordArt object.
The PresetTextEffect parameter can be one of the **[MsoPresetTextEffect](Office.MsoPresetTextEffect.md)** constants declared in the Microsoft Office type library. The **msoTextEffectMixed** constant is not supported.
The FontBold parameter can be one of the **MsoTriState** constants declared in the Microsoft Office type library and shown in the following table.
|Constant|Description|
|:-----|:-----|
| **msoFalse**|Do not format the WordArt text as bold.|
| **msoTrue**|Format the WordArt text as bold.|
The FontItalic parameter can be one of the **MsoTriState** constants declared in the Microsoft Office type library and shown in the following table.
|Constant|Description|
|:-----|:-----|
| **msoFalse**| Do not format the WordArt text as italic.|
| **msoTrue**|Format the WordArt text as italic.|
## Example
The following example adds a WordArt object to the first page of the active publication.
```vb
Dim shpWordArt As Shape
Set shpWordArt = ActiveDocument.Pages(1).Shapes.AddTextEffect _
(PresetTextEffect:=msoTextEffect7, Text:="Annual Report", _
FontName:="Arial Black", FontSize:=24, _
FontBold:=msoFalse, FontItalic:=msoFalse, _
Left:=144, Top:=72)
```
[!include[Support and feedback](~/includes/feedback-boilerplate.md)] | 38.637363 | 261 | 0.746303 | eng_Latn | 0.865378 |
914ddc6977963a3eacf07066bdc5ac8eda5aabf3 | 585 | markdown | Markdown | website/docs/r/thunder_ip_tcp.html.markdown | a10networks/terraform-provider-thunder | 50fe189add4fc51ca17b648945e63685bf350177 | [
"BSD-2-Clause"
] | 4 | 2020-10-17T00:07:06.000Z | 2021-09-11T21:44:42.000Z | website/docs/r/thunder_ip_tcp.html.markdown | a10networks/terraform-provider-thunder | 50fe189add4fc51ca17b648945e63685bf350177 | [
"BSD-2-Clause"
] | 5 | 2020-10-09T06:47:26.000Z | 2021-09-11T21:44:26.000Z | website/docs/r/thunder_ip_tcp.html.markdown | a10networks/terraform-provider-thunder | 50fe189add4fc51ca17b648945e63685bf350177 | [
"BSD-2-Clause"
] | 3 | 2020-10-13T06:09:53.000Z | 2021-12-03T15:29:08.000Z | ---
layout: "thunder"
page_title: "thunder: thunder_ip_tcp"
sidebar_current: "docs-thunder-resource-ip-tcp"
description: |-
Provides details about thunder ip tcp resource for A10
---
# thunder\_ip\_tcp
`thunder_ip_tcp` Provides details about thunder ip tcp
## Example Usage
```hcl
provider "thunder" {
address = "192.0.2.65"
username = "admin"
password = "admin"
}
resource "thunder_ip_tcp" "Iptcp" {
syn_cookie {
threshold = 4
}
}
```
## Argument Reference
* `uuid` - uuid of the object
* `threshold` - SYN cookie expire threshold (seconds (default is 4))
| 17.205882 | 68 | 0.688889 | eng_Latn | 0.724293 |
914e4741c8b60c327f55637612ed31a778e66072 | 2,891 | md | Markdown | programador/preferencia-tipos-dominio-especifico.md | jaimerodas/97cosas | c8f2f7967ca53e58d4eb04d73ba89d474f23c5eb | [
"CC-BY-3.0"
] | 44 | 2015-04-02T14:05:21.000Z | 2022-02-02T08:34:40.000Z | programador/preferencia-tipos-dominio-especifico.md | jaimerodas/97cosas | c8f2f7967ca53e58d4eb04d73ba89d474f23c5eb | [
"CC-BY-3.0"
] | 13 | 2015-06-17T23:47:28.000Z | 2019-10-30T06:23:25.000Z | programador/preferencia-tipos-dominio-especifico.md | esparta/97cosas | f52357df922fea12abe798d9836b5c5121f5732f | [
"CC-BY-3.0"
] | 21 | 2015-04-02T17:49:10.000Z | 2021-06-09T00:19:03.000Z | ---
layout: programador
title: Da preferencia a tipos de Dominio Específico que los tipos primitivos
overview: true
author: Einar Landre
translator: Espartaco Palma
original: https://web.archive.org/web/20150106001512/http://programmer.97things.oreilly.com/wiki/index.php/Prefer_Domain-Specific_Types_to_Primitive_Types
---
El 23 de septiembre de 1999 el _Mars Climate Orbiter_ de U$327.6
millones se perdió mientras entraba a la órbita alrededor de Marte,
debido a un error del software aquí en la Tierra. Error que más tarde
fue llamado de “métrica mixta”. El software de la estación en tierra
estaba trabajando en libras, mientras que la nave esperaba newtons,
llevando a la estación a subestimar el poder de los propulsores de la
nave en un factor de 4.45.
Éste es uno de los muchos ejemplos de fallas de software que se pudo
haber prevenido, si se hubiera aplicado un _tipado_ más fuerte y de
dominio específico. Es también un ejemplo del razonamiento detrás de
muchas características del lenguaje Ada, uno de sus principales metas de
diseño era implementar software de seguridad crítica embebida. Ada
estaba fuertemente _tipado_ con revisiones estáticas de ambos: tipos
primitivos y tipos definidos por el usuario.
type Velocity_In_Knots is new Float range 0.0 .. 500.00;
type Distance_In_Nautical_Miles is new Float range 0.0 .. 3000.00;
Velocity: Velocity_In_Knots;
Distance: Distance_In_Nautical_Miles;
Some_Number: Float;
Some_Number:= Distance + Velocity; -- Será capturado por el compilador como un error de tipos.
Los desarrolladores en dominios menos demandantes también se deberían
beneficiar aplicando más _tipado_ de dominio específico, en el que
pudieran, de otro modo, continuar usando tipos de datos primitivos
ofrecidos por el lenguaje y sus librerías, tales como cadenas y
flotantes. En Java, C++, Python y otros lenguajes modernos, los tipos de
datos abstractos son conocidos como clases. Usar clases como
`Velocity_In_Knots` y `Distance_In_Nautical_Miles` agrega mucho valor
con respecto a la calidad del código:
* El código se vuelve más legible conforme expresa conceptos de un
dominio, no sólo flotantes o cadenas.
* El código se vuelve más _testeable_ conforme encapsula su
comportamiento, así es fácilmente probado.
* El código facilita la reutilización a través de aplicaciones y sistemas.
El enfoque es igualmente válido para usuarios de ambos lenguajes de tipo
estático y dinámico. La única diferencia es que los desarrolladores que
usan lenguajes de _tipado_ estático obtienen más ayuda desde el
compilados, mientras aquellos que adoptan lenguajes de _tipado_ dinámico
es más común que confíen en sus pruebas unitarias. El estilo de revisión
podría ser diferente, pero la motivación y estilo de expresión no.
La moraleja es iniciar explorando los tipos de dominio específico con el
fin de desarrollar software de calidad.
| 46.629032 | 154 | 0.808025 | spa_Latn | 0.996094 |
914e56fe039e6943cc20d26fc4eb02d8d016a707 | 1,102 | md | Markdown | docs/src/SUMMARY.md | roelvdberg/metacontroller | a9ca3730cfba5bfd544f9b65bf42a86f187e478f | [
"Apache-2.0"
] | 375 | 2020-05-21T10:41:15.000Z | 2022-03-29T10:23:08.000Z | docs/src/SUMMARY.md | roelvdberg/metacontroller | a9ca3730cfba5bfd544f9b65bf42a86f187e478f | [
"Apache-2.0"
] | 391 | 2020-05-19T09:33:07.000Z | 2022-03-31T17:27:18.000Z | docs/src/SUMMARY.md | roelvdberg/metacontroller | a9ca3730cfba5bfd544f9b65bf42a86f187e478f | [
"Apache-2.0"
] | 52 | 2020-05-26T03:03:42.000Z | 2022-02-28T03:26:13.000Z | # Summary
- [Getting Started](./intro.md)
- [Introduction](./intro.md)
- [Examples](./examples.md)
- [Concepts](./concepts.md)
- [Features](./features.md)
- [FAQ](./faq.md)
- [Pronunciation](./pronunciation.md)
- [User Guide](./guide.md)
- [Install Metacontroller](./guide/install.md)
- [Install Metacontroller via helm](./guide/helm-install.md)
- [Configuration](./guide/configuration.md)
- [Create a controller](./guide/create.md)
- [Constraints and best practices](./guide/best-practices.md)
- [Troubleshooting](./guide/troubleshooting.md)
- [API Reference](./api.md)
- [Apply Semantics](./api/apply.md)
- [CompositeController](./api/compositecontroller.md)
- [ControllerRevision](./api/controllerrevision.md)
- [DecoratorController](./api/decoratorcontroller.md)
- [Customize Hook](./api/customize.md)
- [Hook](./api/hook.md)
- [Design Docs](./design.md)
- [MapController](./design/map-controller.md)
- [Contributing](./contrib.md)
- [Building](./contrib/build.md)
- [Local development/debug](./contrib/debug.md)
| 38 | 65 | 0.65245 | yue_Hant | 0.539141 |
914e890592492434c7211dafa5cee8f98e0e40dd | 1,390 | md | Markdown | _listings/aws-cloudformation/action-deletestack-get.md | streamdata-gallery-organizations/aws-cloudformation | 4f147d52cf41ec5b07c16a312dff8421187d5ed6 | [
"CC-BY-3.0"
] | null | null | null | _listings/aws-cloudformation/action-deletestack-get.md | streamdata-gallery-organizations/aws-cloudformation | 4f147d52cf41ec5b07c16a312dff8421187d5ed6 | [
"CC-BY-3.0"
] | null | null | null | _listings/aws-cloudformation/action-deletestack-get.md | streamdata-gallery-organizations/aws-cloudformation | 4f147d52cf41ec5b07c16a312dff8421187d5ed6 | [
"CC-BY-3.0"
] | null | null | null | ---
swagger: "2.0"
info:
title: AWS CloudFormation API
version: 1.0.0
schemes:
- http
produces:
- application/json
consumes:
- application/json
paths:
/?Action=DeleteStack:
get:
summary: ' Delete Stack '
description: Deletes a specified stack
operationId: deleteStack
parameters:
- in: query
name: RetainResources.member.N
description: For stacks in the DELETE_FAILED state, a list of resource logical
IDs that are associated with the resources you want to retain
type: string
- in: query
name: RoleARN
description: The Amazon Resource Name (ARN) of an AWS Identity and Access
Management (IAM) role that AWS CloudFormation assumes to delete
the stack
type: string
- in: query
name: StackName
description: The name or the unique stack ID that is associated with the stack
type: string
responses:
200:
description: OK
tags:
- stacks
definitions: []
x-collection-name: AWS CloudFormation
x-streamrank:
polling_total_time_average: 0
polling_size_download_average: 0
streaming_total_time_average: 0
streaming_size_download_average: 0
change_yes: 0
change_no: 0
time_percentage: 0
size_percentage: 0
change_percentage: 0
last_run: ""
days_run: 0
minute_run: 0
--- | 25.740741 | 86 | 0.661871 | eng_Latn | 0.721251 |
914e93e1a084f45b091eaea99f5fbd352f8f8de4 | 26,470 | md | Markdown | common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md | eltociear/common-data-model-and-service | b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd | [
"CC-BY-4.0",
"MIT"
] | null | null | null | common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md | eltociear/common-data-model-and-service | b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd | [
"CC-BY-4.0",
"MIT"
] | null | null | null | common-data-model/schema/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.md | eltociear/common-data-model-and-service | b57ad3a5b8e41699eabb5c424f3c4985d52ac9dd | [
"CC-BY-4.0",
"MIT"
] | 1 | 2020-05-28T15:59:53.000Z | 2020-05-28T15:59:53.000Z | ---
title: BankStatementEntity - Common Data Model | Microsoft Docs
description: undefined
author: nenad1002
ms.service: common-data-model
ms.reviewer: deonhe
ms.topic: article
ms.date: 5/5/2020
ms.author: nebanfic
---
# Bank statement account statement
Latest version of the JSON entity definition is available on <a href="https://github.com/Microsoft/CDM/tree/master/schemaDocuments/core/operationsCommon/Entities/Finance/CashAndBankManagement/BankStatementEntity.cdm.json" target="_blank">GitHub</a>.
## Traits
<details>
<summary>Traits for this entity are listed below.
</summary>
**is.CDM.entityVersion**
<table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>versionNumber</td><td>"1.0.0"</td><td>string</td><td>semantic version number of the entity</td></tr></table>
**is.application.releaseVersion**
<table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>releaseVersion</td><td>"10.0.13.0"</td><td>string</td><td>semantic version number of the application introducing this entity</td></tr></table>
**is.localized.displayedAs**
Holds the list of language specific display text for an object. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>localizedDisplayText</td><td><table><tr><th>languageTag</th><th>displayText</th></tr><tr><td>en</td><td>Bank statement account statement</td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of localized text</td></tr></table>
</details>
## Attributes
|Name|Description|First Included in Instance|
|---|---|---|
|[AccountCurrency](#AccountCurrency)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[AccountIdOtherId](#AccountIdOtherId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[AsOfDateModifier](#AsOfDateModifier)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[BankStatementType](#BankStatementType)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[FromDateTime](#FromDateTime)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[GroupControlTotal](#GroupControlTotal)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[GroupStatus](#GroupStatus)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[IBAN](#IBAN)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[Identification](#Identification)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[SwiftNo](#SwiftNo)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[ToDateTime](#ToDateTime)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[TotalAccountsInGroup](#TotalAccountsInGroup)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[TotalRecordsInGroup](#TotalRecordsInGroup)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[TotalEntriesNumberOfEntries](#TotalEntriesNumberOfEntries)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[OriginatorName](#OriginatorName)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[OriginatorId](#OriginatorId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[UltimateReceiverName](#UltimateReceiverName)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[UltimateReceiverId](#UltimateReceiverId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[OriginatorParty](#OriginatorParty)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[UltimateReceiverParty](#UltimateReceiverParty)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[StatementLineNum](#StatementLineNum)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[DocumentLineNum](#DocumentLineNum)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[BankStmtISODocument](#BankStmtISODocument)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[BankStatementRecId](#BankStatementRecId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[BankAccountTable](#BankAccountTable)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[SequenceNumber](#SequenceNumber)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[Relationship_BankStatementDocumentEntityRelationshipId](#Relationship_BankStatementDocumentEntityRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[BackingTable_BankStmtISOAccountStatementRelationshipId](#BackingTable_BankStmtISOAccountStatementRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
|[Relationship_PrimaryCompanyContextRelationshipId](#Relationship_PrimaryCompanyContextRelationshipId)||<a href="BankStatementEntity.md" target="_blank">CashAndBankManagement/BankStatementEntity</a>|
### <a href=#AccountCurrency name="AccountCurrency">AccountCurrency</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the AccountCurrency attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#AccountIdOtherId name="AccountIdOtherId">AccountIdOtherId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the AccountIdOtherId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#AsOfDateModifier name="AsOfDateModifier">AsOfDateModifier</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the AsOfDateModifier attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#BankStatementType name="BankStatementType">BankStatementType</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the BankStatementType attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#FromDateTime name="FromDateTime">FromDateTime</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the FromDateTime attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#GroupControlTotal name="GroupControlTotal">GroupControlTotal</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the GroupControlTotal attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#GroupStatus name="GroupStatus">GroupStatus</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the GroupStatus attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#IBAN name="IBAN">IBAN</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the IBAN attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#Identification name="Identification">Identification</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the Identification attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#SwiftNo name="SwiftNo">SwiftNo</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the SwiftNo attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#ToDateTime name="ToDateTime">ToDateTime</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the ToDateTime attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#TotalAccountsInGroup name="TotalAccountsInGroup">TotalAccountsInGroup</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the TotalAccountsInGroup attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#TotalRecordsInGroup name="TotalRecordsInGroup">TotalRecordsInGroup</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the TotalRecordsInGroup attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#TotalEntriesNumberOfEntries name="TotalEntriesNumberOfEntries">TotalEntriesNumberOfEntries</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the TotalEntriesNumberOfEntries attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#OriginatorName name="OriginatorName">OriginatorName</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the OriginatorName attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#OriginatorId name="OriginatorId">OriginatorId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the OriginatorId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#UltimateReceiverName name="UltimateReceiverName">UltimateReceiverName</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the UltimateReceiverName attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#UltimateReceiverId name="UltimateReceiverId">UltimateReceiverId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the UltimateReceiverId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#OriginatorParty name="OriginatorParty">OriginatorParty</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the OriginatorParty attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#UltimateReceiverParty name="UltimateReceiverParty">UltimateReceiverParty</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the UltimateReceiverParty attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#StatementLineNum name="StatementLineNum">StatementLineNum</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the StatementLineNum attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#DocumentLineNum name="DocumentLineNum">DocumentLineNum</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the DocumentLineNum attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#BankStmtISODocument name="BankStmtISODocument">BankStmtISODocument</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the BankStmtISODocument attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#BankStatementRecId name="BankStatementRecId">BankStatementRecId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the BankStatementRecId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#BankAccountTable name="BankAccountTable">BankAccountTable</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the BankAccountTable attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#SequenceNumber name="SequenceNumber">SequenceNumber</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>string</td></tr><tr><td>isNullable</td><td>true</td></tr></table>
#### Traits
<details>
<summary>List of traits for the SequenceNumber attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.nullable**
The attribute value may be set to NULL.
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#Relationship_BankStatementDocumentEntityRelationshipId name="Relationship_BankStatementDocumentEntityRelationshipId">Relationship_BankStatementDocumentEntityRelationshipId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table>
#### Traits
<details>
<summary>List of traits for the Relationship_BankStatementDocumentEntityRelationshipId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.dataFormat.guid**
**means.identity.entityId**
**is.linkedEntity.identifier**
Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td>empty table</td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table>
**is.dataFormat.guid**
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#BackingTable_BankStmtISOAccountStatementRelationshipId name="BackingTable_BankStmtISOAccountStatementRelationshipId">BackingTable_BankStmtISOAccountStatementRelationshipId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table>
#### Traits
<details>
<summary>List of traits for the BackingTable_BankStmtISOAccountStatementRelationshipId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.dataFormat.guid**
**means.identity.entityId**
**is.linkedEntity.identifier**
Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td><table><tr><th>entityReference</th><th>attributeReference</th></tr><tr><td><a href="../../../Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.md" target="_blank">/core/operationsCommon/Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.cdm.json/BankStmtISOAccountStatement</a></td><td><a href="../../../Tables/Finance/Bank/Transaction/BankStmtISOAccountStatement.md#RecId" target="_blank">RecId</a></td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table>
**is.dataFormat.guid**
**is.dataFormat.character**
**is.dataFormat.array**
</details>
### <a href=#Relationship_PrimaryCompanyContextRelationshipId name="Relationship_PrimaryCompanyContextRelationshipId">Relationship_PrimaryCompanyContextRelationshipId</a>
First included in: CashAndBankManagement/BankStatementEntity (this entity)
#### Properties
<table><tr><th>Name</th><th>Value</th></tr><tr><td>dataFormat</td><td>guid</td></tr></table>
#### Traits
<details>
<summary>List of traits for the Relationship_PrimaryCompanyContextRelationshipId attribute are listed below.</summary>
**is.dataFormat.character**
**is.dataFormat.big**
**is.dataFormat.array**
**is.dataFormat.guid**
**means.identity.entityId**
**is.linkedEntity.identifier**
Marks the attribute(s) that hold foreign key references to a linked (used as an attribute) entity. This attribute is added to the resolved entity to enumerate the referenced entities. <table><tr><th>Parameter</th><th>Value</th><th>Data type</th><th>Explanation</th></tr><tr><td>entityReferences</td><td><table><tr><th>entityReference</th><th>attributeReference</th></tr><tr><td><a href="../../../Tables/Finance/Ledger/Main/CompanyInfo.md" target="_blank">/core/operationsCommon/Tables/Finance/Ledger/Main/CompanyInfo.cdm.json/CompanyInfo</a></td><td><a href="../../../Tables/Finance/Ledger/Main/CompanyInfo.md#RecId" target="_blank">RecId</a></td></tr></table></td><td>entity</td><td>a reference to the constant entity holding the list of entity references</td></tr></table>
**is.dataFormat.guid**
**is.dataFormat.character**
**is.dataFormat.array**
</details>
| 35.625841 | 854 | 0.730941 | eng_Latn | 0.316566 |
914e98746d91e5398269f530d302a428f65ab202 | 6,046 | md | Markdown | src/pages/posts/Life/kvwpt3k0-03700.md | plutoniumm/frontier-next | 9487714b636a4f8c00feec15b9fd590cc11d1fe0 | [
"MIT"
] | null | null | null | src/pages/posts/Life/kvwpt3k0-03700.md | plutoniumm/frontier-next | 9487714b636a4f8c00feec15b9fd590cc11d1fe0 | [
"MIT"
] | 5 | 2021-10-12T09:58:43.000Z | 2022-02-27T06:43:42.000Z | src/pages/posts/Life/kvwpt3k0-03700.md | plutoniumm/frontier-next | 9487714b636a4f8c00feec15b9fd590cc11d1fe0 | [
"MIT"
] | null | null | null | ---
layout: 'layout:post'
---
# Immature Writers: Star Trek
Modern movies keep getting more and more shite because they are written and directed by children for children. Or more like people with the emotional maturity and the attention span of children.
Consider Star Trek Wrath of Khan (WoK), the undisputed champion of the best Star Trek ever with every metric besides CGI. The main plot is around a 50 year old Kirk who is given back command of the Enterprise to stop Khan who wants chaos everywhere and revenge on Kirk for leaving him stranded on a deserted planet many years ago. The movie starts with a Kirk on his 50th birthday slowly coming to terms with his age, eventually he is even gifted a pair of reading glasses by a member of the crew signifying that his eyesight isn't what is used to be. A VERY VERY RELATABLE BRILLIANT PREMISE, people feel like they are not getting the maximum of their potential and are bored and looking for adventure and that younger and younger people are coming in to replace them everyday.
The new Star Trek Into Darkness portrays him as a 36 year old in the prime of his life literally saving the universe and doing his best work ever. But since they need to project him as a reluctant hero they rip off the 50 year old Kirk's struggles and project it onto a 36 year old one. TF!?.
In the end of WoK as Spock is dying the old Kirk doesn't start screaming or react uncontrollably but instead just slowly sinks to the grown as he slowly simultaneously copes with the death of a man in action and also the loss of a friend at an old age. Even while delivering the Eulogy as expected of a man leading a starship and a crew he delivers it with stoic composure wavering once slightly at the end. He is expectly to lead by example and does so immaculately.
In the new star trek. IT IS A SHITSHOW of characters running around, screaming all the time and discussing relationships in front of superiour officers while responding to high stress situations with emotional outbursts. Even in the end the audience has little time to process what happened as they are forced to come to terms with destruction of a city, a starship, the death of khan and Kirk all within a span of less than 10 minutes. Whatever the fuck just happened? not to mention how in another 5 mins half of this shit will be instantly undone via some magical Trius Ex Machina.
### Interpersonal Conflict
At a brief moment during the collapse of the Klingon empire the federation is arguing of what to do. Spock is on one side where he wants to negociate a peace treaty & bring them into the federation and Kirk is on one side who wants to once and for all eliminate the archenemy who for years and years has terrorised the federation. No one side is shown as right or wrong, just two different perspectives are offered. They both argue their case rationally and slowly at a distance from each other across the table. Even briefly as Kirk's bigoted old nature comes to face with his anger he quickly manages to control it and face with reason the situation. Kirk is also demonstrating a difficulting in coming to terms with the scenario given his old mindset which is now slowly aging and becoming outdated. Despite this neither of the two people talk over each other or scream and disparage the other but instead while leaving unresolved, deal with it peacefully.
How did the new Star Trek deal with this? Spock and Kirk beat each other up.
### Authority
In TNG when Warf openly questions a decision taken by Data when they were left in charge in abscence of the Captain & his First Officer (FO). Data calmly calls Warf into a seperate room where they reason it out quitely in a 2 by 2 exchange. Data asked why he was questioned, Warf denfends that its his job to offer alternatives, data finally explains that after a decision is taken irrespective of his personal feelings the order is to be carried out. In the end Warf recognizes his error and asks data to calmly overlook the incident so that they can continue to be friends. HOLY SHIT THAT WAS IMPECCABLE DIALOGUE.
In Star Trek Discovery in stark contrast. The heirarchy and discipline between the people is worse than a liberal arts college. People say anything to anyone, cuss openly, scream shout and denigerate random people. No one from the Admiral to a Science officer has any sense of heirarchy or conflic resolution skill. They are able to do all this and rhetorics and sarcasm with no repurcussions whatsoever. They are all in Immature, Emotional, Generally Incompetent, Impulsive, Hyperactive, Emotionally Unstable and Unprofessional. These teenagers should not be given a small house so much the whole universe to save under their 'responsibility'.
## Hypothesis
This is all most likely happening for money. Movies are expensive to make now and if your film doesn't rake in a zillion dollars then you might as well.start learning to code. You need to appeal to as wide an audience as possible particularly the younger demographic, which means lots of action, lots of energy, lots of jokes with dialogue and fast-paced storylines that don't demand too much brain power: no time to waste people, go-go-go.
The second reason is this weird trend towards infantilizing modern audiences, carefully shielding them from anything that could be considered difficult scary or threatening: whether it's public information, videos, commercials or government announcements; everything's presented in this weird childish happy-clappy format that looks like the sort of inoffensive crap you'd show to kindergarteners.
The final and most influencial reason is writing. Any character is only as smart, capable and resourceful as their writer is. It is no surprise then that today's writers who consider mean tweets to be equivalent to mass murder aren't beacons of tough, stoic, self reliant characters. TNG was written by WW survivors, Late Star Trek post Wok was written by Cold War generation. The new high paced crap is written by Facebook and Snapchat kids. No wonder its shite. | 195.032258 | 959 | 0.803837 | eng_Latn | 0.999952 |
914ea44796c40786c133873559907899eea8b3a3 | 6,300 | md | Markdown | bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md | frederic-junier/DIU | 6e3b23d06b69c992e03552f4626676e165f772e9 | [
"CC0-1.0"
] | null | null | null | bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md | frederic-junier/DIU | 6e3b23d06b69c992e03552f4626676e165f772e9 | [
"CC0-1.0"
] | null | null | null | bloc4/BDD_Conception_Normalisation/sandbox/TP_algorithmes_jointure.md | frederic-junier/DIU | 6e3b23d06b69c992e03552f4626676e165f772e9 | [
"CC0-1.0"
] | null | null | null | DIU bloc 4 : "Bases de données : création de schémas et normalisation" : TP sur les algorithmes de jointure
====================================================
Dans ce TP, on va s'intéresser **aux algorithmes de jointures**, c'est-à-dire aux algorithmes exécutés par les moteurs des SGBDs quand ils traduisent des requêtes comme la suivante :
```sql
SELECT *
FROM table1 JOIN table2 ON table1.attr1 == table2.attr2
```
Il existe plusieurs algorithmes de jointure et l'optimiseur de requêtes du SGBD va tâcher de choisir le _meilleur_, vis-à-vis de statistiques sur les données et surtout des **index** disponibles sur les tables.
Le but du TP est ainsi de comprendre ces algorithmes fondamentaux et de les comparer entre eux puis de les comparer face à deux de SQLite 3.
**Remarque** la comparaison de performance (_benchmark_) est un exercice complexe car de nombreux paramètres très différents contribuent à la performance finale (matériel, OS, I/O disques ou d'affichage, efficacité de la compilation/interprétation du langage de programmation, caches, temps d'initialisation etc.).
Implanter les algorithmes classiques de jointure en Python
-----------------------------------------------------------
Le fichier [`join_algorithms.py`](join_algorithms.py) contient le squelette à remplir pour les trois algorithmes, à savoir _nested loop_, _hash join_ et _merge join_. Ces algorithmes font la même chose et ont la même signature `def algo(table1, attr1, table2, attr2):` :
* `table1` et `table2` sont des listes (Python) de tuples (Python). Il n'y a pas de garanties d'ordre sur ces listes;
* `attr1` (resp. `attr2`) est _l'indice_ (entier) de l'attribut de `table1` (resp. de `tablee`) sur lequel on fait la jointure;
* ces algorithmes retournent tous une liste de tuples, comme l'aurait fait la requête SQL.
Le fichier [`join_algorithms_test.py`](join_algorithms_test.py) donne un exemple d'entrées et de résultats attendus.
**EXERCICE** : compléter la fonction `join_nested_loop` et tester votre implantation avec `pytest-3` et les tests fournis.
**EXERCICE** : compléter la fonction `join_hash` et tester votre implantation avec `pytest-3` et les tests fournis.
**EXERCICE (POUR ALLER PLUS LOIN)** : compléter la fonction `join_merge` et tester votre implantations avec `pytest-3` et les tests fournis. Vous n'êtes pas obligé de faire cet exercice pour passer à la suite.
Comparer algorithmes implantés en Python
----------------------------------------
On peut maintenant comparer la performance des algorithmes avec la fonction fournie `benchmark`.Pour la fonction `join_merge` on compte séparément le temps pris pour le tris des tables.
En effet, cette étape peut-être _amortie_ car elle est utile pour d'autre opérations que la jointure, comme les clauses `ORDER BY` ou `GROUP BY`.
**EXERCICE** : comprendre ce que fait la fonction `benchmark` (vous pouvez ajouter des commentaires par exempl)e avant de l'exécuter.
Avec les paramètres par défaut de `benchmark`, on obtient les résultats suivants sur une machine portable (Dual Core Intel i7-5600U CPU @ 2.60GHz, 8GB RAM).
```
Temps pour une exécution de join_nested_loop : 47.21451419973164 ms
Temps pour une exécution de join_hash : 0.8530486993549857 ms
Temps pour une exécution des tris : 0.3733556004590355 ms
Temps pour une exécution de join_merge : 0.49316930017084815 ms
```
**EXERCICE (POUR ALLER PLUS LOIN)** : jouer avec les paramètres pour trouver un cas qui soit défavorable à `join_hash` mais favorable à `join_merge`. Sans tenir compte du temps de tri, on peut trouver des cas avec un facteur 10x en faveur de `join_merge`. _Indice_ : remarquez que les rôles de `table1` et `table2` sont asymétriques faire en sorte de passer du temps dans l'étape de construction d'index de `join_hash`.
**EXERCICE (POUR ALLER PLUS LOIN)** : même question que précédement, mais cette fois si il faut trouver un cas qui est favorable à `join_nested_loop` et dévaforable aux deux autres. _Indice_ faites en sorte que la jointure soit aussi grosse que le produit cartésien.
Comparer l'exécution dans Python à celle native dans SQLite
-------------------------------------------------------------
Maintenant, on va comparer la performance de ces implantations Python face aux algorithmes jointures de SQLite (qui est écrit en C). Pour cela on va comparer les deux approches suivantes :
* **Approche A : jointure en SQLite**, on exécute la requête `SELECT * FROM table1 JOIN table2 ON table1.val == table2.val` puis (depuis Python) on récupère l'intégralité du résultat, c'est la fonction `join_sqlite()`
* **Approche B : jointure en Python**, on exécute la requête `SELECT * FROM table1` et on stocke son résultat dans un tableau, de même pour `SELECT * FROM table2` puis on utilise un des algorithmes précedents pour faire le calcul de jointure et enfin on renvoie le résultat, c'est la fonction `join_python()`
**EXERCICE** : créer une nouvelle base de données nommée `join_algorithms_versus_sqlite3.db` et exécuter le script SQL `join_algorithms_schema.sql` pour créer le schéma *et* peupler la base avec un jeu de données similaire à celui du benchmark de l'exercice précédent.
**EXERCICE** : avec la fonction `join_algorithms_versus_sqlite3()` du programme [`join_algorithms_versus_sqlite3.py`](join_algorithms_versus_sqlite3.py) comparer les temps d'exécution des deux méthodes.
Sur ma machine, j'obtiens cet ordre de grandeur :
```
INFO:root:Temps de transfert et de jointure côté Python : 36.440872ms
INFO:root:Temps de transfert et de jointure côté Sqlite3 : 108.951388ms
```
**EXERCICE** : reprendre la comparaison mais cette fois avec la requête `SELECT COUNT(*) FROM table1 JOIN table2 ON table1.val == table2.val`. Ici, `join_python()` renverra _la longueur du tableau_ avec `len(join_hash(table1, 1, table2, 0))` pour l'algorithme de jointure par hash. Une différence _importante doit les séparer_ : comment l'expliquer ?
Sur ma machine, j'obtiens cet ordre de grandeur :
```
INFO:root:Temps de transfert et de jointure côté Python : 36.784499ms
INFO:root:Temps de transfert et de jointure côté Sqlite3 : 6.756878ms
```
**EXERCICE (FACULTATIF ET OUVERT)** : conclure en formulant quelques bonnes pratiques de l'accès à une base de données via un programme (Python).
| 73.255814 | 419 | 0.748095 | fra_Latn | 0.976679 |
914ea952efcb9768c90ca31c5a632f425c666142 | 40 | md | Markdown | README.md | issues2markdown/branding | d6f401fe407b0d7f19f59abc5dcc5fa29052bca2 | [
"Apache-2.0"
] | 1 | 2018-05-06T00:54:39.000Z | 2018-05-06T00:54:39.000Z | README.md | issues2markdown/branding | d6f401fe407b0d7f19f59abc5dcc5fa29052bca2 | [
"Apache-2.0"
] | null | null | null | README.md | issues2markdown/branding | d6f401fe407b0d7f19f59abc5dcc5fa29052bca2 | [
"Apache-2.0"
] | null | null | null | # branding
Branding for issues2markdown
| 13.333333 | 28 | 0.85 | eng_Latn | 0.989059 |
914ee5b146839b893d40df9bdad2e82c91dd6d01 | 7,930 | md | Markdown | README.md | libexus/processing-vscode | 828325ae0b922d0a881b7fa62b476778e6a06b76 | [
"MIT"
] | 176 | 2016-01-31T21:10:43.000Z | 2022-03-20T10:11:51.000Z | README.md | libexus/processing-vscode | 828325ae0b922d0a881b7fa62b476778e6a06b76 | [
"MIT"
] | 95 | 2017-01-08T11:22:45.000Z | 2022-03-01T06:06:01.000Z | README.md | libexus/processing-vscode | 828325ae0b922d0a881b7fa62b476778e6a06b76 | [
"MIT"
] | 42 | 2017-02-11T08:10:27.000Z | 2021-12-18T09:38:13.000Z | # Processing for Visual Studio Code
[](https://github.com/TobiahZ/processing-vscode/actions)
[](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde)
[](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde)
[](https://marketplace.visualstudio.com/items?itemName=Tobiah.language-pde)
## What this extension is
This is a Visual Studio Code extension created by Tobiah Zarlez to add Processing language support.
## What this extension isn't
This extension does not allow you to debug Java or Processing projects.
### Can you add a feature I want?
Possibly! [Let us know](https://github.com/TobiahZ/processing-vscode/issues), we'd love to hear your suggestions.
## Installation Instructions
1. Open [Visual Studio Code](https://code.visualstudio.com/)
1. Open the Command Pallet (`CTRL+SHIFT+P` for Windows/Linux or `CMD+SHIFT+P` on Mac) enter the command “Install Extension”
1. Search for “Processing Language” and click on this extension.
1. Restart Visual Studio Code
## Feature list
### Syntax highlighting
Open any .pde file, or choose "Processing" from the drop down menu in the bottom right corner.
### Snippets
Once the language has been set, you will see code snippets pop up automatically as you type!
### Commands
Installing this extension will add the following commands to your command pallette (`CTRL+SHIFT+P`, or opened by `View -> Command Pallette`). These commands can be selected and run from there, to complete the corresponding tasks.
## Command: Create Task File
Adds a `.vscode/tasks.json` file to your project folder, that has the contents of the `ProcessingTasks.json` located in the root folder of this project.
When you run this task (Keyboard shortcut: `Ctrl+Shift+B`), it will compile and run your project!
If you would like to see output from the compiler, comment out the line `"showOutput": "never",`
**NOTE:** Processing must be added to your path, or you must set the "processing.path" setting!
Follow [these instructions](#add-processing-to-path) to add Processing to your path, or these [alternate instructions](#alternate-method) instead to modify the path setting.
See "[Requirements](#requirements)" for full details.
## Command: Run Processing Project
This is a shortcut for running the `.vscode/tasks.json` file. Same as pressing `Ctrl+Shift+B`
**Note: Must have ran the "Create Processing Task File" command first, [see above](#command-create-task-file)!**
## Command: Open Extension Documentation
Opens this documentation.
By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting.
## Command: Open Documentation for Selection
Use the pallet command "Processing: Open Documentation for Selection" to open the processing documentation for the current selection.
By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting.
## Command: Search Processing Website
Use the pallet command "Processing: Search Processing Website" to quickly search whatever you want on the processing website.
By default uses processing.org's documentation. Can change to p5js's if preferred using the `processing.docs` setting.
By default uses Google for search. Can change to DuckDuckGo if preferred using the `processing.search` setting.
## Requirements
Installing the extension will give you instant access to [syntax highlighting](#syntax-highlighting) and [snippets](#snippets).
However, in order to compile and run your processing project from Visual Studio Code, you will need to do three things:
1. Set up your `.vscode/tasks.json` file. (See: "[Command: Create Task File](#command-create-task-file)")
1. Add Processing to your path **OR** Modify your `.vscode/tasks.json` file. (See: "[Add Processing to path](#add-processing-to-path)" or "[alternate method](#alternate-method)")
1. Have a `<File>.pde` whose filename matches the name of the project's folder (General Processing Requirement). Your file cannot contain any spaces or it will not run correctly.
## Add Processing to path
In order to automatically compile and open from Visual Studio Code, I recommend adding Processing to your path.
### What does that mean?
That means you should be able to type the `processing` from anywhere on your machine, and it will open Processing.
### How do I do that?
It's easier than you might think!
#### Windows
* Open the "Advanced System Settings" by running sysdm.cpl
* In the "System Properties" window, click on the Advanced tab.
* In the "Advanced" section, click the Environment Variables button.
* Edit the "Path" variable. Append the processing path (Example: `;C:\Program Files\Processing-3.0.1\`) to the variable value. Each entry is separated with a semicolon.
#### Mac
Open Processing, and click the `Tools` -> `Install "processing-java"` menu item.
**Note:** You will have to install processing-java for all users for this to work
#### Linux
Set your `PATH` to where your processing application is located.
Example: `export PATH=$PATH:/opt/processing/processing-2.0b4`
You also need to create an alias for `processing-java` in `/bin/` instead of `/usr/bin/`.
Example: `sudo ln -s /opt/processing/processing-java /bin/processing-java`
### Then what?
Once you've installed Processing to your path, you need to add the appropriate `.vscode/tasks.json` file to every Processing project.
See the command "[Create Task File](#command-create-task-file)"
### Alternate Method
What if you cannot, or do not want to add Processing to your path?
You can modify the `processing.path` setting to follow the path to wherever processing is installed on your machine. Be sure to remember to keep the `processing-java` at the end of the path!
To change settings in VSCode, here is a link to the [official documentation](https://code.visualstudio.com/docs/getstarted/settings).
(Remember, for Windows be sure to turn any "`\`" into "`\\`"!)
Example:
```json
"processing.path": "C:\\Program Files\\processing-3.0.1\\processing-java",
```
**NOTE:** This is untested on Mac and Linux
### If needed: Overwrite default terminal
You may need to also overwrite your default terminal in order to get your task file to run correctly.
Following [the instructions on the official VSCode documentation](https://code.visualstudio.com/docs/editor/tasks#_common-questions), all you have to do is add a few extra lines to your task file once you generate it.
For example, if you are running Windows and want the task file to use Command Prompt ('cmd.exe') you can add an 'options' parameter under the 'windows' portion:
```json
"windows": {
"options": {
"shell": {
"executable": "cmd.exe",
"args": [
"/d", "/c"
]
}
},
"args": [
"--force",
{
"value": "--sketch=${workspaceRoot}",
"quoting": "strong"
},
{
"value": "--output=${workspaceRoot}\\out",
"quoting": "strong"
},
"--run"
]
}
```
## To Do List
* Take nice looking (Animated?) screen shots for README/Instructions
## Credits
Syntax highlighting and snippets code based on the [Processing Sublime Text plugin](https://github.com/b-g/processing-sublime).
## Other resources
Here are some other resources I recommend:
* [Processing's official site](https://processing.org/)
* [Tobiah Zarlez Blog](http://www.TobiahZ.com)
| 40.050505 | 229 | 0.731526 | eng_Latn | 0.953566 |
914f5a560aa7c944bf528cbb6b77b933b3f50731 | 9,848 | md | Markdown | content/en/docs/quarks-statefulset/development/_index.md | cloudfoundry-incubator/quarks-docs | 2bc24a58e81b3a4c333e786ce79f75901dfbdbb2 | [
"Apache-2.0"
] | 1 | 2021-03-03T13:33:16.000Z | 2021-03-03T13:33:16.000Z | content/en/docs/quarks-statefulset/development/_index.md | cloudfoundry-incubator/quarks-docs | 2bc24a58e81b3a4c333e786ce79f75901dfbdbb2 | [
"Apache-2.0"
] | 4 | 2020-07-13T05:33:20.000Z | 2020-08-28T20:53:18.000Z | content/en/docs/quarks-statefulset/development/_index.md | cloudfoundry-incubator/quarks-docs | 2bc24a58e81b3a4c333e786ce79f75901dfbdbb2 | [
"Apache-2.0"
] | 1 | 2020-09-08T17:43:20.000Z | 2020-09-08T17:43:20.000Z | ---
title: "Development"
linkTitle: "Development"
weight: 4
description: >
The QuarksStatefulSet component can be understood as the set of controllers responsible for translating the BOSH manifest `instance_groups` into Kubernetes resources.
---
## Description
The QuarksStatefulSet component can be understood as the set of controllers responsible for translating the BOSH manifest `instance_groups` into Kubernetes resources.
## QuarksStatefulset Component
The **QuarksStatefulset** component is a categorization of a set of controllers, under the same group. Inside the **QuarksStatefulset** component, we have a set of 2 controllers together with one separate reconciliation loop per controller.
Figure 1 illustrates a **QuarksStatefulset** component diagram that covers the set of controllers it uses.

*Fig. 1: The QuarksStatefulset component*
### **_QuarksStatefulSet Controller_**

*Fig. 2: The QuarksStatefulset controller*
This controller will generate a Kubernetes statefulset for each `instance_group` defined in the BOSH manifest. This `Statefulset` will also include a set of Kubernetes services, so that each component can be accessed on specific ports.
#### Watches in sts controller
- `QuarksStatefulset`: Creation
- `Configmaps`: Update
- `Secrets`: Update
#### Reconciliation in sts controller
Will generate versioned `Statefulsets` with the required data to make all jobs of the `instance_group` runnable.
#### Scaling Restrictions (not implemented)
Ability to set restrictions on how scaling can occur: min, max, odd replicas.
#### Automatic Restart of Containers
When an env value or mount changes due to a `ConfigMap` or `Secret` change, containers are restarted.
The operator watches all the `ConfigMaps` and `Secrets` referenced by the `StatefulSet`, and automatically performs the update, without extra workarounds.
#### Exposing QuarksStatefulSets Publicly
Exposing `quarksstatefulsets` is similar to exposing `statefulsets` in kubernetes. A Kubernetes service makes use of labels to select the pods which should be in the service. We need to use two labels to group the pods of a single instance group.
1. `quarks.cloudfoundry.org/instance-group-name: ((instanceGroupName))`
2. `quarks.cloudfoundry.org/deployment-name: ((deploymentName))`
#### Cluster IP
Following is the example which creates a service with type **ClusterIP** for a single instance group named `nats` in deployment `nats-deployment` for exposing port 4222.
```yaml
apiVersion: v1
kind: Service
metadata:
name: nats-service
spec:
type: ClusterIP
selector:
quarks.cloudfoundry.org/instance-group-name: nats
quarks.cloudfoundry.org/deployment-name: nats-deployment
ports:
- protocol: TCP
port: 80
targetPort: 4222
```
Complete example can be found [here](https://github.com/cloudfoundry-incubator/quarks-operator/tree/master/docs/examples/bosh-deployment/boshdeployment-with-service.yaml).
Though, by default, quarks creates three services of type **ClusterIP** as defined [here](../../quarks-operator/concepts/from_bosh_to_kube/#naming-conventions) for any instance group.
#### Load Balancer
For creating a service type **LoadBalancer**, we just need to change the `.spec.type` to `LoadBalancer` in the above example. The LoadBalancer Ingress is your public IP specified in the output of this command `kubectl describe service nats-service`.
#### Ingress
**Ingress** doesn't use any labels but just sits on top of services and acts as a smart router. You can create services of different types based on the above examples and use them as values in the ingress Kubernetes spec. An example of Ingress can be found [here](https://kubernetes.io/docs/tasks/access-application-cluster/ingress-minikube/#create-an-ingress-resource)
For more information about Kubernetes services, we recommend you to read [this](https://kubernetes.io/docs/concepts/services-networking/service/).
#### Extended Upgrade Support
When an update needs to happen, a second `StatefulSet` for the new version is deployed, and both coexist until canary conditions are met.
Annotated with a version (auto-incremented on each update). The annotation key is `quarks.cloudfoundry.org/version`.
Ability to upgrade even though `StatefulSet` pods are not ready.
#### Detects if StatefulSet versions are running
During upgrades, there is more than one `StatefulSet` version for an `QuarksStatefulSet` resource. The operator lists available versions and keeps track of which are running.
A running version means that at least one pod that belongs to a `StatefulSet` is running. When a version **n** is running, any version lower than **n** is deleted.
The controller continues to reconcile until there's only one version.
#### AZ Support
The `zones` key defines the availability zones the `QuarksStatefulSet` needs to span.
The `zoneNodeLabel` defines the node label that defines a node's zone.
The default value for `zoneNodeLabel` is `failure-domain.beta.kubernetes.io/zone`.
The example below defines an `QuarksStatefulSet` that should be deployed in two availability zones, **us-central1-a** and **us-central1-b**.
```yaml
apiVersion: quarks.cloudfoundry.org/v1alpha1
kind: QuarksStatefulSet
metadata:
name: MyQuarksStatefulSet
spec:
zoneNodeLabel: "failure-domain.beta.kubernetes.io/zone"
zones: ["us-central1-a", "us-central1-b"]
...
template:
spec:
replicas: 2
...
```
The `QuarksStatefulSet` controller creates one `StatefulSet` version for each availability zone, and adds affinity information to the pods of those `StatefulSets`:
```yaml
affinity:
nodeAffinity:
requiredDuringSchedulingIgnoredDuringExecution:
nodeSelectorTerms:
- matchExpressions:
- key: "failure-domain.beta.kubernetes.io/zone"
operator: In
values: ["us-central1-a"]
```
If zones are set for an `QuarksStatefulSet`, the following occurs:
- The name of each created `StatefulSet` is generated as `<quarks statefulset name>-z<index of az>`.
```text
myquarksstatefulset-z0
```
- The `StatefulSet` and its `Pods` are labeled with the following:
```yaml
quarks.cloudfoundry.org/az-index: "0"
quarks.cloudfoundry.org/az-name: "us-central1-a"
```
- The `StatefulSet` and its `Pods` are annotated with an **ordered** JSON array of all the availability zones:
```yaml
quarks.cloudfoundry.org/zones: '["us-central1-a", "us-central1-b"]'
```
- As defined above, each pod is modified to contain affinity rules.
- Each container and init container of each pod have the following env vars set:
```shell
KUBE_AZ="zone name"
BOSH_AZ="zone name"
CF_OPERATOR_AZ="zone name"
AZ_INDEX="zone index"
```
##### Tolerations
Taints and tolerations is a concept defined in kubernetes to repel pods from nodes [link](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/). Defining tolerations is same as defined in the kubernetes docs. Keep in mind the affinity rules added by the controller when az's are defined. An example is specified in the examples folder.
#### Restarting on Config Change
`QuarksStatefulSets` can be automatically updated when the environment/mounts have changed due to a referenced
`ConfigMap` or a `Secret` being updated. This behavior is controlled by the `updateOnConfigChange` flag which defaults to `false`.
#### Watches in cleanup controller
- `StatefulSet`: Creation/Update
#### Reconciliation in cleanup controller
It will delete statefulsets with old versions, only after the new statefulset version instances are up and running.
### **_QuarksStatefulSet Active-Passive Controller_**

*Fig. 3: The QuarksStatefulset active/passive controller*
Active/passive model is application model that have multiple running instances, but only one instance is active and all other instances are passive (standby). If the active instance is down, one of the passive instances will be promoted to active immediately.
The `activePassiveProbes` key defines active probe to be performed on a container. The controller examines the active probe periodically to see if the active one is still active. If active pod is down or there isn’t an active pod, the first running pod will be promoted as active and label it as `quarks.cloudfoundry.org/pod-active: active`.
```yaml
apiVersion: quarks.cloudfoundry.org/v1alpha1
kind: QuarksStatefulSet
metadata:
name: myquarksstatefulset
spec:
activePassiveProbes:
busybox:
# define a active probe on the container
exec:
command:
- /bin/sh
- -c
- /root/check-active.sh
template:
spec:
replicas: 2
template:
spec:
containers:
- name: busybox
image: busybox
command:
- sleep
- "3600"
...
```
The controller manages this active probing and provides pod designation label to the service's selectors. Any requests sent to the service will then only be sent to the active pod.
## Relationship with the BDPL component

*Fig. 4: Relationship with the BPM controller*
Figure 4 illustrates the interaction of the **BPM** Controller with the **QuarksStatefulSet** Controller. Once the BPM controller consumes the data persisted in secrets from the `QuarksJob` Component, it will use that data to generate new `QuarksStatefulset` instances. When these resources are generated, the QuarksStatefulSet controller will be watching and trigger its reconciliation loop.
## Examples
See https://github.com/cloudfoundry-incubator/quarks-operator/tree/master/docs/examples/quarks-statefulset
| 41.033333 | 392 | 0.762998 | eng_Latn | 0.985505 |
914fc01bb8fc851df13112b02e3a565dfaa1ea95 | 3,993 | md | Markdown | event-pipeline-evm/README.md | 0xProject/0x-event-pipeline | 060e7b04bd73fe3d60d9646b7eb03ddabeb2641b | [
"Apache-2.0"
] | 11 | 2020-03-03T19:32:27.000Z | 2022-03-30T03:56:53.000Z | event-pipeline-evm/README.md | 0xProject/0x-event-pipeline | 060e7b04bd73fe3d60d9646b7eb03ddabeb2641b | [
"Apache-2.0"
] | 11 | 2019-12-03T23:57:38.000Z | 2021-09-02T03:16:53.000Z | event-pipeline-evm/README.md | 0xProject/0x-event-pipeline | 060e7b04bd73fe3d60d9646b7eb03ddabeb2641b | [
"Apache-2.0"
] | 10 | 2020-10-22T22:03:28.000Z | 2022-03-30T03:56:55.000Z | # 0x-event-pipeline
A node.js app that was originally designed for pulling 0x staking events info, but now expanded to all other 0x related events.
## Getting started
Test locally:
- Step 1
Rename the `.env.exmaple` file to `.env`, or create a new `.env` file. Add the required env variables (see below for configureation details)
- Step 2
Set up the database variables in `docker-compose.yml` file to the desired database destination.
```
$ docker-compose up # get postgres up
```
- Step 3 Test build & Debug
```
$ yarn install
$ yarn build
```
- Step 4 Build migrations
If there are new tables to be created, or schema changes, you will need to create migration files first:
```
yarn migrate:create -n <YourMigrationName>
```
Modify the migration file in `migrations/` folder with necessary changes.
Run migration:
```
$ yarn migrate:run
```
To revert migration:
```
$ yarn migrate:revert
```
- Step 5
Start the scraper:
```
$ yarn start
```
## Configuration
### Environment variables:
**Required**
`ETHEREUM_RPC_URL` - The RPC URL to use. Must match `CHAIN_ID`.
`CHAIN_ID` - The EVM chain id.
`EP_DEPLOYMENT_BLOCK` - The block on which the proxy contract was deployed
`SCHEMA` - The schema to use to store events in the DB
**Optional**
`POSTGRES_URI` - The full postgres URI to connect to. Defaults to local development.
`START_BLOCK_OFFSET` - How many blocks before the current block to search for events, allowing for updates to previously scraped events that may be in orphaned blocks.
`MAX_BLOCKS_TO_PULL` - The maximum number of blocks to pull at once.
`MAX_BLOCKS_TO_SEARCH` - The maximum number of blocks to search for events at once.
`BLOCK_FINALITY_THRESHOLD` - How many blocks before the current block to end the search, allowing you to limit your event scrape to blocks that are relatively more settled.
`MINUTES_BETWEEN_RUNS` - How long to wait between scrapes.
`SHOULD_SYNCHRONIZE` - Whether typeorm should synchronize with the database from `POSTGRES_URI`.
`STAKING_DEPLOYMENT_BLOCK` - The block on which the staking contract was deployed
`STAKING_POOLS_JSON_URL` - The source for the JSON mapping of staking pools to UUIDs (for grabbing metadata about pools). Defaults to the 0x staking pool registry GitHub repo.
`STAKING_POOLS_METADATA_JSON_URL` - The source for the JSON mapping of UUIDs to metadata. Defaults to the 0x staking pool registry GitHub repo.
`BASE_GITHUB_LOGO_URL` - The base URL for grabbing logos for staking pools. Defaults to the 0x staking pool registry GitHub repo.
## Database snapshots
When running the app on a new database it can take a long time to find new events depending on how much time has passed since the contracts were deployed. There are options to dump and restore data from other sources using `pg_dump` ([Documentation](https://www.postgresql.org/docs/9.6/app-pgdump.html)) and `pg_restore` ([Documentation](https://www.postgresql.org/docs/9.2/app-pgrestore.html)). Some examples are outlined below.
These examples will require `postgresql` to be installed.
```
$ brew install postgresql
```
### Getting data from another database
If you know of another database that contains up-to-date data, you can `pg_dump` data from the relevant schemas from that database by running:
```
$ pg_dump -h <host> -U <user> -p <port> --schema staking --schema events --data-only --file events.dump --format=c <database name>
```
To save a `pg_dump` archive file named `events.dump`. The command will prompt you for the password.
### Restoring data from a pg_dump
If you have access to a `.dump` file you can `pg_restore` data from that file into another database.
To restore data into the default development database that is spun up by `docker-compose up`, you can run:
```
$ pg_restore --data-only --dbname events --host localhost --port 5432 -U user events.dump
```
Assuming you have access to an `events.dump` file. The command will prompt you for the password.
| 31.944 | 429 | 0.749562 | eng_Latn | 0.98751 |
9150199e8c4ddf912e25e7c581aec7ab62c76ebb | 93 | md | Markdown | README.md | djoc3/react-markup | d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9 | [
"MIT"
] | null | null | null | README.md | djoc3/react-markup | d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9 | [
"MIT"
] | 5 | 2020-07-19T08:25:45.000Z | 2022-02-13T04:24:54.000Z | README.md | djoc3/react-markdown | d80d3eb4c4e916f5a5a230ccf6dea252df90f6a9 | [
"MIT"
] | null | null | null | Markdown Preview is built with React and Marked.js
Developed and maintained by Dan O'Connor
| 23.25 | 50 | 0.817204 | eng_Latn | 0.999521 |
91502d1e660651cc068a74cce475bd7f5297be53 | 981 | md | Markdown | README.md | m-pawelczyk/skrzynkomaty-domain-explorers | ef80c9e8d2f0abaab2602a35d12b430a36183ed0 | [
"MIT"
] | null | null | null | README.md | m-pawelczyk/skrzynkomaty-domain-explorers | ef80c9e8d2f0abaab2602a35d12b430a36183ed0 | [
"MIT"
] | null | null | null | README.md | m-pawelczyk/skrzynkomaty-domain-explorers | ef80c9e8d2f0abaab2602a35d12b430a36183ed0 | [
"MIT"
] | null | null | null | # SKRZYNKOMATY #DomainExplorers
Oryginalny pomysł na ten projekt wyszedł od [Mariusza Gila](https://github.com/mariuszgil/domain-explorers) i jest próbą odpowiedzenia na jego inicjatywę #DomainExplorers
Ciekawą zapowiedzią i punktem wyjścia na początek był film opublikowany na YouTube: [Rozpoznanie domeny z Big Picture EventStorming | Domain Explorers](https://youtu.be/LqIF_pGI3wk)
## O repozytorium
Ten projekt stanowi zapis analizy, modelowania oraz implementacji w ramach [Domain Explorers](https://explorers.bettersoftwaredesign.pl). Znajdziesz tutaj zarówno artefakty wynikowe poszczególnych sesji EventStormingu, modelowania, jak i (docelowo) powstałą na tej bazie implementację (miejmy nadzieję :D).
## Analiza domeny
### Big Picture EventStorming
Pierwsze rozwiązanie - Mikołaj Nowogrodzki i Michał Pawełczyk:

Board w MIRO - [LINK](https://miro.com/app/board/o9J_lU7sET0=/)
| 49.05 | 306 | 0.808359 | pol_Latn | 0.998602 |