hexsha
stringlengths
40
40
size
int64
5
1.04M
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
344
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
listlengths
1
11
max_stars_count
int64
1
368k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
344
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
listlengths
1
11
max_issues_count
int64
1
116k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
344
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
listlengths
1
11
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
1.04M
avg_line_length
float64
1.14
851k
max_line_length
int64
1
1.03M
alphanum_fraction
float64
0
1
lid
stringclasses
191 values
lid_prob
float64
0.01
1
878ae0fa7bfcf611c2182fdc16642e2be5bd0b69
5,013
mkd
Markdown
sesion7.mkd
luisfiallos9/Administracion-ssoo
7942a546c1c1fe10b1681c4313c32f0d13c4972f
[ "MIT" ]
null
null
null
sesion7.mkd
luisfiallos9/Administracion-ssoo
7942a546c1c1fe10b1681c4313c32f0d13c4972f
[ "MIT" ]
null
null
null
sesion7.mkd
luisfiallos9/Administracion-ssoo
7942a546c1c1fe10b1681c4313c32f0d13c4972f
[ "MIT" ]
null
null
null
# Sesión 7 - 2018/2019 ### Imagenes ``` $ docker pull luisf10/img-luishello:labdock $ ccarbonero/img-carbonerodocker $ ivanhn1/image-caronte:lab1 ``` ## Construir un balanceador de carga con Docker-compose y nginx como servidor web - La ruta se va a ejecutar sobre ~/lb. - Se van a crear 2 servidores web nginx y un balanceador de carga. - Por lo que para empezar es necesario crear un directorio __servers__ en el cual se establecerán los parámetros de configuración y otro directorio __load-balancer__. 1. Empezando con los servers, accedemos a el y creamos un Dockerfile. Vamos a configurar el archivo nginx.conf por lo que vamos a crear un archivo nginx.conf y en el DockerFile especificamos que queremos insertar el nuestro (Por defecto nginx trae el suyo, vamos a añadirle una linea) 2. Copiamos el archivo nginx.conf de la imagen a nuestro directorio con docker cp. * `docker cp container_id:origen destino` * `$ docker cp 881bd08c0b08:/etc/nginx/nginx.conf nginx.conf.original` 3. Añadimos en el archivo `nginx.conf` del directorio servers: - `add_header X-Backend-Server $hostname;` - Especificamos en el Dockerfile un run de nginx:latest, que copie el archivo modificado a su directorio /etc/nginx y abra el puerto 80. ``` FROM nginx:latest COPY nginx.conf /etc/nginx EXPOSE 80 ``` - Creamos directorio html y dentro de él un fichero `index.html` que especifica que página mostrar al principio. - Crear fichero `nginx.conf` y `Dockerfile` en ~/lb/load-balancer - load-balancer Dockerfile igual que el de servers - Poner esto __SOLO__ en fichero `nginx.conf` del __load balancer__ para decirle que balancee las cargas ``` upstream app_servers { server webserver_1:80; server webserver_2:80; } access_log /var/log/nginx/access.log main; server { listen 80; location / { proxy_pass http://app_servers; proxy_redirect off; proxy_set_header Host $host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header X-Forwarded-Host $server_name; } } ``` - Por último crear __docker-compose.yml__ en ~/lb - Con services le especificamos que servicios debe crear, crea primero webserver_1 dentro de la carpeta servers ejecutando el Dockerfile y abriendo el puerto 80 a la máquina local, __no a internet__ - Despues ejecuta load-balancer especificando que dependa de los dos servidores webs creados anteriormente, accediendo a la red desde el puerto 80. ``` version: '3' services: webserver_1: build: context: ./servers dockerfile: Dockerfile expose: - "80" webserver_2: build: context: ./servers dockerfile: Dockerfile expose: - "80" lb: build: context: ./load-balancer dockerfile: Dockerfile depends_on: - webserver_1 - webserver_2 ports: - "80:80" ``` - Dentro de la carpeta ~/lb - `$ docker-compose build webserver_1` Construye el servicio webserver_1 especificado en el docker-compose.yml - `$ docker-compose build webserver_2` - `$ docker-compose build lb` Construye el servicio lb especificado en el docker-compose.yml - `$ docker-compose up lb` Ejecuta el servicio __lb__, como depende de webserver_1 y webserver_2 se ejecutan tambien esos servicios. - `$ docker-compose up -d --build`Construye la imagen y la ejecuta en segundo plano - `$ docker-compose stop` Para la imagen - `$ docker-compose logs` Imprime los logs a esa página web en ejecución. ### Ejercicio - Crea un script (loadBalancer.sh) que realice 100 solicitudes a la página web creada y nos imprima por pantalla cuantas solicitudes han sido aceptadas por qué servidor. ```bash #!/bin/bash numerof51=$(echo e2c899031f51) for i in $(seq 100) do maquina=$(curl -I http://35.234.93.74/ | grep X-Backend | cut -d ':' -f 2 | tr -d ' ') echo $maquina done ``` - `$ ./loadBalancer.sh 2> /dev/null | sort | uniq -c` Realiza las peticiones, ordena las peticiones y cuenta las peticiones de cada tipo. ### Diferencia entre ports y expose (Docker) - `expose` Significa que el puerto docker esta reservado, no se puede utilizar, para que desde tu maquina puedas acceder al docker de dentro, pero desde internet no se puede acceder. - `$ docker run --expose=1234 my_app` - `port, -p` `localPort:DockerPort` Es 80:80, de la maquina anfitriona a la docker, publica el puerto en la maquina exterior para que se localice desde fuera. - `$ docker run -p 8080:80/tcp my_app` ## Extra - `$ vimdiff nginx.conf nginx.conf.original` Para ver la diferencia entre archivos. - `curl -I webPage` Para ver solo las cabeceras. - `Kubernetes` Sistema para la gestión de aplicaciones en contenedores Docker permitiendo acciones como programar el despliegue, escalado y la monitorización de nuestros contenedores. - `Heartbeat` Daemon de docker que consulta el estado de los servicios y determinan si están disponibles. Envía controles de salud, por ejemplo una query nula para saber si está levantado el servicio.
33.198675
284
0.741073
spa_Latn
0.941713
878b12dc126756d21e9d91f184688a91313fba6f
729
md
Markdown
_posts/2015-05-11-busy.md
xiaoronglv/xiaoronglv.github.io
48784a298e377b6efcfbaf4a70b507d3b81608ee
[ "Apache-2.0" ]
10
2016-08-02T23:36:06.000Z
2021-02-03T12:41:03.000Z
_posts/2015-05-11-busy.md
xiaoronglv/xiaoronglv.github.io
48784a298e377b6efcfbaf4a70b507d3b81608ee
[ "Apache-2.0" ]
4
2020-05-07T01:57:23.000Z
2021-09-27T21:00:20.000Z
_posts/2015-05-11-busy.md
xiaoronglv/xiaoronglv.github.io
48784a298e377b6efcfbaf4a70b507d3b81608ee
[ "Apache-2.0" ]
13
2015-08-13T04:26:57.000Z
2019-04-08T06:57:49.000Z
--- title: 我不在 layout: post guid: 75d9d0b31ab date: 2015-05-11 21:54:44 tags: - --- > 什么是骄傲,什么是缺乏爱心? > 这篇文章就是活生生的案例。 > updated_at 2015-05-30 22:48:09 这是微信上经常上演的一幕。 > 2015-05-11 21:55:43 > > 某某:在? > 2015-05-12 07:00:00 > > 吕小荣:恩,什么事? > 2015-05-12 12:10:00 > > 某某:问你个事? > 2015-05-12 12:10:30 > > 某某:哪家医院治疗糖尿病最好?...(缓慢的打字...) > 2015-05-12 12:10:50 > > 吕小荣:上海地区吗? > 2015-05-12 12:13:00 > > 某某:...(继续非常缓慢的打字) > 2015-05-12 12:14:50 > > 吕小荣:六院或瑞金 整个沟通耗时一天,让我十分烦躁。为什么不一次性把时间、地点、任务、事情全部交代清楚啊? > 这样表达多好啊。 > > 2015-05-11 21:55:43 > > A 君: Hi 小荣,我的叔叔 54 岁,刚刚检查出得了糖尿病,暂没有发现并发症,这是初筛的检查资料。你帮忙看一下。顺便问一下,上海哪个医院治疗糖尿病最好? > 2015-05-11 21:55:50 > > 吕小荣:六院或瑞金。尤其是六院,经常会组织糖尿病科普讲座,你可以留意一下。 以后再有人啰里啰嗦问「在不在」,直接拉黑,绝不搭理。 珍爱生命,远离 IM 工具。
11.390625
80
0.648834
yue_Hant
0.23018
878b713042746d4baa3179761e0d7e3ada26d6b3
1,169
md
Markdown
technical-reference/msbts-messageinstance-referencetype-property-wmi.md
SicongLiuSimon/biztalk-docs
85394b436d277504d9e759c655608888123785bd
[ "CC-BY-4.0", "MIT" ]
1
2020-06-16T22:06:46.000Z
2020-06-16T22:06:46.000Z
technical-reference/msbts-messageinstance-referencetype-property-wmi.md
AzureMentor/biztalk-docs
16b211f29ad233c26d5511475c7e621760908af3
[ "CC-BY-4.0", "MIT" ]
7
2020-01-09T22:34:58.000Z
2020-02-18T19:42:16.000Z
technical-reference/msbts-messageinstance-referencetype-property-wmi.md
AzureMentor/biztalk-docs
16b211f29ad233c26d5511475c7e621760908af3
[ "CC-BY-4.0", "MIT" ]
2
2017-06-23T18:30:28.000Z
2017-11-28T01:11:25.000Z
--- title: MSBTS_MessageInstance.ReferenceType Property (WMI) TOCTitle: MSBTS_MessageInstance.ReferenceType Property (WMI) ms:assetid: 7d204971-e7f8-42e4-9e9f-4d33a72fcf0d ms:mtpsurl: https://msdn.microsoft.com/library/Aa560993(v=BTS.80) ms:contentKeyID: 51529177 ms.date: 08/30/2017 mtps_version: v=BTS.80 --- # MSBTS\_MessageInstance.ReferenceType Property (WMI)   Contains information about how message is referenced by a service. *The syntax shown is language neutral.* ## Syntax ```C# uint32 ReferenceType; ``` ## Remarks This property is read-only. The following table contains the permissible values for this property: <table> <thead> <tr class="header"> <th>Description</th> <th>Value</th> </tr> </thead> <tbody> <tr class="odd"> <td>Delivered, not consumed</td> <td>1</td> </tr> <tr class="even"> <td>Consumed</td> <td>2</td> </tr> <tr class="odd"> <td>Suspended</td> <td>4</td> </tr> <tr class="even"> <td>Abandoned</td> <td>8</td> </tr> </tbody> </table> Note that the integer values must be used in code and script. ## Requirements **Header:** Declared in BTSWMISchemaXP.mof. **Namespace:** Included in \\root\\MicrosoftBizTalkServer.
17.191176
70
0.715141
eng_Latn
0.523834
878b8729f7a479e9fe8a5a2b299d6da5526233cb
4,750
markdown
Markdown
_posts/2018-10-14-what-im-listening-to.markdown
richardcadman/rich.fyi
4768b0bc7640cfbedf445d40f034e5987ae539f0
[ "MIT" ]
null
null
null
_posts/2018-10-14-what-im-listening-to.markdown
richardcadman/rich.fyi
4768b0bc7640cfbedf445d40f034e5987ae539f0
[ "MIT" ]
3
2021-05-19T18:16:54.000Z
2022-02-26T04:26:07.000Z
_posts/2018-10-14-what-im-listening-to.markdown
richardcadman/rich.fyi
4768b0bc7640cfbedf445d40f034e5987ae539f0
[ "MIT" ]
null
null
null
--- layout: post title: "What I'm listening to" date: 2018-10-14 12:09:18 +0100 categories: music --- `What I'm listening to` tl;dr, tends to be Jazz #### March 2019 Been a while: - [My March 2019 Playlist](spotify:user:kingcadders:playlist:65VQ9BIBh8ktnbas5yIyIy) - [Jordan Rakei - Minds Eye](spotify:track:3UYKvI6JYzkp8eep9fcWT0), love Jordan - [Japanese House - Good at Falling](spotify:album:7qcAgj6CoV0VeRLo6ILiF2), The 1975 protege - [Snarky Puppy - Immigrance](spotify:album:6a1HtLhd3zNccXRNUZ23ge), so so good, also nuts live vid of some old stuff [here](https://www.youtube.com/watch?v=L_XJ_s5IsQc) - [Alfa Mist - Retainer](spotify:track:7juADD4wfjO6YhL433dQLC) - [Cory Wong - The Optimist](spotify:album:0A8cq3cjiMMP8lI1SeG4GN), fresh funk - [The Cinematic Orchestra - To Believe](spotify:album:3coLBjEk3wxbYrmIKVuSon), with Moses Sumney #### November 2018 - [My Nov 2018 Playlist](spotify:user:kingcadders:playlist:1MN1Vz26NTFTZafFJ5oR1I) - [Anderson .Paak - Oxnard](spotify:album:3rqqwtJE89WoWvMyPTvbZc), incredible album - [Classical Fix Playlist](spotify:user:rl0t0ut0zu27ouetrby7doect:playlist:4N19PpjoOnom1bq6QMaKBm), nice podcast - [Honne - Love Me / Love Me Not](spotify:album:0fwZXPXf41aF6H0CN3UtXV), sounds the same as their last album - [Michael Kiwanuka - Love & Hate](spotify:album:71kreixSBb0aUsk2gx0lcA), forgot how easy this is to listen to - [Lord of the Rings - Fellowship of the Ring](spotify:album:04rz93AqGy9JduzV3K81Dh), I mean, come on - [Vulfpeck - Darwin Derby](spotify:album:1ntmUvbS2ycBWoIWgmTVtx), bit much, but new Vulfpeck #### October 2018 - [My Oct 2018 Playlist](spotify:user:kingcadders:playlist:5n0YG0jrRmjxWqYKTNGGiZ) - [Noname - Room 25](spotify:album:7oHM3Sj0l2nXAzGAxW0KOt), still listening - [Loyle Carner - Ottolenghi](spotify:track:0KTvHwi5cDdZh86cpZztfu), great cookbook, great rapper + Jordan Rakei naughtiness - [Jazz Classics Blue Note Playlist](spotify:user:spotify:playlist:37i9dQZF1DWTR4ZOXTfd9K), smooth as hell - [Alex Somers - Black Mirror Hang the DJ Soundtrack](spotify:album:2l1cg4CKpr6YzWnb9nESCB), great TV show, great anthemic ending - [Alex Somers - Captain Fantastic Soundtrack](spotify:album:5g2NoUFQCXUfaB8xWJEQI3), great film, great playlist for a rain day - [All of Alfa Mist's stuff](spotify:artist:2i1CPudyCUjL50Wqjv8AMI), particularly [Antiphon](spotify:album:3yvubzjqmhnZhVwp6qDXPq), [7th October Epilogue](spotify:album:7yIqcIfBI2iMDNTN9RAJZU), [Blacked Out](spotify:album:5KWeMYzDSclzFCDeOd1v7j) - [All of Barney Artist's stuff](spotify:artist:5iRM7qYip6UNfQaPe2reCz), particularly his [debut album](spotify:album:1GpJeoYxcez8hKdk6tforL) and [this with Tom Misch and Loyle Carner](spotify:album:5tbvov0fii1eMXkTYZmVdp) - New [1975](spotify:track:0D4yVl9Pn45xW2s63MFCmT) single #### September 2018 - [My Sep 2018 playlist](spotify:user:kingcadders:playlist:6S3mw9fXtpdwdMZZYdyuM8) - [Oriental/Middle Eastern Jazz Mix](spotify:user:joeshinataan:playlist:4YSSozC4JrfuHLbYJpiLxF), had a middle eastern dinner party - [Jordan Rakei - Songs I love](spotify:user:jordanrakei:playlist:0D9v1n9pIJlDFJOAuZpivT), chunky playlist from Jordan - [Boogie Nights Playlist](spotify:user:spotify:playlist:37i9dQZF1DXaJN8jVBUEiZ), funk/soul/boogie/disco vibes - [Jazz UK Playlist](spotify:user:spotify:playlist:37i9dQZF1DXbHcQpOiXk1D), best modern jazz playlist - [Noname - Room 25](spotify:album:7oHM3Sj0l2nXAzGAxW0KOt), back with some [cosmic jazz / neo-soul](https://pitchfork.com/reviews/albums/noname-room-25/) - [Robert Glasper - Black Radio](spotify:album:1yqUCdbw73DpnHBVDwNa3X), incredible album, especially _Afro Blue_ and _Ah Yeah_ - [Parcels - Overnight](spotify:track:1kgws2l8gsvDhtsVyzWbu9), basically a Nile Rodgers tribute act #### August 2018 - [My Aug 2018 Playlist](spotify:user:kingcadders:playlist:7B3GNreoTkQxw7g8eJY6R1), big month for Snarky Puppy - [Wu Funk Playlist](spotify:user:soulfood17:playlist:5GvYvebTBhy4lE7oaF1AyH), the man known as Yussef Kamal (half of), Kamal Williams, Henry Wu - [Liminial Playlist](spotify:user:thesigurros:playlist:5GQrd8uzIV6q3ck8UqJmPk), an 'endless playlist' curated by Alex Somers, Jonsi, etc #### July 2018 - [My Jul 2018 Playlist](spotify:user:kingcadders:playlist:3XY64RHWnF6uxTaipAE2tc), chirpy Aaron Taylor, Prince, 1975, James Vincent, Blue Lab Beats, etc - [Summer Playlist](spotify:user:kingcadders:playlist:53MzLLqitHNhxiGSmr9C0k), don't judge me, chirpy summer playlist for the car - [Low Key Playlist](spotify:user:spotify:playlist:37i9dQZF1DX2yvmlOdMYzV), according to spotify "high-key bangers for low-key enjoyment" - [Blue Lab Beats - Xover](spotify:album:08OWYxVuGW8F7Ne2TdNMUp) - [Chance the Rapper - I might need security](spotify:track:3EApebexZ7YqDIqw2EMTDh), one of a few new Chance singles
69.852941
245
0.794526
eng_Latn
0.291157
878bc9e3ca0ac11c4eaebc93a2114a18dd6dab62
506
md
Markdown
en/_includes/create-folder.md
teminalk0/docs
2067fdc72e78b3a9ff9987723a56a2a1b4eea41d
[ "CC-BY-4.0" ]
117
2018-12-29T10:20:17.000Z
2022-03-30T12:30:13.000Z
en/_includes/create-folder.md
teminalk0/docs
2067fdc72e78b3a9ff9987723a56a2a1b4eea41d
[ "CC-BY-4.0" ]
205
2018-12-29T14:58:45.000Z
2022-03-30T21:47:12.000Z
en/_includes/create-folder.md
teminalk0/docs
2067fdc72e78b3a9ff9987723a56a2a1b4eea41d
[ "CC-BY-4.0" ]
393
2018-12-26T16:53:47.000Z
2022-03-31T17:33:48.000Z
1. Click **Create folder** in the [Home page]({{ link-console-main }}) of the management console. 2. Enter the folder name. {% include [name-format](name-format.md) %} 1. Select **Create a default network**. A [network](../vpc/concepts/network.md#network) is created with subnets in each availability zone. A [default security group](../vpc/concepts/security-groups.md#default-security-group) will also be created in this network, inside which all network traffic is allowed. 1. Click **Create**.
46
307
0.727273
eng_Latn
0.991413
878d584aa3ec1d1fb71f8804411117b70de829c3
475
md
Markdown
README.md
Vladimir-Patrushin/ptq
fe89cd28e623f818c848e3972af2a74633778c7c
[ "MIT" ]
null
null
null
README.md
Vladimir-Patrushin/ptq
fe89cd28e623f818c848e3972af2a74633778c7c
[ "MIT" ]
null
null
null
README.md
Vladimir-Patrushin/ptq
fe89cd28e623f818c848e3972af2a74633778c7c
[ "MIT" ]
null
null
null
This is the Point Tracker Quiz. Its just a normal quiz. it reveals your points at the end, normal, right? This program uses python3 with terminal. put the 'pqt.py' in a directory you like, cd into that directory, type 'python3 pqt.py' and you're done! <br>That simple! the program has run. it has run in the terminal, but thats because its suposed to be a simple quiz game played with terminal. it was a test at first, but i like sharing my ideas for inspiration. good luck!
118.75
222
0.766316
eng_Latn
0.999894
878dd1b44949fa68380c8bcb0ac85e75ce0a7473
119
md
Markdown
README.md
hoangq1211/react-quiz
db1b5bd7238040bc22657581bb4f8e3e0d66e4eb
[ "MIT" ]
null
null
null
README.md
hoangq1211/react-quiz
db1b5bd7238040bc22657581bb4f8e3e0d66e4eb
[ "MIT" ]
null
null
null
README.md
hoangq1211/react-quiz
db1b5bd7238040bc22657581bb4f8e3e0d66e4eb
[ "MIT" ]
null
null
null
## Installation - Install dependencies ``` $ npm install / yarn install ``` - Run it ``` $ npm start / yarn start ```
10.818182
28
0.621849
eng_Latn
0.545156
87907e0e9555f60f2bac3044a0bf059b216931ba
51,219
md
Markdown
articles/app-service-mobile/app-service-mobile-node-backend-how-to-use-server-sdk.md
OpenLocalizationTestOrg/azure-docs-pr15_de-AT
ca82887d8067662697adba993b87860bdbefea29
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
1
2020-11-29T22:55:06.000Z
2020-11-29T22:55:06.000Z
articles/app-service-mobile/app-service-mobile-node-backend-how-to-use-server-sdk.md
Allyn69/azure-docs-pr15_de-CH
211ef2a7547f43e3b90b3c4e2cb49e88d7fe139f
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
null
null
null
articles/app-service-mobile/app-service-mobile-node-backend-how-to-use-server-sdk.md
Allyn69/azure-docs-pr15_de-CH
211ef2a7547f43e3b90b3c4e2cb49e88d7fe139f
[ "CC-BY-3.0", "CC-BY-4.0", "MIT" ]
2
2019-07-03T20:05:49.000Z
2020-11-29T22:55:15.000Z
<properties pageTitle="Arbeiten mit Node.js Backend-Server SDK für Mobile Apps | Azure App Service" description="Informationen Sie zum Arbeiten mit Node.js Backend-Server SDK für Azure App Service Mobile Apps." services="app-service\mobile" documentationCenter="" authors="adrianhall" manager="erikre" editor=""/> <tags ms.service="app-service-mobile" ms.workload="mobile" ms.tgt_pltfrm="mobile-multiple" ms.devlang="node" ms.topic="article" ms.date="10/01/2016" ms.author="adrianha"/> # <a name="how-to-use-the-azure-mobile-apps-nodejs-sdk"></a>Verwendung des Azure Mobile Apps Node.js-SDK [AZURE.INCLUDE [app-service-mobile-selector-server-sdk](../../includes/app-service-mobile-selector-server-sdk.md)] Dieser Artikel enthält detaillierte Informationen und Beispiele zum Backend Node.js in Azure App Service Mobile Apps arbeiten. ## <a name="Introduction"></a>Einführung Azure App Service Mobile Apps ermöglicht einen Mobil optimierte Datenzugriff Web API eine Anwendung hinzufügen. ASP.NET und Node.js ASP.NET-Webanwendungen Azure App Service Mobile Apps SDK vorgesehen. Das SDK enthält die folgenden Vorgänge: - Tabellenvorgänge (Lesen, Insert, Update, Delete) für den Datenzugriff - Benutzerdefinierte API-Operationen Beide Vorgänge bieten für die Authentifizierung über alle Identitätsanbieter Azure App Service, einschließlich der sozialen Identitätsanbieter wie Facebook, Twitter, Google und Microsoft Azure Active Directory für Enterprise Identity zulässig. Sie finden Beispiele für jeden Anwendungsfall im [Beispielverzeichnis GitHub]. ## <a name="supported-platforms"></a>Unterstützte Plattformen Azure Mobile Apps Knoten SDK unterstützt die aktuelle LTS Version von Knoten und höher. Schreibe ist die neueste LTS Knoten v4.5.0. Andere Versionen von Knoten funktioniert jedoch nicht unterstützt. Azure Mobile Apps Knoten SDK unterstützt zwei Datenbank - Knoten Mssql-Treiber unterstützt SQL Azure und lokalen SQL Server-Instanzen. Sqlite3-Treiber unterstützt SQLite-Datenbanken in einer einzigen Instanz. ### <a name="howto-cmdline-basicapp"></a>Gewusst wie: verwenden grundlegende Node.js Backend erstellen Jede Azure App Service Mobile App Node.js Backend startet eine ExpressJS Anwendung. ExpressJS ist der am häufigsten verwendeten Webdienst-Framework für Node.js. Erstellen Sie ein [Express] -Anwendung wie folgt: 1. Erstellen Sie ein Verzeichnis für das Projekt in einem Befehl oder PowerShell-Fenster. mkdir basicapp 2. Führen Sie Npm Init um die Paketstruktur zu initialisieren. cd basicapp npm init Befehl Init Npm fordert eine Reihe von Fragen, um das Projekt nicht initialisieren. Finden Sie in der Ausgabe: ![Die Init-Ausgabe npm][0] 3. Installieren Sie express und Azure-Mobile-apps Bibliotheken von Npm Repository. npm install --save express azure-mobile-apps 4. Erstellen Sie eine Datei app.js um einfache mobile Server implementieren. var express = require('express'), azureMobileApps = require('azure-mobile-apps'); var app = express(), mobile = azureMobileApps(); // Define a TodoItem table mobile.tables.add('TodoItem'); // Add the mobile API so it is accessible as a Web API app.use(mobile); // Start listening on HTTP app.listen(process.env.PORT || 3000); Diese Anwendung erstellt Mobil optimierte WebAPI mit einem einzigen Endpunkt (`/tables/TodoItem`), nicht authentifizierten Zugriff auf einem zugrunde liegenden SQL-Datenspeicher mit einem dynamischen Schema bereitstellt. Es ist für folgende Client Bibliothek schnell starten: - [Schnelleinstieg Android-Client] - [Apache Cordova Client Schnellstart] - [iOS-Client Schnellstart] - [Windows Store Client Schnellstart] - [Xamarin.iOS Client-Schnellstart] - [Xamarin.Android Client-Schnellstart] - [Xamarin.Forms Client-Schnellstart] Den Code finden Sie für diese einfache Anwendung in die [Basicapp Probe auf GitHub]. ### <a name="howto-vs2015-basicapp"></a>Gewusst wie: Erstellen Sie Knoten Backend mit Visual Studio 2015 Visual Studio 2015 erfordert eine Erweiterung in der IDE Node.js Anwendungsentwicklung. Installieren Sie dazu [Node.js Tools 1.1 für Visual Studio]. Nach der Installation Node.js-Tools für Visual Studio erstellen Sie eine Express 4.x Anwendung: 1. Das Dialogfeld **Neues Projekt** öffnen ( **Datei** > **neu** > **Projekt...**). 2. **Vorlagen** > **JavaScript** > **Node.js**. 3. Wählen Sie die **grundlegenden Azure Node.js Express 4 Anwendung**. 4. Geben Sie den Namen des Projekts. Klicken Sie auf *OK*. ![Neues Projekt von Visual Studio 2015][1] 5. Den Knoten **Npm** Maustaste und wählen Sie **neue installieren... Npm Pakete**. 6. Sie müssen Npm-Katalog zum Erstellen Ihrer ersten Node.js-Anwendung aktualisieren. Klicken Sie ggf. auf **Aktualisieren** . 7. _Azure-Mobile-apps_ in das Suchfeld eingeben. **Azure-Mobile-apps 2.0.0** -Paket, klicken Sie auf **Paket installieren**. ![Neue Npm-Pakete installieren][2] 8. Klicken Sie auf **Schließen**. 9. Öffnen der _app.js_ Azure Mobile Apps SDK unterstützen. In Zeile 6 zu der Bibliothek Anweisungen erforderlich, fügen Sie den folgenden Code hinzu: var bodyParser = require('body-parser'); var azureMobileApps = require('azure-mobile-apps'); Fügen Sie an rund 27 nach anderen app.use Anweisungen den folgenden Code: app.use('/users', users); // Azure Mobile Apps Initialization var mobile = azureMobileApps(); mobile.tables.add('TodoItem'); app.use(mobile); Speichern Sie die Datei. 10. Führen Sie die Anwendung lokal (http://localhost: 3000 ist die API bereitgestellt) oder Azure veröffentlichen. ### <a name="create-node-backend-portal"></a>Gewusst wie: Node.js Backend mit Azure-Portal erstellen Sie können Recht Backend-Mobile-Anwendung in [Azure-Portal]erstellen. Führen Sie die folgenden Schritte oder Client und Server gemeinsam nach dem [Erstellen einer Apps](app-service-mobile-ios-get-started.md) Tutorial erstellen. Das Lernprogramm enthält eine vereinfachte Version der Bedienungsanleitung und am besten Nachweis Konzept Projekte. [AZURE.INCLUDE [app-service-mobile-dotnet-backend-create-new-service-classic](../../includes/app-service-mobile-dotnet-backend-create-new-service-classic.md)] Wählen Sie im Blatt _Erste Schritte_ unter **Erstellen eines API** **Node.js** als **Back-End-Sprache**. Aktivieren Sie das Kontrollkästchen für "**ich bestätige, dass dadurch alle Website Inhalt. überschrieben**" und dann auf **Erstellen TodoItem-Tabelle**. ### <a name="download-quickstart"></a>Gewusst wie: Downloaden Node.js Backend-Schnellstart Codeprojekt mit Git Beim Erstellen Backend Node.js Mobile-Anwendung über das Portal **Schnellstart** Blade Node.js-Projekt erstellt und auf Ihrer Website bereitgestellt. Sie können Tabellen und APIs hinzufügen und Codedateien für das Backend Node.js im Portal bearbeiten. Verschiedene Bereitstellungstools können Sie das Back-End-Projekt herunterladen, hinzufügen oder Ändern von Tabellen und APIs und veröffentlichen Sie das Projekt erneut. Weitere Informationen finden Sie im [Bereitstellungshandbuch für Azure App Service]. Die folgende Prozedur verwendet ein Git Repository Quickstart Projektcode herunterladen. 1. Installieren Sie Git, falls Sie dies nicht bereits getan haben. Die Schritte zum Installieren der Git unterschiedlich Betriebssysteme. Siehe [Installation von Git](http://git-scm.com/book/en/Getting-Started-Installing-Git) für betriebssystemspezifische Verteilung und Installation. 2. Folgen Sie den Schritten Git Repository für die Back-End-Website eine Notiz Bereitstellung Benutzername und Kennwort aktivieren [App Service app Repository aktivieren](../app-service-web/app-service-deploy-local-git.md#Step3) . 3. Blatt für die Mobile Anwendung Backend Notieren der **Git Clone URL** -Einstellung. 4. Ausführen der `git clone` Befehl Git clone Passworteingabe bei Bedarf wie im folgenden Beispiel-URL: $ git clone https://username@todolist.scm.azurewebsites.net:443/todolist.git 5. Wechseln Sie zum lokalen Verzeichnis, das im vorherigen Beispiel /todolist ist und beachten Sie, dass Dateien heruntergeladen wurden. Suchen Sie die `todoitem.json` Datei in der `/tables` Verzeichnis. Diese Datei definiert die Berechtigungen für die Tabelle. Auch die `todoitem.js` -Datei im gleichen Verzeichnis definiert, CRUD-Vorgang Skripts für die Tabelle. 6. Nachdem Sie Projektdateien geändert haben, führen Sie die folgenden Befehle hinzufügen, übernehmen und dann die Änderungen auf die Website hochladen: $ git commit -m "updated the table script" $ git push origin master Wenn Sie das Projekt neue Dateien hinzufügen, müssen Sie zunächst führen die `git add .` Befehl. Die Website wird veröffentlicht, jedes Mal ein neuer Satz von Commits auf der Website abgelegt ist. ### <a name="howto-publish-to-azure"></a>Gewusst wie: Backend Node.js in Azure veröffentlichen Microsoft Azure bietet viele Mechanismen für Azure App Service Mobile Apps Node.js Backend Azure Service veröffentlichen. Dazu gehören mit Bereitstellungstools in Visual Studio integriert, Befehlszeilenprogramme und kontinuierliche Bereitstellungsoptionen auf Datenquellen-Steuerelement. Weitere Informationen zu diesem Thema finden Sie im [Bereitstellungshandbuch für Azure App Service]. Azure App Service hat Tipps für Node.js-Anwendung, die Sie vor der Bereitstellung überprüfen sollten: - [Knoten-Version] angeben - Wie Sie [Knoten Module] ### <a name="howto-enable-homepage"></a>Gewusst wie: Aktivieren einer Homepage für Ihre Anwendung Viele sind eine Kombination von Web- und mobile apps und ExpressJS Framework ermöglicht zwei Facetten kombinieren. In einigen Fällen können Sie jedoch nur eine mobile Schnittstelle implementieren. Es ist vorgesehen, dass Zielseite um die app Service ausgeführt wird. Geben Ihre eigene Homepage oder eine temporäre Homepage aktivieren. Um eine temporäre Homepage zu aktivieren, verwenden Sie folgende Azure Mobile Apps instanziieren: var mobile = azureMobileApps({ homePage: true }); Wollen Sie nur diese Option verfügbar bei lokal, können Sie diese Einstellung, um Ihre `azureMobile.js` Datei. ## <a name="TableOperations"></a>Tabellenvorgänge Azure-Mobile-apps Node.js Server SDK bietet Mechanismen zum Verfügbarmachen von Datentabellen in Azure SQL-Datenbank als eine WebAPI gespeichert. Fünf dienen. | Vorgang | Beschreibung | | --------- | ----------- | | /Tables/_Tablename_ abrufen | Alle Datensätze in der Tabelle abrufen | | /Tables/_Tabellenname_/:id abrufen | Abrufen eines bestimmten Datensatzes in der Tabelle | | /Tables/_Tablename_ buchen | Erstellen Sie einen Datensatz in der Tabelle | | PATCH /tables/_Tabellenname_/:id | Ein Datensatz in der Tabelle aktualisieren | | /Tables/_Tabellenname_/:id löschen | Löschen eines Datensatzes in der Tabelle | Diese WebAPI [OData] unterstützt und erweitert das Schema zur Unterstützung von [offline-Daten synchronisieren]. ### <a name="howto-dynamicschema"></a>Gewusst wie: Definieren von Tabellen mit einem dynamischen Schema Bevor eine Tabelle verwendet werden kann, muss er definiert. Tabellen können definiert werden, mit einem statischen Schema (wo Entwickler die Spalten im Schema definiert) oder dynamisch (wo das SDK steuert das Schema basierend auf Anfragen). Darüber hinaus kann Entwickler bestimmte Aspekte der WebAPI steuern die Definition Javascript-Code hinzufügen. Als bewährte Methode sollte jede Tabelle in einer Javascript-Datei im Verzeichnis Tabellen definieren Sie mithilfe der tables.import()-Methode die Tabellen importieren. Basic-Anwendung erweitern, würde die Datei app.js angepasst werden: var express = require('express'), azureMobileApps = require('azure-mobile-apps'); var app = express(), mobile = azureMobileApps(); // Define the database schema that is exposed mobile.tables.import('./tables'); // Provide initialization of any tables that are statically defined mobile.tables.initialize().then(function () { // Add the mobile API so it is accessible as a Web API app.use(mobile); // Start listening on HTTP app.listen(process.env.PORT || 3000); }); Definieren die Tabelle. / tables/TodoItem.js: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Additional configuration for the table goes here module.exports = table; Tabellen mit dynamischen Schema Standardeinstellung. Zum Deaktivieren des dynamischen Schema Global App Einstellung **MS_DynamicSchema** auf False festgelegt in Azure-Portal. Ein vollständiges Beispiel finden Sie im [Beispiel Todo auf GitHub]. ### <a name="howto-staticschema"></a>Gewusst wie: Definieren von Tabellen mit einem statischen Schema Sie können Spalten über die WebAPI verfügbar machen explizit definieren. Azure-Mobile-apps Node.js SDK fügt automatisch alle zusätzlichen Spalten erforderlich für offline-Daten synchronisieren der Liste, die Sie bereitstellen. Schnellstart-Clientanwendungen erfordern beispielsweise eine Tabelle mit zwei Spalten: Text (String) und (boolesch). Die Tabelle kann in der Tabelle JavaScript Definitionsdatei (befindet sich im Verzeichnis Tabellen) wie folgt definiert: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Define the columns within the table table.columns = { "text": "string", "complete": "boolean" }; // Turn off dynamic schema table.dynamicSchema = false; module.exports = table; Wenn Sie Tabellen statisch definieren, müssen Sie auch die tables.initialize()-Methode, um das Datenbankschema beim Start erstellen aufrufen. Die tables.initialize()-Methode gibt ein [Versprechen] , damit der Webdienst keine Anfragen vor der Datenbank initialisiert bedienen. ### <a name="howto-sqlexpress-setup"></a>Gewusst wie: SQL Express als Datenspeicher Entwicklung auf dem lokalen Computer verwenden Azure Mobile Apps die AzureMobile Apps Knoten SDK bietet drei Optionen für die Bereitstellung von Daten aus dem Feld: SDK bietet drei Optionen für die Bereitstellung von Daten aus dem Feld: - Verwenden Sie den Treiber **Speicher** wird nicht beständige Speicher bereitstellen - Verwenden Sie den **Mssql** -Treiber zu einem Datenspeicher SQL Express für die Entwicklung - Verwenden Sie **Mssql** -Treiber zu einem Datenspeicher Azure SQL-Datenbank für die Produktion Azure Mobile Apps Node.js SDK verwendet [Mssql Node.js Paket] und eine Verbindung zur SQL Express und SQL-Datenbank verwenden. Dieses Paket erfordert, dass der SQL Express-Instanz TCP-Verbindungen aktivieren. > [AZURE.TIP]Memory-Treiber bietet keine umfassende Funktionen zum Testen. Die Back-End-lokal testen möchten, sollten die Verwendung von einem Datenspeicher SQL Express und der Mssql-Treiber. 1. Downloaden Sie und installieren Sie [Microsoft SQL Server 2014 Express]. Stellen Sie sicher, dass SQL Server 2014 Express Edition Tools installieren. Wenn Sie explizit 64-Bit-Unterstützung benötigen, verbraucht 32-Bit-Version weniger Arbeitsspeicher beim Ausführen. 2. Führen Sie den Konfigurations-Manager von SQL Server 2014. 1. Erweitern Sie in der linken Struktur Menü **SQL Server-Netzwerkkonfiguration** . 2. Klicken Sie auf **Protokolle für SQLEXPRESS**. 3. Maustaste auf **TCP/IP** und **Aktivieren**. Klicken Sie im Popupfenster auf **OK** . 4. **TCP/IP** Maustaste, und wählen Sie **Eigenschaften**. 5. Klicken Sie auf die Registerkarte **IP-Adressen** . 6. Suchen Sie den Knoten **IPAll** . Geben Sie im Feld **TCP-Port** **1433**. ![Configure SQL Express for TCP/IP][3] 7. Klicken Sie auf **OK**. Klicken Sie im Popupfenster auf **OK** . 8. Klicken Sie im linken Struktur auf **SQL Server-Dienste** . 9. **SQL Server (SQLEXPRESS)** Maustaste und wählen Sie **neu starten** 10. Schließen Sie SQL Server 2014 Configuration Manager. 3. Führen Sie SQL Server 2014 Management Studio und Ihrer lokalen SQL Express-Instanz herstellen 1. Maustaste auf die Instanz im Objekt-Explorer und wählen Sie **Eigenschaften** 2. Wählen Sie die Seite **Sicherheit** . 3. Stellen Sie sicher, dass **SQL Server und Windows-Authentifizierungsmodus** ausgewählt ist 4. Klicken Sie auf **OK** ![Konfigurieren der SQL Express-Authentifizierung][4] 5. Erweitern Sie **Sicherheit** > **Benutzernamen** im Objekt-Explorer 6. Klicken Sie auf **Anmeldung** , und wählen Sie **Neue Anmeldung...** 7. Geben Sie einen Benutzernamen ein. Wählen Sie **SQL Server-Authentifizierung**. Geben Sie ein Kennwort und Kennwort **bestätigen**Geben Sie dasselbe Kennwort ein. Das Kennwort muss Windows Komplexität erfüllen. 8. Klicken Sie auf **OK** ![Hinzufügen eines neuen Benutzers zu SQL Express][5] 9. Ihr neuen Benutzername Maustaste, und wählen Sie **Eigenschaften** 10. Die Seite **Serverrollen** auswählen 11. Aktivieren Sie das Kontrollkästchen neben der Serverrolle **dbcreator** 12. Klicken Sie auf **OK** 13. Schließen Sie SQL Server 2015 Management Studio Sicherstellen Sie, dass Sie Benutzername und Kennwort gewählte aufzeichnen. Sie müssen zusätzliche Server-Rollen oder Berechtigungen je nach Bedarf bestimmte Datenbank zuweisen. Node.js-Anwendung liest die Umgebungsvariable **SQLCONNSTR_MS_TableConnectionString** für die Verbindungszeichenfolge für die Datenbank. Sie können diese Variable in Ihrer Umgebung festlegen. PowerShell können Sie diese Umgebungsvariable festlegen: $env:SQLCONNSTR_MS_TableConnectionString = "Server=127.0.0.1; Database=mytestdatabase; User Id=azuremobile; Password=T3stPa55word;" Zugriff auf die Datenbank über eine TCP-Verbindung und einen Benutzernamen und ein Kennwort für die Verbindung. ### <a name="howto-config-localdev"></a>Gewusst wie: Konfigurieren des Projekts für lokale Entwicklung Azure Mobile Apps liest eine JavaScript-Datei mit _azureMobile.js_ aus dem lokalen Dateisystem. Verwenden Sie diese Datei nicht konfigurieren Azure Mobile Apps SDK in Produktion-Appeinstellungen in [Azure-Portal] verwenden. Die _azureMobile.js_ -Datei sollte ein Konfigurationsobjekt exportieren. Die gängigsten sind: - Database Settings - Diagnoseprotokolle Settings - Alternative CORS Settings Eine Implementierung der vorherigen datenbankeinstellungen _azureMobile.js_ Beispieldatei folgt: module.exports = { cors: { origins: [ 'localhost' ] }, data: { provider: 'mssql', server: '127.0.0.1', database: 'mytestdatabase', user: 'azuremobile', password: 'T3stPa55word' }, logging: { level: 'verbose' } }; Es wird empfohlen, die _.gitignore_ Datei _azureMobile.js_ hinzufügen (oder anderen Quellcode Datei ignorieren) zu verhindern, dass Kennwörter in der Cloud gespeichert werden. Immer konfigurieren Sie Produktion im Anwendung in [Azure-Portal]. ### <a name="howto-appsettings"></a>Vorgehensweise: Konfigurieren App für die Mobile Anwendung Die meisten Einstellungen in der Datei _azureMobile.js_ haben eine entsprechende Einstellung der Anwendung in [Azure-Portal]. Anhand der folgenden Liste Ihre app im App konfigurieren: | App-Einstellung | _azureMobile.js_ festlegen | Beschreibung | Gültige Werte | | :-------------------------- | :------------------------ | :---------------------------------------- | :------------------------------------------ | | **MS_MobileAppName** | Name | Der Name der Anwendung | Zeichenfolge | | **MS_MobileLoggingLevel** | Logging.Level | Minimale Protokollebene Nachrichten protokollieren | Fehler, Warnung, Informationen, ausführliche Debug, dumm | | **MS_DebugMode** | Debuggen | Aktivieren oder deaktivieren Debugmodus | True, false | | **MS_TableSchema** | Data.Schema | Standardname für SQL-Tabellen-schema | Zeichenfolge (Standard: Dbo) | | **MS_DynamicSchema** | data.dynamicSchema | Aktivieren oder deaktivieren Debugmodus | True, false | | **MS_DisableVersionHeader** | Version (auf undefined gesetzt)| Deaktiviert den Header X-ZUMO-Server-Version | True, false | | **MS_SkipVersionCheck** | skipversioncheck | Deaktiviert das Kontrollkästchen Client-API-version | True, false | So richten Sie eine App-Einstellung 1. Auf der [Azure-Portal]anmelden. 2. **Alle Ressourcen** oder **Anwendungsdienste** klicken Sie auf den Namen der Mobile-Anwendung. 3. Einstellungen-Blades wird standardmäßig geöffnet. **Wenn nicht, klicken Sie auf.** 4. Klicken Sie im Menü Allgemein auf **ApplicationSettings** . 5. Gehen Sie zum Abschnitt App-Einstellungen. 6. Wenn Ihre Anwendung Einstellung bereits vorhanden ist, klicken Sie auf den Wert der app-Einstellung den Wert bearbeiten. 7. Wenn Ihre app-Einstellung nicht vorhanden ist, geben Sie App-Einstellung in den Schlüssel und den Wert in das Feld Wert ein. 8. Sobald Sie abgeschlossen sind, klicken Sie auf **Speichern**. Die meisten app Einstellungen erfordert einen Neustart der Dienste. ### <a name="howto-use-sqlazure"></a>Gewusst wie: SQL-Datenbank als Datenspeicher Produktion verwenden <!--- ALTERNATE INCLUDE - we can't use ../includes/app-service-mobile-dotnet-backend-create-new-service.md - slightly different semantics --> Verwendung von Azure SQL-Datenbank als Datenspeicher ist für alle Arten von Azure App Service-Anwendung identisch. Wenn dies noch nicht getan haben, folgendermaßen Sie vor, um Mobile App-Back-End erstellen. 1. Auf der [Azure-Portal]anmelden. 2. Oben links im Fenster, klicken Sie auf **+ neu** > **Web + Mobile** > **Mobile-Anwendung**, geben Sie einen Namen für Ihre Mobile App-Backend. 3. Geben Sie im Feld **Ressourcengruppe** den gleichen Namen wie Ihre app. 4. App-Standarddienst Plan ausgewählt ist. Möchten Sie Ihren App Service-Plan ändern, Sie können dazu die App Service-Plan auf > **+ neu erstellen**. Name der neuen App Service-Plan und wählen Sie eine geeignete Stelle. Klicken Sie auf die Preisstufe und wählen Sie einen entsprechenden Tarif für den Dienst. Wählen Sie **Ansicht alle** Weitere Preisoptionen **frei** und **Shared**anzeigen. Nach Auswahl den Tarif klicken Sie **Wählen** . Klicken Sie in **App Service-Plan** -Blade auf **OK**. 5. Klicken Sie auf **Erstellen**. Eine Mobile Anwendung Back-End-Bereitstellung kann einige Minuten dauern. Nach der Bereitstellung von Mobile-Anwendung Back-End-Öffnet das Portal Blatt **Einstellungen** für Mobile App-Backend. Erstellte Mobile App-Back-End können Sie Ihre Mobile App-Back-End eine vorhandene SQL-Datenbank herzustellen oder eine neue SQL-Datenbank erstellen. In diesem Abschnitt erstellen Sie eine SQL-Datenbank. > [AZURE.NOTE]Haben Sie bereits eine Datenbank in demselben Speicherort wie die mobile Anwendung Back-End-, können stattdessen **vorhandene Datenbank verwenden** , und wählen Sie dann die Datenbank. Die Verwendung einer Datenbank an einem anderen Speicherort wird durch höhere Wartezeiten nicht empfohlen. 6. **Klicken Sie im neuen Mobile App Backend** > **Mobile-Anwendung** > **Daten** > **+ Hinzufügen**. 7. Blade **Datenquelle hinzufügen** klicken Sie **SQL - Einstellungen konfigurieren** > **eine neue Datenbank erstellen**. Geben Sie im Feld **Name** den Namen der neuen Datenbank. 8. Klicken Sie auf **Server**. **Neue Server** -Blade Geben Sie einen eindeutigen Namen im Feld **Servername** und einen geeigneten **Server Administrator-Benutzernamen** und **Kennwort**. Sicherstellen Sie, dass **Zulassen Azure Services Zugriff auf Server** aktiviert ist. Klicken Sie auf **OK**. ![Erstellen Sie eine SQL Azure-Datenbank][6] 9. Das Blade **neue Datenbank** klicken Sie auf **OK**. 10. Wählen Sie auf das Blade **Datenquelle hinzufügen** **Verbindungszeichenfolge**, Benutzernamen und Kennwort, die Sie beim Erstellen der Datenbank bereitgestellt. Wenn Sie eine vorhandene Datenbank verwenden, bieten Sie die Anmeldeinformationen für die Datenbank. Einmal eingegeben haben, klicken Sie auf **OK**. 11. Wieder auf die **Datenquelle hinzufügen** , klicken Sie auf **OK** , um die Datenbank zu erstellen. <!--- END OF ALTERNATE INCLUDE --> Erstellung der Datenbank kann einige Minuten dauern. Verwenden **der Infobereich** den Fortschritt der Bereitstellung überwachen. Keine ausgeführt, bis die Datenbank erfolgreich bereitgestellt wurde. Nachdem erfolgreich bereitgestellt haben, wird eine Verbindungszeichenfolge für die SQL-Datenbankinstanz im mobilen Backend App-Einstellungen erstellt. Diese app Einstellung **Einstellungen**Siehe > **ApplicationSettings** > **Verbindungszeichenfolgen**. ### <a name="howto-tables-auth"></a>Gewusst wie: Authentifizierung für den Zugriff auf Tabellen Möchten Sie mit den Tabellen App-Authentifizierung verwenden, müssen Sie zuerst App-Authentifizierung in [Azure-Portal] konfigurieren. Weitere Informationen zum Konfigurieren der Authentifizierung in Azure App Service finden Sie im Konfigurationshandbuch für Identitätsanbieter verwenden möchten: - [Azure Active Directory-Authentifizierung konfigurieren] - [Facebook-Authentifizierung konfigurieren] - [Google-Authentifizierung konfigurieren] - [Microsoft Authentication konfigurieren] - [Twitter-Authentifizierung konfigurieren] Jede Tabelle verfügt über eine Access-Eigenschaft, die zum Steuern des Zugriffs auf die Tabelle verwendet werden kann. Im folgende Beispiel wird eine statisch definierte Tabelle mit Authentifizierung erforderlich. var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Define the columns within the table table.columns = { "text": "string", "complete": "boolean" }; // Turn off dynamic schema table.dynamicSchema = false; // Require authentication to access the table table.access = 'authenticated'; module.exports = table; Die Eigenschaft kann einen der drei Werte annehmen - *anonyme* bedeutet, dass die Clientanwendung ohne Authentifizierung lesen - *Authentifizierte* gibt an, dass die Clientanwendung ein gültigen Authentifizierungstokens mit der Anforderung gesendet werden müssen - *deaktiviert* bedeutet, dass diese Tabelle gesperrt ist Wenn die Eigenschaft nicht definiert ist, ist der nicht authentifizierter Zugriff zulässig. ### <a name="howto-tables-getidentity"></a>Gewusst wie: Authentifizierung Ansprüche von Tabellen verwenden Sie können verschiedene Ansprüche einrichten, die angefordert werden, wenn Authentifizierung eingerichtet wird. Diese Anträge sind nicht normalerweise durch die `context.user` Objekt. Aber sie können abgerufen werden mit der `context.user.getIdentity()` Methode. Die `getIdentity()` -Methode gibt ein Versprechen in ein Objekt aufgelöst wird. Das Objekt wird durch die Authentifizierungsmethode (Facebook, Google, Twitter, Microsoftaccount oder Aad) eingegeben. Wenn Sie Microsoft Account Authentifizierung und Anforderung Anspruch die e-Mail-Adressen einrichten, können Sie die e-Mail-Adresse auf den Datensatz mit der folgenden Tabelle hinzufügen: var azureMobileApps = require('azure-mobile-apps'); // Create a new table definition var table = azureMobileApps.table(); table.columns = { "emailAddress": "string", "text": "string", "complete": "boolean" }; table.dynamicSchema = false; table.access = 'authenticated'; /** * Limit the context query to those records with the authenticated user email address * @param {Context} context the operation context * @returns {Promise} context execution Promise */ function queryContextForEmail(context) { return context.user.getIdentity().then((data) => { context.query.where({ emailAddress: data.microsoftaccount.claims.emailaddress }); return context.execute(); }); } /** * Adds the email address from the claims to the context item - used for * insert operations * @param {Context} context the operation context * @returns {Promise} context execution Promise */ function addEmailToContext(context) { return context.user.getIdentity().then((data) => { context.item.emailAddress = data.microsoftaccount.claims.emailaddress; return context.execute(); }); } // Configure specific code when the client does a request // READ - only return records belonging to the authenticated user table.read(queryContextForEmail); // CREATE - add or overwrite the userId based on the authenticated user table.insert(addEmailToContext); // UPDATE - only allow updating of record belong to the authenticated user table.update(queryContextForEmail); // DELETE - only allow deletion of records belong to the authenticated uer table.delete(queryContextForEmail); module.exports = table; Um festzustellen, welche verfügbar sind, verwenden Sie einen Webbrowser an der `/.auth/me` Endpunkt der Website. ### <a name="howto-tables-disabled"></a>Gewusst wie: Deaktivieren des Zugriffs auf bestimmte Tabellenvorgänge Zusätzlich in der Tabelle angezeigt werden, kann die Eigenschaft steuern einzelner Arbeitsgänge verwendet werden. Es gibt vier Vorgänge aus: - *Lesen* ist RESTful GET-Vorgang für die Tabelle - *Legen Sie* ist den Rest POST-Vorgang für die Tabelle - RESTful PATCH-Vorgang für die Tabelle wird *aktualisiert* - *Löschen* ist der Arbeitsgang Rest Löschen der Tabelle Möglicherweise möchten eine schreibgeschützte nicht authentifizierte Tabelle: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Read-Only table - only allow READ operations table.read.access = 'anonymous'; table.insert.access = 'disabled'; table.update.access = 'disabled'; table.delete.access = 'disabled'; module.exports = table; ### <a name="howto-tables-query"></a>Gewusst wie: Anpassen die Abfrage, mit Tabelle Eine allgemeine Anforderung für Tabellenvorgänge soll eine eingeschränkte Ansicht der Daten. Beispielsweise können Sie eine Tabelle, die mit authentifizierten Benutzer-ID markiert ist, kann nur gelesen oder Ihre eigenen Datensätze aktualisieren. Die folgende Tabellendefinition stellt diese Funktionalität bereit: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Define a static schema for the table table.columns = { "userId": "string", "text": "string", "complete": "boolean" }; table.dynamicSchema = false; // Require authentication for this table table.access = 'authenticated'; // Ensure that only records for the authenticated user are retrieved table.read(function (context) { context.query.where({ userId: context.user.id }); return context.execute(); }); // When adding records, add or overwrite the userId with the authenticated user table.insert(function (context) { context.item.userId = context.user.id; return context.execute(); }); module.exports = table; Vorgänge, die normalerweise eine Abfrage ist eine Abfrageeigenschaft, die Sie, mit einer anpassen können Klausel. Abfrageeigenschaft ist ein [QueryJS] -Objekt, eine OData-Abfrage in etwas zu konvertieren, das Backend Daten verarbeiten kann. Einfache Übereinstimmungsvergleiche (wie die vorhergehenden) Fällen kann eine Zuordnung verwendet werden. Sie können auch SQL-Klauseln hinzufügen: context.query.where('myfield eq ?', 'value'); ### <a name="howto-tables-softdelete"></a>Gewusst wie: Weiches Löschen einer Tabelle konfigurieren Weiche wird nicht tatsächlich Datensätze löschen. Stattdessen werden sie in der Datenbank gelöscht, indem die gelöschte Spalte auf true festgelegt. Azure Mobile Apps SDK entfernt automatisch Datensätze gelöscht Ergebnisse, wenn Mobile Client SDK IncludeDeleted() verwendet. Zum Konfigurieren einer Tabelle für weiche löschen legen die `softDelete` -Eigenschaft in der Tabellendefinitionsdatei: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Define the columns within the table table.columns = { "text": "string", "complete": "boolean" }; // Turn off dynamic schema table.dynamicSchema = false; // Turn on Soft Delete table.softDelete = true; // Require authentication to access the table table.access = 'authenticated'; module.exports = table; Sie sollten einen Mechanismus zum Löschen von Datensätzen – entweder von einer Clientanwendung über ein Webauftrag Azure-Funktion oder eine benutzerdefinierte API einrichten. ### <a name="howto-tables-seeding"></a>Gewusst wie: Seeding die Datenbank mit Daten Beim Erstellen einer neuen Anwendung können Sie eine Tabelle mit Daten Startwert. Dies kann in der Tabelle Definition JavaScript-Datei wie folgt erfolgen: var azureMobileApps = require('azure-mobile-apps'); var table = azureMobileApps.table(); // Define the columns within the table table.columns = { "text": "string", "complete": "boolean" }; table.seed = [ { text: 'Example 1', complete: false }, { text: 'Example 2', complete: true } ]; // Turn off dynamic schema table.dynamicSchema = false; // Require authentication to access the table table.access = 'authenticated'; module.exports = table; Das Seeding der Daten erfolgt nur bei Azure Mobile Apps SDK ist. Wenn die Tabelle in der Datenbank vorhanden ist, werden keine Daten in die Tabelle eingefügt. Wenn dynamische Schema aktiviert ist, wird das Schema vergebene Daten abgeleitet. Wir empfehlen explizit aufrufen der `tables.initialize()` Methode, um die Tabelle zu erstellen, wenn der Dienst gestartet wird. ### <a name="Swagger"></a>Gewusst wie: Aktivieren der Unterstützung von stolz Azure App Service Mobile Apps kommt mit eingebauten [Swagger] unterstützt. Damit stolz Support zuerst installieren Sie stolz-Benutzeroberfläche als Abhängigkeit: npm install --save swagger-ui Nach der Installation können Sie stolz Unterstützung in Azure Mobile Apps-Konstruktor: var mobile = azureMobileApps({ swagger: true }); Sie wahrscheinlich nur stolz Unterstützung in Development Edition aktivieren möchten. Nutzen Sie dazu die `NODE_ENV` app-Einstellung: var mobile = azureMobileApps({ swagger: process.env.NODE_ENV !== 'production' }); Der stolz Endpunkt befindet sich unter http://_Standortname_.azurewebsites.net/swagger. Swagger-Benutzeroberfläche über Zugriff der `/swagger/ui` Endpunkt. Wunsch der gesamten Anwendung Authentifizierung erzeugt stolz einen Fehler. Für optimale Ergebnisse festlegen, nicht authentifizierte Anfragen in Azure App-Authentifizierung / Autorisierung, dann Einstellungen Authentifizierung der `table.access` Eigenschaft. Sie können auch die Option stolz Ihre `azureMobile.js` Datei, wenn Sie nur stolz Unterstützung bei lokal. ## <a name="a-namepushpush-notifications"></a><a name="push">Push-Benachrichtigung Mobile Apps integriert Azure Notification Hubs können Millionen von Geräten für alle wichtigen Plattformen gezielte Pushbenachrichtigungen an. Mit Notification Hubs, schicken Sie Pushbenachrichtigungen zu iOS, Android und Windows. Sie erfahren können mit Notification Hubs, finden Sie unter [Übersicht über Notification Hubs](../notification-hubs/notification-hubs-push-notification-overview.md). ### </a><a name="send-push"></a>Gewusst wie: Senden von Pushbenachrichtigungen Im folgenden Codebeispiel wird die Verwendung das Push-Objekt registrierten iOS Geräte broadcast Push-Benachrichtigung an: // Create an APNS payload. var payload = '{"aps": {"alert": "This is an APNS payload."}}'; // Only do the push if configured if (context.push) { // Send a push notification using APNS. context.push.apns.send(null, payload, function (error) { if (error) { // Do something or log the error. } }); } Erstellen einer Vorlage Push Registrierung vom Client, können Sie stattdessen eine Vorlagennachricht Geräte auf allen unterstützten Plattformen senden. Der folgende Code zeigt, wie eine Vorlage Benachrichtigung: // Define the template payload. var payload = '{"messageParam": "This is a template payload."}'; // Only do the push if configured if (context.push) { // Send a template notification. context.push.send(null, payload, function (error) { if (error) { // Do something or log the error. } }); } ###<a name="push-user"></a>Gewusst wie: Senden von Pushbenachrichtigungen für einen authentifizierten Benutzer mit Tags Wenn ein authentifizierter Benutzer für Pushbenachrichtigungen registriert, ist eine Benutzer-ID-Tag die Registrierung automatisch hinzugefügt. Mithilfe dieses Tags können Sie alle Geräte, die von einem bestimmten Benutzer registriert Push-Benachrichtigung senden. Der folgende Code Ruft die SID des Benutzers und sendet eine Push-vorlagenbenachrichtigung jedes Gerät Registrierung für diesen Benutzer: // Only do the push if configured if (context.push) { // Send a notification to the current user. context.push.send(context.user.id, payload, function (error) { if (error) { // Do something or log the error. } }); } Anmeldung von einem authentifizierten Client Pushbenachrichtigungen, sicherzustellen Sie, dass die Authentifizierung abgeschlossen ist, bevor Sie versuchen, die Registrierung. ## <a name="CustomAPI"></a>Benutzerdefinierte APIs ### <a name="howto-customapi-basic"></a>Gewusst wie: definieren eine benutzerdefinierte API Neben der Datenzugriffs-API über den Endpunkt Tables bieten Azure Mobile Apps benutzerdefinierte API-Abdeckung. Benutzerdefinierte APIs ähnlich die Tabellendefinitionen definiert und können die gleichen Funktionen, einschließlich Authentifizierung zugreifen. Wenn Sie eine benutzerdefinierte API App-Authentifizierung verwenden möchten, müssen Sie zuerst App-Authentifizierung in [Azure-Portal] konfigurieren. Weitere Informationen zum Konfigurieren der Authentifizierung in Azure App Service finden Sie im Konfigurationshandbuch für Identitätsanbieter verwenden möchten: - [Azure Active Directory-Authentifizierung konfigurieren] - [Facebook-Authentifizierung konfigurieren] - [Google-Authentifizierung konfigurieren] - [Microsoft Authentication konfigurieren] - [Twitter-Authentifizierung konfigurieren] Benutzerdefinierte APIs sind in fast genauso wie die Tabellen-API definiert. 1. Ein **api** -Verzeichnis erstellen 2. Erstellen Sie eine Definitionsdatei JavaScript API in **API-** Verzeichnis. 3. Verwenden Sie die Importmethode **API-** Verzeichnis importieren. Hier ist der Prototyp-API-Definition basierend auf den zuvor verwendeten Basic-app. var express = require('express'), azureMobileApps = require('azure-mobile-apps'); var app = express(), mobile = azureMobileApps(); // Import the Custom API mobile.api.import('./api'); // Add the mobile API so it is accessible as a Web API app.use(mobile); // Start listening on HTTP app.listen(process.env.PORT || 3000); Nehmen Sie eine Beispiel-API, die das Datum des Servers mithilfe der _Date.now()_ -Methode zurückgegeben. Hier wird die api/date.js-Datei: var api = { get: function (req, res, next) { var date = { currentTime: Date.now() }; res.status(200).type('application/json').send(date); }); }; module.exports = api; Jeder Parameter ist der Rest Standardverben - GET, POST, PATCH oder löschen. Die Methode ist eine Standardfunktion [ExpressJS Middleware] die erforderliche Ausgabe gesendet. ### <a name="howto-customapi-auth"></a>Gewusst wie: Authentifizierung für den Zugriff auf eine benutzerdefinierte API Azure Mobile Apps SDK implementiert Authentifizierung auf die gleiche Weise für die Tabellen Endpunkt und benutzerdefinierte APIs. Um Authentifizierung API entwickelt im vorherigen Abschnitt hinzuzufügen, fügen Sie einer **Access** -Eigenschaft: var api = { get: function (req, res, next) { var date = { currentTime: Date.now() }; res.status(200).type('application/json').send(date); }); }; // All methods must be authenticated. api.access = 'authenticated'; module.exports = api; Sie können auch auf bestimmte Vorgänge Authentifizierung angeben: var api = { get: function (req, res, next) { var date = { currentTime: Date.now() }; res.status(200).type('application/json').send(date); } }; // The GET methods must be authenticated. api.get.access = 'authenticated'; module.exports = api; Für Tabellen-Endpunkt wird auch muss für benutzerdefinierte Authentifikation APIs verwendet werden. ### <a name="howto-customapi-auth"></a>Gewusst wie: Behandeln von großen Dateiuploads Azure Mobile Apps SDK verwendet [Text-Parser Middleware](https://github.com/expressjs/body-parser) akzeptieren und Inhalt Ihrer Einsendung decodieren. Sie können Text-Parser, um größere Dateiuploads akzeptieren vorkonfigurieren: var express = require('express'), bodyParser = require('body-parser'), azureMobileApps = require('azure-mobile-apps'); var app = express(), mobile = azureMobileApps(); // Set up large body content handling app.use(bodyParser.json({ limit: '50mb' })); app.use(bodyParser.urlencoded({ limit: '50mb', extended: true })); // Import the Custom API mobile.api.import('./api'); // Add the mobile API so it is accessible as a Web API app.use(mobile); // Start listening on HTTP app.listen(process.env.PORT || 3000); Die Datei ist Base64-codierte vor der Übertragung. Das tatsächliche Upload vergrößert und somit die Größe Sie berücksichtigt. ### <a name="howto-customapi-sql"></a>Gewusst wie: Ausführen von benutzerdefinierten SQL-Anweisung Azure Mobile Apps SDK ermöglicht den Zugriff auf den gesamten Kontext durch das Anforderungsobjekt ermöglicht parametrisierte SQL-Anweisungen Anbieter definierten Daten problemlos ausführen: var api = { get: function (request, response, next) { // Check for parameters - if not there, pass on to a later API call if (typeof request.params.completed === 'undefined') return next(); // Define the query - anything that can be handled by the mssql // driver is allowed. var query = { sql: 'UPDATE TodoItem SET complete=@completed', parameters: [{ completed: request.params.completed }] }; // Execute the query. The context for Azure Mobile Apps is available through // request.azureMobile - the data object contains the configured data provider. request.azureMobile.data.execute(query) .then(function (results) { response.json(results); }); } }; api.get.access = 'authenticated'; module.exports = api; ## <a name="Debugging"></a>Debuggen, einfache Tabellen und einfache APIs ### <a name="howto-diagnostic-logs"></a>Gewusst wie: Debuggen, diagnose und Problembehandlung bei Azure Mobile apps Azure App Service bietet verschiedene Debuggen und Problembehandlungsverfahren für Node.js-Applikationen. Finden Sie in den folgenden Artikeln bei Ihrem Node.js Mobile Back-End-Einstieg: - [Überwachen von Azure App Service] - [Diagnoseprotokoll in Azure App Service aktivieren] - [Problembehandlung bei einer Azure App Service in Visual Studio] Node.js Applikationen haben Zugriff auf eine Vielzahl von Tools Diagnoseprotokoll. Intern verwendet Azure Mobile Apps Node.js SDK [Winston] für Diagnoseprotokoll. Protokollierung ist Debug-Modus aktivieren oder **MS_DebugMode** app Einstellung true im [Azure-Portal]automatisch aktiviert. Generierte Protokolle in die Diagnoseprotokolle [Azure-Portal]angezeigt. ### <a name="in-portal-editing"></a><a name="work-easy-tables"></a>Gewusst wie: Arbeiten mit einfachen Tabellen im Azure-Portal Einfache Tabellen im Portal können Sie erstellen und Arbeiten mit Tabellen im Portal. Sie können auch Tabellenvorgänge im App Service-Editor bearbeiten. Wenn Sie **einfache Tabellen** in der Back-End-Standortparameter klicken, können Sie hinzufügen, ändern oder Löschen einer Tabelle. Sie können auch Daten in der Tabelle angezeigt. ![Einfache Tabellen arbeiten](./media/app-service-mobile-node-backend-how-to-use-server-sdk/mobile-apps-easy-tables.png) Die folgenden Befehle stehen auf der Befehlsleiste eine Tabelle: + **Berechtigungen ändern** - ändern Sie die Berechtigung zum Lesen, einfügen, aktualisieren und löschen Vorgänge in der Tabelle. Optionen sind anonymen Zugriff, Authentifizierung und Zugriff auf die Operation deaktivieren. + **Skript bearbeiten** - Skriptdatei für die Tabelle wird im App Service-Editor geöffnet. + **Schema verwalten** - hinzufügen oder Löschen von Spalten oder ändern den Tabellenindex. + **Tabelle löschen** - schneidet eine vorhandene Tabelle löschen alle Datenzeilen aber Verlassen des Schemas unverändert. + **Zeilen löschen** - Löschen einzelner Datenzeilen. + **Streaming-Protokolle anzeigen** - Verbindung streaming Protokolldienst für Ihre Website. ###<a name="work-easy-apis"></a>Gewusst wie: Arbeiten mit einfachen APIs im Azure-Portal Einfache APIs im Portal ermöglichen das Erstellen und Verwenden von benutzerdefinierten APIs direkt im Portal. Sie können API Skripts mit dem App-Service-Editor bearbeiten. Klick **Einfach APIs** in der Back-End-Standortparameter können Sie hinzufügen, ändern oder Löschen einen benutzerdefinierten API-Endpunkt. ![Arbeiten mit einfachen APIs](./media/app-service-mobile-node-backend-how-to-use-server-sdk/mobile-apps-easy-apis.png) Im Portal können Sie Zugriffsberechtigungen für HTTP-Aktionen ändern, API-Skriptdatei im App Service-Editor bearbeiten oder Streaming-Protokolle anzeigen. ###<a name="online-editor"></a>Gewusst wie: Bearbeiten von Code im App Service-Editor Azure-Portal können Sie Node.js Backend-Skriptdateien in App Service-Editor bearbeiten, ohne das Projekt auf dem lokalen Computer herunterladen. So bearbeiten Sie Skriptdateien in den online-editor 1. Klicken Sie in der Back-End-Blade Mobile-Anwendung auf **Alle** > **einfache Tabellen** oder **Einfach APIs**, einer Tabelle oder einer API klicken **Skript bearbeiten**. Die Skriptdatei wird der App Service-Editor geöffnet. ![App Service-Editor](./media/app-service-mobile-node-backend-how-to-use-server-sdk/mobile-apps-visual-studio-editor.png) 2. Ändern Sie die Datei in der online-Editor. Daten werden während der Eingabe automatisch gespeichert. <!-- Images --> [0]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/npm-init.png [1]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/vs2015-new-project.png [2]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/vs2015-install-npm.png [3]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/sqlexpress-config.png [4]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/sqlexpress-authconfig.png [5]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/sqlexpress-newuser-1.png [6]: ./media/app-service-mobile-node-backend-how-to-use-server-sdk/dotnet-backend-create-db.png <!-- URLs --> [Schnelleinstieg Android-Client]: app-service-mobile-android-get-started.md [Apache Cordova Client Schnellstart]: app-service-mobile-cordova-get-started.md [iOS-Client Schnellstart]: app-service-mobile-ios-get-started.md [Xamarin.iOS Client-Schnellstart]: app-service-mobile-xamarin-ios-get-started.md [Xamarin.Android Client-Schnellstart]: app-service-mobile-xamarin-android-get-started.md [Xamarin.Forms Client-Schnellstart]: app-service-mobile-xamarin-forms-get-started.md [Windows Store Client Schnellstart]: app-service-mobile-windows-store-dotnet-get-started.md [HTML/Javascript Client QuickStart]: app-service-html-get-started.md [Offline-Daten synchronisieren]: app-service-mobile-offline-data-sync.md [Azure Active Directory-Authentifizierung konfigurieren]: app-service-mobile-how-to-configure-active-directory-authentication.md [Facebook-Authentifizierung konfigurieren]: app-service-mobile-how-to-configure-facebook-authentication.md [Google-Authentifizierung konfigurieren]: app-service-mobile-how-to-configure-google-authentication.md [Microsoft Authentication konfigurieren]: app-service-mobile-how-to-configure-microsoft-authentication.md [Twitter-Authentifizierung konfigurieren]: app-service-mobile-how-to-configure-twitter-authentication.md [Azure App Service Deployment Guide]: ../app-service-web/web-sites-deploy.md [Überwachen von Azure App Service]: ../app-service-web/web-sites-monitor.md [Diagnoseprotokoll in Azure App Service aktivieren]: ../app-service-web/web-sites-enable-diagnostic-log.md [Problembehandlung bei einer Azure App Service in Visual Studio]: ../app-service-web/web-sites-dotnet-troubleshoot-visual-studio.md [Geben Sie die Knoten-Version]: ../nodejs-specify-node-version-azure-apps.md [Knoten Module]: ../nodejs-use-node-modules-azure-apps.md [Create a new Azure App Service]: ../app-service-web/ [azure-mobile-apps]: https://www.npmjs.com/package/azure-mobile-apps [Express]: http://expressjs.com/ [Stolz]: http://swagger.io/ [Azure-portal]: https://portal.azure.com/ [OData]: http://www.odata.org [Promise]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Promise [Basicapp-Beispiel GitHub]: https://github.com/azure/azure-mobile-apps-node/tree/master/samples/basic-app [TODO-Beispiel GitHub]: https://github.com/azure/azure-mobile-apps-node/tree/master/samples/todo [Beispielverzeichnis GitHub]: https://github.com/azure/azure-mobile-apps-node/tree/master/samples [static-schema sample on GitHub]: https://github.com/azure/azure-mobile-apps-node/tree/master/samples/static-schema [QueryJS]: https://github.com/Azure/queryjs [Node.js Tools 1.1 für Visual Studio]: https://github.com/Microsoft/nodejstools/releases/tag/v1.1-RC.2.1 [MSSQL Node.js-Paket]: https://www.npmjs.com/package/mssql [Microsoft SQL Server 2014 Express]: http://www.microsoft.com/en-us/server-cloud/Products/sql-server-editions/sql-server-express.aspx [ExpressJS-Middleware]: http://expressjs.com/guide/using-middleware.html [Winston]: https://github.com/winstonjs/winston
54.721154
595
0.745407
deu_Latn
0.96701
87917505f16decce74176a46314d2427dd9a9962
2,294
md
Markdown
_posts/2009-04-06-slidetablerows.md
aterai/jekyll-bootstrap
4d3db7adc582963df7a9e6dbbb6e49f1a3c18b4a
[ "MIT" ]
null
null
null
_posts/2009-04-06-slidetablerows.md
aterai/jekyll-bootstrap
4d3db7adc582963df7a9e6dbbb6e49f1a3c18b4a
[ "MIT" ]
null
null
null
_posts/2009-04-06-slidetablerows.md
aterai/jekyll-bootstrap
4d3db7adc582963df7a9e6dbbb6e49f1a3c18b4a
[ "MIT" ]
null
null
null
--- layout: post category: swing folder: SlideTableRows title: JTableで行の追加、削除アニメーション tags: [JTable, Animation] author: aterai pubdate: 2009-04-06T14:03:11+09:00 description: JTableの行追加や削除をスライドアニメーションで強調します。 image: https://lh6.googleusercontent.com/_9Z4BYR88imo/TQTTP0i2yxI/AAAAAAAAAkE/DQKpmn3BIQo/s800/SlideTableRows.png hreflang: href: https://java-swing-tips.blogspot.com/2009/04/animating-jtable-rows.html lang: en comments: true --- ## 概要 `JTable`の行追加や削除をスライドアニメーションで強調します。 {% download https://lh6.googleusercontent.com/_9Z4BYR88imo/TQTTP0i2yxI/AAAAAAAAAkE/DQKpmn3BIQo/s800/SlideTableRows.png %} ## サンプルコード <pre class="prettyprint"><code>private void testCreateActionPerformed(ActionEvent e) { model.addTest(new Test("New name", "")); (new Timer(DELAY, new ActionListener() { int i = table.convertRowIndexToView(model.getRowCount() - 1); int h = START_HEIGHT; @Override public void actionPerformed(ActionEvent e) { if (h &lt; END_HEIGHT) { table.setRowHeight(i, h++); } else { ((Timer) e.getSource()).stop(); } } })).start(); } private void deleteActionPerformed(ActionEvent evt) { final int[] selection = table.getSelectedRows(); if (selection.length &lt;= 0) { return; } (new Timer(DELAY, new ActionListener() { int h = END_HEIGHT; @Override public void actionPerformed(ActionEvent e) { h--; if (h &gt; START_HEIGHT) { for (int i = selection.length - 1; i &gt;= 0; i--) table.setRowHeight(selection[i], h); } else { ((Timer) e.getSource()).stop(); for (int i = selection.length - 1; i &gt;= 0; i--) { model.removeRow(table.convertRowIndexToModel(selection[i])); } } } })).start(); } </code></pre> ## 解説 上記のサンプルでは、`javax.swing.Timer`を使用して徐々に行の高さを拡大、または縮小することで、追加と削除のアニメーションを行っています。 - 行の追加アニメーション - 高さ`0`の行を追加したあと`JTable#setRowHeight(int, int)`メソッドを使用してその高さをデフォルトの高さになるまで拡大 - 行の削除アニメーション - 選択された行の高さを`JTable#setRowHeight(int, int)`メソッドを使用して縮小 - 高さが一定以下になったらその行を削除 <!-- dummy comment line for breaking list --> ## 参考リンク - [JTableの行を追加、削除](https://ateraimemo.com/Swing/AddRow.html) - [JTableの行の高さを変更する](https://ateraimemo.com/Swing/FishEyeTable.html) <!-- dummy comment line for breaking list --> ## コメント
29.410256
121
0.685266
yue_Hant
0.627949
87926ee69f52605eb80376a29ae4be24f1158033
1,658
md
Markdown
azure-monitor-ref/tables/containerlogv2.md
UCOwner/azure-reference-other
70d137672684cf9f137df24f46224ca2744bfcea
[ "CC-BY-4.0", "MIT" ]
null
null
null
azure-monitor-ref/tables/containerlogv2.md
UCOwner/azure-reference-other
70d137672684cf9f137df24f46224ca2744bfcea
[ "CC-BY-4.0", "MIT" ]
null
null
null
azure-monitor-ref/tables/containerlogv2.md
UCOwner/azure-reference-other
70d137672684cf9f137df24f46224ca2744bfcea
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Azure Monitor Logs reference - ContainerLogV2 description: Reference for ContainerLogV2 table in Azure Monitor Logs. ms.topic: reference ms.service: azure-monitor ms.subservice: logs ms.author: bwren author: bwren ms.date: 12/30/2021 --- # ContainerLogV2 Kubernetes Container logs in V2 schema. This is the successor for ContainerLog. This has more friendlier schema, specifically for kubernetes orchestrated containers in pods. ## Categories - Containers ## Solutions - ContainerInsights ## Resource types - Kubernetes Services ## Columns | Column | Type | Description | | --- | --- | --- | | Computer | string | Name of the Computer/Node generating the log. | | ContainerId | string | Container ID of the log source as seen by the Container engine. | | ContainerName | string | Name of the Container generating the log. | | LogMessage | dynamic | Log message from stdout or stderr. Being a dynmic field, json log messages can be queried without parse_json. | | LogSource | string | Source of the Log message. Possible vlaues are stdout or stderr. | | PodName | string | Kubernetes Pod name for the Container generating the log. | | PodNamespace | string | Kubernetes Namespace for the container's pod. | | _ResourceId | string | A unique identifier for the resource that the record is associated with | | SourceSystem | string | | | _SubscriptionId | string | A unique identifier for the subscription that the record is associated with | | TenantId | string | | | TimeGenerated | datetime | The timestamp (UTC) of when the log was generated. | | Type | string | The name of the table |
36.043478
175
0.725573
eng_Latn
0.975123
87936a5ab9156a48d87e6941ccad8e7788a91e91
24
md
Markdown
README.md
Arnukk/Matlab
6ab5c5c2160c96bb418700a5e635f41899e9fb24
[ "MIT" ]
null
null
null
README.md
Arnukk/Matlab
6ab5c5c2160c96bb418700a5e635f41899e9fb24
[ "MIT" ]
null
null
null
README.md
Arnukk/Matlab
6ab5c5c2160c96bb418700a5e635f41899e9fb24
[ "MIT" ]
null
null
null
# Matlab Matlab Scripts
8
14
0.791667
zsm_Latn
0.994126
8793f79c1d7e1e0d90a384b951c4671a559024ea
20
md
Markdown
README.md
huomiaowang/studyssh
36207a1ea2fdeb474fb2af4487e4fdeced23b760
[ "Apache-2.0" ]
null
null
null
README.md
huomiaowang/studyssh
36207a1ea2fdeb474fb2af4487e4fdeced23b760
[ "Apache-2.0" ]
null
null
null
README.md
huomiaowang/studyssh
36207a1ea2fdeb474fb2af4487e4fdeced23b760
[ "Apache-2.0" ]
null
null
null
# studyssh studyssh
6.666667
10
0.8
eng_Latn
0.696064
879462c2aa01686a3a6e7fd2f3b1ed670098cd3f
5,968
md
Markdown
node_modules/atlassian-oauth2/node_modules/jsuri/README.md
karloconnor67/Jira-StoryPoint-Dashboard-Item
c9ed0d0ef3c9be75144de7ffbda429861c77674e
[ "Apache-2.0" ]
null
null
null
node_modules/atlassian-oauth2/node_modules/jsuri/README.md
karloconnor67/Jira-StoryPoint-Dashboard-Item
c9ed0d0ef3c9be75144de7ffbda429861c77674e
[ "Apache-2.0" ]
null
null
null
node_modules/atlassian-oauth2/node_modules/jsuri/README.md
karloconnor67/Jira-StoryPoint-Dashboard-Item
c9ed0d0ef3c9be75144de7ffbda429861c77674e
[ "Apache-2.0" ]
null
null
null
jsUri ===== Uri and query string manipulation in javascript. This project incorporates the excellent [parseUri](http://blog.stevenlevithan.com/archives/parseuri) regular expression library by Steven Levithan. You can safely parse URLs of all shapes and sizes, however invalid or hideous. Usage ----- Pass anything that your browser would recognize as a url to the new Uri() constructor var uri = new Uri('http://user:pass@www.test.com:81/index.html?q=books#fragment'); and then use the following accessor methods to get at the various parts. uri.protocol(); // http uri.userInfo(); // user:pass uri.host(); // www.test.com uri.port(); // 81 uri.path(); // /index.html uri.query(); // q=books uri.anchor(); // fragment The accessor methods accept an optional value for setting the property uri.protocol('https'); uri.toString(); // https://user:pass@www.test.com:81/index.html?q=books#fragment uri.host('mydomain.com'); uri.toString(); // https://user:pass@www.mydomain.com:81/index.html?q=books#fragment Fluent Setters -------------- The fluent interface provides a simple way to chain property assignment new Uri() .setPath('/index.html') .setAnchor('content') .setHost('www.test.com') .setPort(8080) .setUserInfo('username:password') .setProtocol('https') .setQuery('this=that&some=thing') // https://username:password@www.test.com:8080/index.html?this=that&some=thing#content new Uri('http://www.test.com') .setHost('www.yahoo.com') .setProtocol('https') // https://www.yahoo.com new Uri() .setPath('/archives/1979/') .setQuery('?page=1') // /archives/1979?page=1 Query Parameter Access and Manipulation --------------------------------------- Special methods are available for fetching, building and modifying query string parameters. An emhpasis is placed on query string integrity; duplicate parameter names and values are preserved. Parameter ordering is preserved when possible. URI Components are decoded for comparision, but are otherwise left in their original state. ### Getting query param values by name Returns the first query param value for the key new Uri('?cat=1&cat=2&cat=3').getQueryParamValue('cat') // 1 Returns all query param values for the given key new Uri('?cat=1&cat=2&cat=3').getQueryParamValues('cat') // [1, 2, 3] ### Getting all query param keys and values Internally, query key/value pairs are stored as a series of two-value arrays in the Query object new Uri('?a=b&c=d').query().params // [ ['a', 'b'], ['c', 'd']] ### Adding query param values new Uri().addQueryParam('q', 'books') // ?q=books new Uri('http://www.github.com') .addQueryParam('testing', '123') .addQueryParam('one', 1) // http://www.github.com/?testing=123&one=1 // insert param at index 0 new Uri('?b=2&c=3&d=4').addQueryParam('a', '1', 0) // ?a=1&b=2&c=3&d=4 ### Replacing query param values Replaces every query string parameter named `key` with a single instance with the value `newVal`. If `oldValue` is supplied, only parameters valued `oldVal` will be replaced. new Uri('?a=1&b=2&c=3') .replaceQueryParam('a', 'eh') // ?a=eh&b=2&c=3 new Uri('?a=1&b=2&c=3&c=4&c=5&c=6') .replaceQueryParam('c', 'five', '5') // ?a=1&b=2&c=3&c=4&c=five&c=6 new Uri().replaceQueryParam('page', 2) // ?page=2 ### Deleting query param values Removes instances of query parameters named `key`. If `value` is passed, only params named `key` and valued `value` will be deleted. new Uri('?a=1&b=2&c=3') .deleteQueryParam('a') // ?b=2&c=3 new Uri('test.com?a=1&b=2&c=3&a=eh') .deleteQueryParam('a', 'eh') // test.com/?a=1&b=2&c=3 Object Cloning -------------- Duplication (via `.clone()`) is an easy way to inflate an identical uri object, which you can muck around with as much as you like without destroying the original. var baseUri = new Uri('http://localhost/'); baseUri.clone().setProtocol('https'); // https://localhost/ baseUri; // http://localhost/ Testing ------- There is a comprensive set of unit tests written in [jasmine](http://pivotal.github.com/jasmine/). To run them, simply open `testrunner.html` from the root of the project in a browser. License ------- Copyright (c) 2012 Derek Watson Copyright (c) 2007 Steven Levithan Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
39.263158
331
0.629692
eng_Latn
0.692617
8794d400c1b78587192b768d8f96d8b101732e6f
181
md
Markdown
node_modules/github/README.md
hrithiksh101/final-pds-project
c7edac20b6f5212db84ab325c4b67e3b4f363c26
[ "MIT" ]
null
null
null
node_modules/github/README.md
hrithiksh101/final-pds-project
c7edac20b6f5212db84ab325c4b67e3b4f363c26
[ "MIT" ]
null
null
null
node_modules/github/README.md
hrithiksh101/final-pds-project
c7edac20b6f5212db84ab325c4b67e3b4f363c26
[ "MIT" ]
null
null
null
# "github" has been renamed to "[@octokit/rest](https://www.npmjs.com/package/@octokit/rest)" See https://github.com/octokit/rest.js/releases/tag/v14.0.0 for upgrade instructions.
45.25
93
0.745856
eng_Latn
0.488671
8795aa129a544b83655e1466ebd2e4e78a51dd73
113
md
Markdown
README.md
moeinrahimi/snail
c164546df33d291cb7d5c799ccd98a0bbf8b671e
[ "MIT" ]
null
null
null
README.md
moeinrahimi/snail
c164546df33d291cb7d5c799ccd98a0bbf8b671e
[ "MIT" ]
null
null
null
README.md
moeinrahimi/snail
c164546df33d291cb7d5c799ccd98a0bbf8b671e
[ "MIT" ]
null
null
null
# snail example of crawling a website(here http://wallpaperswide.com/) to get all wallpapers based on given name
37.666667
104
0.787611
eng_Latn
0.998029
8795b7b23fc7c29eed56eb4c7ca52f0ea2d77b63
592
md
Markdown
content/authors/Maciek_Tomczak/_index.md
maxdiluca/ARME
7190560b607b9f983266064681d477f185980df4
[ "MIT" ]
4
2021-11-03T15:19:03.000Z
2022-02-08T16:23:04.000Z
content/authors/Maciek_Tomczak/_index.md
maxdiluca/ARME
7190560b607b9f983266064681d477f185980df4
[ "MIT" ]
null
null
null
content/authors/Maciek_Tomczak/_index.md
maxdiluca/ARME
7190560b607b9f983266064681d477f185980df4
[ "MIT" ]
2
2021-09-17T12:02:53.000Z
2021-11-04T11:31:21.000Z
--- # Display name title: Maciek Tomczak # Role/position role: Post-doctoral Research Fellow # Username (this should match the folder name) authors: - Maciek_Tomczak # Is this the primary user of the site? superuser: false # Organizations/Affiliations organizations: - name: Birmingham City University url: "" highlight_name: false avatar_filename: avatar interests: - Audio engineering - Digital signal processing - Music informatics - Machine learning social: - icon: web icon_pack: fab link: https://maciek-tomczak.github.io/ email: "" user_groups: - Researchers ---
14.095238
46
0.744932
eng_Latn
0.729464
8795e19531a27ba7ede600370c22b372ccdc30b9
5,487
md
Markdown
_posts/2016-10-20-rest-v2-3-1-patch-notes.md
Naltrexone/developer-dot
6fb248bfe8c3b66877af49edd6bea9981c41cce8
[ "MIT" ]
null
null
null
_posts/2016-10-20-rest-v2-3-1-patch-notes.md
Naltrexone/developer-dot
6fb248bfe8c3b66877af49edd6bea9981c41cce8
[ "MIT" ]
1
2018-10-25T22:31:25.000Z
2018-10-25T22:31:25.000Z
_posts/2016-10-20-rest-v2-3-1-patch-notes.md
Naltrexone/developer-dot
6fb248bfe8c3b66877af49edd6bea9981c41cce8
[ "MIT" ]
1
2018-09-21T19:53:27.000Z
2018-09-21T19:53:27.000Z
--- layout: post title: REST v2.3.1 Patch Notes date: 2016-10-20 11:00 author: Ted Spence comments: true categories: [avatax, patch notes] product: blog doctype: blog disqus: 1 --- <h2>REST v2.3.1 Patch Notes</h2> For those of you who participated in the AvaTax REST v2 Preview program, I'd like to take this opportunity to thank you for your time and effort helping us debug a huge software release. We've now implemented a clean, modern, consistent REST API that covers tax functionality from top to bottom - and we've established a great platform for continuing improvements. Now that the program is winding down, please take note of a few differences between the final release of REST v2 and the preview program you tested in August/September: <h3>New URL for REST v2</h3> The existing URL, <a href="https://rest-sbx-preview.avalara.net">https://rest-sbx-preview.avalara.net</a>, will be retired and no longer available after October 25th, 2016. If you have any applications or services using this URL, please update them to point to the new URLs: <ul class="normal"> <li><a href="https://sandbox-rest.avatax.com">https://sandbox-rest.avatax.com</a> (for users of the "Sandbox" test development environment, available now)</li> <li><a href="https://rest.avatax.com">https://rest.avatax.com</a> (for Production accounts, available October 25th 2016)</li> </ul> <h3>Changes to /api/v2/transactions/create</h3> The `/api/v2/transactions/create` endpoint now only allows creation of one transaction at a time. If any of your existing programs were creating multiple transactions with a single API call, you will need to: <ul class="normal"> <li>Modify them to use one API call per each transaction, or</li> <li>Modify them to use the <code class="highlighter-rouge">/api/v2/companies/123/batches/create</code> endpoint instead.</li> </ul> <h3>Posting and Committing Transactions</h3> To avoid confusion between the action "POST" and the HTTP verb "POST", we have renamed and simplified the `/api/v2/transactions/123/post` API call to "settle". Some users were also confused about the difference between a call to `/api/v2/transactions/123/post` vs a call to `/api/v2/transactions/123/commit`. To reduce this confusion, we have split these API calls into individual functions: <pre>POST /api/v2/companies/ABC/transactions/DEF/verify</pre> Ensures that a transaction's amounts match an expected value, and returns an error if they do not match. <pre>POST /api/v2/companies/ABC/transactions/DEF/changecode</pre> Changes the transaction code of a specified transaction. <pre>POST /api/v2/companies/ABC/transactions/DEF/commit</pre> Commits a transaction so that it can be reported in a tax return. Each of these three APIs is separate and does not depend on the others. Once you have created a transaction, you can call any of these three APIs to execute any one of these three functions separately. If you'd still like to call all three of these API calls at once, you can use "settle": <pre>POST /api/v2/companies/ABC/transactions/DEF/settle</pre> This endpoint allows you to execute all three actions, verify, changecode, and commit, in a single API call. <h3>Single Object Creation</h3> If you call an API that allows multiple object creation, but you forget to send an array, our API will now accept a single object and continue to work rather than reporting an error. <h3>Improved API Documentation</h3> When browsing the online swagger documentation for REST v2, you will notice that objects have rewritten help text and example objects visible within the online documentation. <h3>Error Messaging Improvements</h3> Error Message objects have been rewritten to match Microsoft's updated Error Message standards, and all error messages contain a URL that points to a web page explaining the error in detail. Here's an example of the new error objects: ```json { "error": { "code": "ModelStateInvalid", "message": "One or more parameters in your request were incorrect.", "target": "HttpRequest", "details": [ { "code": "ModelStateInvalid", "number": 70, "message": "One or more parameters in your request were incorrect.", "description": "The value '-0-' is not a valid 'id'.", "faultCode": "Client", "helpLink": "/avatax/errors/ModelStateInvalid", "severity": "Error" } ] } } ``` You can preview a list of error messages recognized by AvaTax REST v2 on the <a href="/avatax/errors/">/avatax/errors/</a> page. <h3>Client Identification</h3> In each AvaTax API request, you can provide a request header called `X-Avalara-Client`. This value identifies your application and helps Avalara track down problems and identify when customers are using an outdated version of the software. The client identification tag looks like this: ``` X-Avalara-Client: (application name); (application version); (library name); (library version); (machine name) ``` <h3>Other Improvements</h3> We made hundreds of individual improvements to the API based on your feedback and based on internal testing. Many of these changes helped to clarify error messages that were confusing, or identify unsafe API calls - calls that would result in invalid tax configuration settings, for example. We welcome your feedback - please continue to share your experience of the REST v2 API and we will be happy to continue to work with you! --Ted Spence, Director, AvaTax Core Engine
48.557522
365
0.753053
eng_Latn
0.989321
8796714238bf3bf2c5b0f6621807aa0a5d55d526
2,327
md
Markdown
other/cheeki-breeki-chebureki.md
caligin/actual-cookbook
26bc039f749761417438ff2945b45b88ad6bbc6e
[ "Unlicense" ]
6
2019-03-28T23:04:53.000Z
2021-08-15T17:47:26.000Z
other/cheeki-breeki-chebureki.md
caligin/actual-cookbook
26bc039f749761417438ff2945b45b88ad6bbc6e
[ "Unlicense" ]
28
2017-12-30T20:12:29.000Z
2021-10-31T17:26:42.000Z
other/cheeki-breeki-chebureki.md
caligin/actual-cookbook
26bc039f749761417438ff2945b45b88ad6bbc6e
[ "Unlicense" ]
2
2017-12-29T23:45:59.000Z
2022-01-26T17:05:15.000Z
# cheeki breeki chebureki From: [Cheeki Breeki Chebureki (ЧЕБУРЕКИ) - Cooking with Boris](https://youtu.be/x9NEXPoepO4) Not for the faint of heart. Serves: a small army ## ingredients - 250g *pork* mince - 1 garlic clove - 1/2 onion - 2 1cspn salt - 1spn pepper - 1cspn sunflower oil - 1spn dill - 1cspn parsley - 1 egg - 200ml water - 300g flour - 2 shots of vodka or 20 - 1spn sugar - sunflower oil for deep frying ## preparation ### foundation - set aside one shot of vodka, that's for the cheburek shells. the rest is for the chef ### filling - chop the onion coarsely - mince garlic - combine mince, chopped onion, 1cspn salt, pepper, dill, parsley in a bowl - mix with your hands - cover and let sit while you prepare the shells ### shells - combine flour, 1cspn salt, sugar, egg, water, 1cspn oil, 1 shot vodka in a bowl - mix to form dough - knead until it's uniform enough - roll in a fat sausage, cut in 8 parts - use a rolling pin to flatten each piece to a circle/oval shape ### kombine - heat a pan of oil or a deep frier, around 170c - divide the filling in 8 parts - put each portion in the middle of a shell and fold, then use a fork to press on the edges to seal the cheburek - deep fry until golden - serve with some fresh dill (if you have it) and plenty of mayonez ## notes Recipe works fine even if it's not middle of night or not cousin's place. And it works perfectly well with vegetable oil instead of sunflower. There's no "too big" cheburek but having them bigget than the size of your pan doesn't help. Works better with pork mince or a mix that is mainly pork. Actually, don't even try to use beef, it's not great and does not go that well with either the seasonings nor mayonez. If you don't have a rolling pin, you can use a bottle of vodka instead. It also adds that "nailed it" flavour to the dish that you can't really with other means. Just be aware that if the vodka you have is a Русский Стандарт (as I would indeed recommend), its bottle shape just sucks at being a rolling pin: using one of those would make it double hard to prepare the shells and increase the "nailed it" flavours tenfold. Well suited to last long in the freezer and serve as emergency rations. Can easily be cooked in a deep fryer, set around 170c and fry for a few (5?) minutes-ish, until golden.
33.242857
259
0.745595
eng_Latn
0.999173
8796811f3274aa3880cc4d827165448c2957dfb5
4,424
md
Markdown
docs/library/tic80.md
arnoldsecret/docs
085237ce14bc6d31d7a88f94c9a34b9636137e74
[ "MIT" ]
4
2020-11-03T23:16:16.000Z
2021-04-09T22:11:40.000Z
docs/library/tic80.md
arnoldsecret/docs
085237ce14bc6d31d7a88f94c9a34b9636137e74
[ "MIT" ]
null
null
null
docs/library/tic80.md
arnoldsecret/docs
085237ce14bc6d31d7a88f94c9a34b9636137e74
[ "MIT" ]
null
null
null
# TIC-80 ## Background [TIC-80](https://tic.computer) is a fantasy computer for making, playing and sharing tiny games. The TIC-80 core has been authored by - Rob Loach The TIC-80 core is licensed under - [MIT](https://github.com/nesbox/TIC-80/blob/master/LICENSE) A summary of the licenses behind RetroArch and its cores can be found [here](../development/licenses.md). ## Requirements None ## BIOS The TIC-80 core does not feature BIOS use. ## Extensions Content that can be loaded by the TIC-80 core have the following file extensions: - `.tic` (TIC-80 cart) RetroArch database(s) that are associated with the TIC-80 core: - [`TIC-80.rdb`](https://github.com/libretro/libretro-database/blob/master/rdb/TIC-80.rdb) ## Features Frontend-level settings or features that the Theodore core respects. | Feature | Supported | |-------------------|:---------:| | Restart | ✔ | | Saves | ✔ | | States | ✕ | | Rewind | ✕ | | Netplay | [✔](https://tic.computer/play?cart=893) | | Core Options | ✔ | | RetroAchievements | ✕ | | RetroArch Cheats | ✕ | | Native Cheats | [✔](https://github.com/libretro/libretro-database/tree/master/cht/TIC-80) | | Controls | ✔ | | Remapping | ✕ | | Multi-Mouse | ✕ | | Rumble | ✕ | | Sensors | ✕ | | Camera | ✕ | | Location | ✕ | | Subsystem | ✕ | | [Softpatching](../guides/softpatching.md) | ✕ | | Disk Control | ✕ | | Username | ✕ | | Language | ✕ | | Crop Overscan | ✕ | | LEDs | ✕ | ### Directories The TIC-80 core's internal core name is `tic80`. ### Geometry and timing - The TIC-80 core's base width is 240 pixels. - The TIC-80 core's base height is 136 pixels. ## Usage 1. Download a [TIC-80 cart](https://tic.computer/play) - Example: [TIC-80 Mario Bros?](https://tic.computer/play?cart=223) 2. Launch the cart through the TIC-80 core ## Core options The TIC-80 core has the following option(s) that can be tweaked from the core options menu. The default setting is bolded. - **Mouse API instead of Pointer** [tic80_mouse] (**disabled**|enabled) - **Mouse Cursor** [tic80_mouse_cursor] (**disabled**|dot|cross|arrow) ## Controllers The TIC-80 core supports the following device type(s) in the controls menu, bolded device types are the default for the specified user(s): ### User 1 - 4 device types - None - Doesn't disable input. - **RetroPad** - Joypad - Stay on this. - RetroPad w/Analog - Joypad - There's no reason to switch to this. ### Controller tables #### Joypad !!! attention What the buttons do are game specific. | User 1 - 4 Remap descriptors | RetroPad Inputs | |------------------------------|------------------------------------------------| | A | ![](../image/retropad/retro_b.png) | | X | ![](../image/retropad/retro_y.png) | | Select | ![](../image/retropad/retro_select.png) | | D-Pad Down | ![](../image/retropad/retro_dpad_down.png) | | D-Pad Left | ![](../image/retropad/retro_dpad_left.png) | | D-Pad Right | ![](../image/retropad/retro_dpad_right.png) | | B | ![](../image/retropad/retro_a.png) | | Y | ![](../image/retropad/retro_x.png) | | L | ![](../image/retropad/retro_l1.png) | | R | ![](../image/retropad/retro_r1.png) | ## External Links - [TIC-80 Github Repository](https://github.com/nesbox/TIC-80/) - [Report Libretro TIC-80 Core Issues Here](https://github.com/nesbox/TIC-80/issues) - [Libretro TIC-80 Core info file](https://github.com/libretro/libretro-super/blob/master/dist/info/tic80_libretro.info) - [Libretro TIC-80 Database](https://github.com/libretro/libretro-database/blob/master/rdb/TIC-80.rdb) - [Libretro TIC-80 Thumbnails](https://github.com/libretro-thumbnails/TIC-80) - [Libretro TIC-80 Cheats](https://github.com/libretro/libretro-database/tree/master/cht/TIC-80)
35.111111
138
0.555154
eng_Latn
0.648681
8797db554072e61dc66e9e47def2904aa02cd67c
20,712
md
Markdown
repos/openjdk/remote/8-jdk.md
NeilHanlon/repo-info
c513545202b9690e12fe679317a79e0fffa81765
[ "Apache-2.0" ]
null
null
null
repos/openjdk/remote/8-jdk.md
NeilHanlon/repo-info
c513545202b9690e12fe679317a79e0fffa81765
[ "Apache-2.0" ]
null
null
null
repos/openjdk/remote/8-jdk.md
NeilHanlon/repo-info
c513545202b9690e12fe679317a79e0fffa81765
[ "Apache-2.0" ]
1
2022-01-26T17:51:39.000Z
2022-01-26T17:51:39.000Z
## `openjdk:8-jdk` ```console $ docker pull openjdk@sha256:92ea5796fd2e2d34aa650d3649b15915379a468cc10be8199e80e78851582da8 ``` - Manifest MIME: `application/vnd.docker.distribution.manifest.list.v2+json` - Platforms: 4 - linux; amd64 - linux; arm64 variant v8 - windows version 10.0.20348.473; amd64 - windows version 10.0.17763.2458; amd64 ### `openjdk:8-jdk` - linux; amd64 ```console $ docker pull openjdk@sha256:be8d393bf05323d837ba4a4ec5d50eb06f16cdc5ec184921e5d32c72cbc640b7 ``` - Docker Version: 20.10.7 - Manifest MIME: `application/vnd.docker.distribution.manifest.v2+json` - Total Size: **236.9 MB (236928351 bytes)** (compressed transfer size, not on-disk size) - Image ID: `sha256:3bc5f7759e81182b118ab4d74087103d3733483ea37080ed5b6581251d326713` - Default Command: `["bash"]` ```dockerfile # Wed, 26 Jan 2022 01:40:22 GMT ADD file:9cca7f8e4abcd8309501cad216ff33a28932386ded66461a7591c0fdf2c859d3 in / # Wed, 26 Jan 2022 01:40:23 GMT CMD ["bash"] # Wed, 26 Jan 2022 02:11:57 GMT RUN set -eux; apt-get update; apt-get install -y --no-install-recommends ca-certificates curl netbase wget ; rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 02:12:03 GMT RUN set -ex; if ! command -v gpg > /dev/null; then apt-get update; apt-get install -y --no-install-recommends gnupg dirmngr ; rm -rf /var/lib/apt/lists/*; fi # Wed, 26 Jan 2022 02:12:22 GMT RUN apt-get update && apt-get install -y --no-install-recommends git mercurial openssh-client subversion procps && rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 09:24:13 GMT RUN set -eux; apt-get update; apt-get install -y --no-install-recommends bzip2 unzip xz-utils fontconfig libfreetype6 ca-certificates p11-kit ; rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 09:29:49 GMT ENV JAVA_HOME=/usr/local/openjdk-8 # Wed, 26 Jan 2022 09:29:50 GMT RUN { echo '#/bin/sh'; echo 'echo "$JAVA_HOME"'; } > /usr/local/bin/docker-java-home && chmod +x /usr/local/bin/docker-java-home && [ "$JAVA_HOME" = "$(docker-java-home)" ] # backwards compatibility # Wed, 26 Jan 2022 09:29:50 GMT ENV PATH=/usr/local/openjdk-8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin # Wed, 26 Jan 2022 09:29:50 GMT ENV LANG=C.UTF-8 # Wed, 26 Jan 2022 09:29:50 GMT ENV JAVA_VERSION=8u312 # Wed, 26 Jan 2022 09:30:47 GMT RUN set -eux; arch="$(dpkg --print-architecture)"; case "$arch" in 'amd64') downloadUrl='https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_x64_linux_8u312b07.tar.gz'; ;; 'arm64') downloadUrl='https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_aarch64_linux_8u312b07.tar.gz'; ;; *) echo >&2 "error: unsupported architecture: '$arch'"; exit 1 ;; esac; wget --progress=dot:giga -O openjdk.tgz "$downloadUrl"; wget --progress=dot:giga -O openjdk.tgz.asc "$downloadUrl.sign"; export GNUPGHOME="$(mktemp -d)"; gpg --batch --keyserver keyserver.ubuntu.com --recv-keys EAC843EBD3EFDB98CC772FADA5CD6035332FA671; gpg --batch --keyserver keyserver.ubuntu.com --keyserver-options no-self-sigs-only --recv-keys CA5F11C6CE22644D42C6AC4492EF8D39DC13168F; gpg --batch --list-sigs --keyid-format 0xLONG CA5F11C6CE22644D42C6AC4492EF8D39DC13168F | tee /dev/stderr | grep '0xA5CD6035332FA671' | grep 'Andrew Haley'; gpg --batch --verify openjdk.tgz.asc openjdk.tgz; gpgconf --kill all; rm -rf "$GNUPGHOME"; mkdir -p "$JAVA_HOME"; tar --extract --file openjdk.tgz --directory "$JAVA_HOME" --strip-components 1 --no-same-owner ; rm openjdk.tgz*; { echo '#!/usr/bin/env bash'; echo 'set -Eeuo pipefail'; echo 'trust extract --overwrite --format=java-cacerts --filter=ca-anchors --purpose=server-auth "$JAVA_HOME/jre/lib/security/cacerts"'; } > /etc/ca-certificates/update.d/docker-openjdk; chmod +x /etc/ca-certificates/update.d/docker-openjdk; /etc/ca-certificates/update.d/docker-openjdk; find "$JAVA_HOME/lib" -name '*.so' -exec dirname '{}' ';' | sort -u > /etc/ld.so.conf.d/docker-openjdk.conf; ldconfig; javac -version; java -version ``` - Layers: - `sha256:0c6b8ff8c37e92eb1ca65ed8917e818927d5bf318b6f18896049b5d9afc28343` Last Modified: Wed, 26 Jan 2022 01:46:07 GMT Size: 54.9 MB (54917164 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:412caad352a3ecbb29c080379407ae0761e7b9b454f7239cbfd1d1da25e06b29` Last Modified: Wed, 26 Jan 2022 02:22:09 GMT Size: 5.2 MB (5152685 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:e6d3e61f7a504fa66d7275123969e9917570188650eb84b2280a726b996040f6` Last Modified: Wed, 26 Jan 2022 02:22:09 GMT Size: 10.9 MB (10871783 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:461bb1d8c517c7f9fc0f1df66c9dc34c85a23421c1e1c540b2e28cbb258e75f5` Last Modified: Wed, 26 Jan 2022 02:22:32 GMT Size: 54.6 MB (54567492 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:e442ee9d8dd9896a5b3b67117060f2af4ae8e997af7297009e7d0ea4b6595163` Last Modified: Wed, 26 Jan 2022 09:50:58 GMT Size: 5.4 MB (5420014 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:36b5bcea956cbbcc7e16356354d5eb4b5b3834a43d708ca6c3fd048f33ddc579` Last Modified: Wed, 26 Jan 2022 09:55:46 GMT Size: 211.0 B MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:c3879cb950b1f93eff09fbecc9c1d4ed3760d3f81df96b99f8090e3955578b57` Last Modified: Wed, 26 Jan 2022 09:55:59 GMT Size: 106.0 MB (105999002 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip ### `openjdk:8-jdk` - linux; arm64 variant v8 ```console $ docker pull openjdk@sha256:aba441bcd92702429e6bdb076473001e5cbbf15dac870f5c14b6f5840efc1f5a ``` - Docker Version: 20.10.7 - Manifest MIME: `application/vnd.docker.distribution.manifest.v2+json` - Total Size: **234.5 MB (234507132 bytes)** (compressed transfer size, not on-disk size) - Image ID: `sha256:dd85bd25e528125d81189a2ca573db55728c224c1e294e06ace66f789ee37c18` - Default Command: `["bash"]` ```dockerfile # Wed, 26 Jan 2022 01:42:15 GMT ADD file:b6338a9c3c2f8d7ba1a23dcb7a1389c7d0eab75a7489ef66f21c773be56aa9ed in / # Wed, 26 Jan 2022 01:42:16 GMT CMD ["bash"] # Wed, 26 Jan 2022 02:13:03 GMT RUN set -eux; apt-get update; apt-get install -y --no-install-recommends ca-certificates curl netbase wget ; rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 02:13:09 GMT RUN set -ex; if ! command -v gpg > /dev/null; then apt-get update; apt-get install -y --no-install-recommends gnupg dirmngr ; rm -rf /var/lib/apt/lists/*; fi # Wed, 26 Jan 2022 02:13:28 GMT RUN apt-get update && apt-get install -y --no-install-recommends git mercurial openssh-client subversion procps && rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 05:56:06 GMT RUN set -eux; apt-get update; apt-get install -y --no-install-recommends bzip2 unzip xz-utils fontconfig libfreetype6 ca-certificates p11-kit ; rm -rf /var/lib/apt/lists/* # Wed, 26 Jan 2022 06:00:46 GMT ENV JAVA_HOME=/usr/local/openjdk-8 # Wed, 26 Jan 2022 06:00:48 GMT RUN { echo '#/bin/sh'; echo 'echo "$JAVA_HOME"'; } > /usr/local/bin/docker-java-home && chmod +x /usr/local/bin/docker-java-home && [ "$JAVA_HOME" = "$(docker-java-home)" ] # backwards compatibility # Wed, 26 Jan 2022 06:00:48 GMT ENV PATH=/usr/local/openjdk-8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin # Wed, 26 Jan 2022 06:00:49 GMT ENV LANG=C.UTF-8 # Wed, 26 Jan 2022 06:00:50 GMT ENV JAVA_VERSION=8u312 # Wed, 26 Jan 2022 06:01:01 GMT RUN set -eux; arch="$(dpkg --print-architecture)"; case "$arch" in 'amd64') downloadUrl='https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_x64_linux_8u312b07.tar.gz'; ;; 'arm64') downloadUrl='https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_aarch64_linux_8u312b07.tar.gz'; ;; *) echo >&2 "error: unsupported architecture: '$arch'"; exit 1 ;; esac; wget --progress=dot:giga -O openjdk.tgz "$downloadUrl"; wget --progress=dot:giga -O openjdk.tgz.asc "$downloadUrl.sign"; export GNUPGHOME="$(mktemp -d)"; gpg --batch --keyserver keyserver.ubuntu.com --recv-keys EAC843EBD3EFDB98CC772FADA5CD6035332FA671; gpg --batch --keyserver keyserver.ubuntu.com --keyserver-options no-self-sigs-only --recv-keys CA5F11C6CE22644D42C6AC4492EF8D39DC13168F; gpg --batch --list-sigs --keyid-format 0xLONG CA5F11C6CE22644D42C6AC4492EF8D39DC13168F | tee /dev/stderr | grep '0xA5CD6035332FA671' | grep 'Andrew Haley'; gpg --batch --verify openjdk.tgz.asc openjdk.tgz; gpgconf --kill all; rm -rf "$GNUPGHOME"; mkdir -p "$JAVA_HOME"; tar --extract --file openjdk.tgz --directory "$JAVA_HOME" --strip-components 1 --no-same-owner ; rm openjdk.tgz*; { echo '#!/usr/bin/env bash'; echo 'set -Eeuo pipefail'; echo 'trust extract --overwrite --format=java-cacerts --filter=ca-anchors --purpose=server-auth "$JAVA_HOME/jre/lib/security/cacerts"'; } > /etc/ca-certificates/update.d/docker-openjdk; chmod +x /etc/ca-certificates/update.d/docker-openjdk; /etc/ca-certificates/update.d/docker-openjdk; find "$JAVA_HOME/lib" -name '*.so' -exec dirname '{}' ';' | sort -u > /etc/ld.so.conf.d/docker-openjdk.conf; ldconfig; javac -version; java -version ``` - Layers: - `sha256:39ab78bc09e79a21f719ced771689354d1948f4afd57e86afed8dac6ffb47826` Last Modified: Wed, 26 Jan 2022 01:48:34 GMT Size: 53.6 MB (53608848 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:0cf88d09fc28807e643d4f619b2e0c559aaeddc7d7b8176e1144b065d63fa160` Last Modified: Wed, 26 Jan 2022 02:23:47 GMT Size: 5.1 MB (5141567 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:e019dd0368ba992803442a16cc7792c1bd5a3d06c3a1ae6fae17cb838822fb4c` Last Modified: Wed, 26 Jan 2022 02:23:48 GMT Size: 10.7 MB (10655902 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:4b9f1769d50d3693321e1b48f527d9c920830ad24d931b642c116bf7823bed6a` Last Modified: Wed, 26 Jan 2022 02:24:10 GMT Size: 54.7 MB (54669460 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:af5cb8edd8beada37ea034ed391c47dfd106ff9cc9fac852e0f52db0b063374f` Last Modified: Wed, 26 Jan 2022 06:23:05 GMT Size: 5.4 MB (5420184 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:e3973acc73ccb4db4685136c67eb73dd02f02b1518d5381cf376d82710758caa` Last Modified: Wed, 26 Jan 2022 06:28:15 GMT Size: 211.0 B MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:568c3474f69d85e6c2c507e089aacb5f23548845e35e993a3df6640ca80a6a61` Last Modified: Wed, 26 Jan 2022 06:28:29 GMT Size: 105.0 MB (105010960 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip ### `openjdk:8-jdk` - windows version 10.0.20348.473; amd64 ```console $ docker pull openjdk@sha256:c3b70f632b545a4fbe78dcda5a96dcba900036a0839495e19323e4a6650ed229 ``` - Docker Version: 20.10.8 - Manifest MIME: `application/vnd.docker.distribution.manifest.v2+json` - Total Size: **2.3 GB (2310202756 bytes)** (compressed transfer size, not on-disk size) - Image ID: `sha256:ae64bf65c06a870952918d31bf3a46692d575c4abc60f3a022c170f2f089c485` - Default Command: `["c:\\windows\\system32\\cmd.exe"]` - `SHELL`: `["powershell","-Command","$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"]` ```dockerfile # Sat, 08 May 2021 09:40:24 GMT RUN Apply image 2022-RTM-amd64 # Sun, 16 Jan 2022 05:17:24 GMT RUN Install update ltsc2022-amd64 # Wed, 19 Jan 2022 22:21:54 GMT SHELL [powershell -Command $ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';] # Wed, 19 Jan 2022 22:23:11 GMT RUN Write-Host 'Enabling TLS 1.2 (https://githubengineering.com/crypto-removal-notice/) ...'; $tls12RegBase = 'HKLM:\\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\TLS 1.2'; if (Test-Path $tls12RegBase) { throw ('"{0}" already exists!' -f $tls12RegBase) }; New-Item -Path ('{0}/Client' -f $tls12RegBase) -Force; New-Item -Path ('{0}/Server' -f $tls12RegBase) -Force; New-ItemProperty -Path ('{0}/Client' -f $tls12RegBase) -Name 'DisabledByDefault' -PropertyType DWORD -Value 0 -Force; New-ItemProperty -Path ('{0}/Client' -f $tls12RegBase) -Name 'Enabled' -PropertyType DWORD -Value 1 -Force; New-ItemProperty -Path ('{0}/Server' -f $tls12RegBase) -Name 'DisabledByDefault' -PropertyType DWORD -Value 0 -Force; New-ItemProperty -Path ('{0}/Server' -f $tls12RegBase) -Name 'Enabled' -PropertyType DWORD -Value 1 -Force; Write-Host 'Complete.' # Wed, 19 Jan 2022 22:37:53 GMT ENV JAVA_HOME=C:\openjdk-8 # Wed, 19 Jan 2022 22:38:16 GMT RUN $newPath = ('{0}\bin;{1}' -f $env:JAVA_HOME, $env:PATH); Write-Host ('Updating PATH: {0}' -f $newPath); setx /M PATH $newPath; Write-Host 'Complete.' # Wed, 19 Jan 2022 22:38:17 GMT ENV JAVA_VERSION=8u312 # Wed, 19 Jan 2022 22:38:18 GMT ENV JAVA_URL=https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_x64_windows_8u312b07.zip # Wed, 19 Jan 2022 22:39:13 GMT RUN Write-Host ('Downloading {0} ...' -f $env:JAVA_URL); [Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12; Invoke-WebRequest -Uri $env:JAVA_URL -OutFile 'openjdk.zip'; Write-Host 'Expanding ...'; New-Item -ItemType Directory -Path C:\temp | Out-Null; Expand-Archive openjdk.zip -DestinationPath C:\temp; Move-Item -Path C:\temp\* -Destination $env:JAVA_HOME; Remove-Item C:\temp; Write-Host 'Removing ...'; Remove-Item openjdk.zip -Force; Write-Host 'Verifying install ...'; Write-Host ' javac -version'; javac -version; Write-Host ' java -version'; java -version; Write-Host 'Complete.' ``` - Layers: - `sha256:8f616e6e9eec767c425fd9346648807d1b658d20ff6097be1d955aac69c26642` Size: 1.3 GB (1251699055 bytes) MIME: application/vnd.docker.image.rootfs.foreign.diff.tar.gzip - `sha256:0e02c12b1310e6c76c29fcd6f81905400fdb6a01caac9dc825939ad004baffef` Size: 955.8 MB (955800778 bytes) MIME: application/vnd.docker.image.rootfs.foreign.diff.tar.gzip - `sha256:2e1f45a873642f0aab3474828d75469362741244e7c714068ac1afe056102cd6` Last Modified: Thu, 20 Jan 2022 00:40:19 GMT Size: 1.4 KB (1424 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:fa436893917b4b24f7020ae30c97fa33ab983d9c24e6fd7534d56f8b19bf786f` Last Modified: Thu, 20 Jan 2022 02:19:54 GMT Size: 601.2 KB (601200 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:1d4e5cb5c1db0f54a5490ae158422cc0b99b83174adf1eea6454b0543c595c5c` Last Modified: Thu, 20 Jan 2022 02:39:49 GMT Size: 1.4 KB (1404 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:83df987ecf41597ac882d38a9c8f6daa4aa6503ebf73f388dc77a6ae9db30c72` Last Modified: Thu, 20 Jan 2022 02:39:50 GMT Size: 506.6 KB (506564 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:dffeb7c5e533c5f3370ac5fa503f0b0e46c5a96ec022983bd774f729f5e3cef8` Last Modified: Thu, 20 Jan 2022 02:39:49 GMT Size: 1.4 KB (1426 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:6c832107834b0f366877adcceacc064dd1123d4b9674717dc3667318501e3f74` Last Modified: Thu, 20 Jan 2022 02:39:49 GMT Size: 1.4 KB (1440 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:a71fa2f8f8435d2a9e6c33c02d9cb7d6a72da0f0663f4c5f7142948651f3f1f6` Last Modified: Thu, 20 Jan 2022 02:40:02 GMT Size: 101.6 MB (101589465 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip ### `openjdk:8-jdk` - windows version 10.0.17763.2458; amd64 ```console $ docker pull openjdk@sha256:17414fc0766b06d69d50f2512026ce02de88524c615987d96389bd4da26255e3 ``` - Docker Version: 20.10.8 - Manifest MIME: `application/vnd.docker.distribution.manifest.v2+json` - Total Size: **2.8 GB (2815394065 bytes)** (compressed transfer size, not on-disk size) - Image ID: `sha256:95a24e496fb2d5f6baa382df617872bafb1e6f0e011183b847f34fb2a7510dc2` - Default Command: `["c:\\windows\\system32\\cmd.exe"]` - `SHELL`: `["powershell","-Command","$ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';"]` ```dockerfile # Thu, 07 May 2020 05:09:25 GMT RUN Apply image 1809-RTM-amd64 # Tue, 18 Jan 2022 01:52:23 GMT RUN Install update 1809-amd64 # Wed, 19 Jan 2022 23:21:48 GMT SHELL [powershell -Command $ErrorActionPreference = 'Stop'; $ProgressPreference = 'SilentlyContinue';] # Wed, 19 Jan 2022 23:23:41 GMT RUN Write-Host 'Enabling TLS 1.2 (https://githubengineering.com/crypto-removal-notice/) ...'; $tls12RegBase = 'HKLM:\\SYSTEM\CurrentControlSet\Control\SecurityProviders\SCHANNEL\Protocols\TLS 1.2'; if (Test-Path $tls12RegBase) { throw ('"{0}" already exists!' -f $tls12RegBase) }; New-Item -Path ('{0}/Client' -f $tls12RegBase) -Force; New-Item -Path ('{0}/Server' -f $tls12RegBase) -Force; New-ItemProperty -Path ('{0}/Client' -f $tls12RegBase) -Name 'DisabledByDefault' -PropertyType DWORD -Value 0 -Force; New-ItemProperty -Path ('{0}/Client' -f $tls12RegBase) -Name 'Enabled' -PropertyType DWORD -Value 1 -Force; New-ItemProperty -Path ('{0}/Server' -f $tls12RegBase) -Name 'DisabledByDefault' -PropertyType DWORD -Value 0 -Force; New-ItemProperty -Path ('{0}/Server' -f $tls12RegBase) -Name 'Enabled' -PropertyType DWORD -Value 1 -Force; Write-Host 'Complete.' # Wed, 19 Jan 2022 23:39:21 GMT ENV JAVA_HOME=C:\openjdk-8 # Wed, 19 Jan 2022 23:40:22 GMT RUN $newPath = ('{0}\bin;{1}' -f $env:JAVA_HOME, $env:PATH); Write-Host ('Updating PATH: {0}' -f $newPath); setx /M PATH $newPath; Write-Host 'Complete.' # Wed, 19 Jan 2022 23:40:23 GMT ENV JAVA_VERSION=8u312 # Wed, 19 Jan 2022 23:40:25 GMT ENV JAVA_URL=https://github.com/AdoptOpenJDK/openjdk8-upstream-binaries/releases/download/jdk8u312-b07/OpenJDK8U-jdk_x64_windows_8u312b07.zip # Wed, 19 Jan 2022 23:42:01 GMT RUN Write-Host ('Downloading {0} ...' -f $env:JAVA_URL); [Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12; Invoke-WebRequest -Uri $env:JAVA_URL -OutFile 'openjdk.zip'; Write-Host 'Expanding ...'; New-Item -ItemType Directory -Path C:\temp | Out-Null; Expand-Archive openjdk.zip -DestinationPath C:\temp; Move-Item -Path C:\temp\* -Destination $env:JAVA_HOME; Remove-Item C:\temp; Write-Host 'Removing ...'; Remove-Item openjdk.zip -Force; Write-Host 'Verifying install ...'; Write-Host ' javac -version'; javac -version; Write-Host ' java -version'; java -version; Write-Host 'Complete.' ``` - Layers: - `sha256:4612f6d0b889cad0ed0292fae3a0b0c8a9e49aff6dea8eb049b2386d9b07986f` Size: 1.7 GB (1718332879 bytes) MIME: application/vnd.docker.image.rootfs.foreign.diff.tar.gzip - `sha256:567fd00846e9a9f44eea5925b497356dda00fe89b8335d2a3b2a8b9d84b0bb69` Size: 995.0 MB (994988595 bytes) MIME: application/vnd.docker.image.rootfs.foreign.diff.tar.gzip - `sha256:ba7b1090ce9545f6aac90d390785f6c5e3846304cb7596289dfc36c169d7b1da` Last Modified: Thu, 20 Jan 2022 00:40:41 GMT Size: 1.4 KB (1428 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:a98c73c8d54eeae51e9c93499966b4413ea63f218bdf175bed7b7d83c76e6d58` Last Modified: Thu, 20 Jan 2022 02:20:37 GMT Size: 357.3 KB (357347 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:246842bd6c122f1b034197679204f1fb3ddbd4f74cd7385ae12e05f57df8e0bd` Last Modified: Thu, 20 Jan 2022 02:40:17 GMT Size: 1.4 KB (1402 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:b49f2f0163e3bd31b4712f6d8d88a90630e5c1e88dda4c13eb5469c995df3370` Last Modified: Thu, 20 Jan 2022 02:40:18 GMT Size: 310.7 KB (310700 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:af2612e59aaf2cae73a8d0898d51a46f2802c664336faeb0666e93558608da11` Last Modified: Thu, 20 Jan 2022 02:40:17 GMT Size: 1.4 KB (1417 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:fb919c8a5d29718072bf68785f82bacdb5ca0c5fad59c6c775fee9f928010afe` Last Modified: Thu, 20 Jan 2022 02:40:17 GMT Size: 1.4 KB (1384 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip - `sha256:04cf9ec7167e3d2709427ad944f5dbbfb765c3b30b72509bb57241f1f8ed7eb2` Last Modified: Thu, 20 Jan 2022 02:42:08 GMT Size: 101.4 MB (101398913 bytes) MIME: application/vnd.docker.image.rootfs.diff.tar.gzip
70.210169
1,792
0.737978
yue_Hant
0.406966
8798573a1339568170b0377d366172f34f23019a
10,273
md
Markdown
docs/calliope/readme.md
JackGamesFTW/framework
d3856c4d4468bd74398abbf5f05e3cf71b25e2c4
[ "MIT" ]
null
null
null
docs/calliope/readme.md
JackGamesFTW/framework
d3856c4d4468bd74398abbf5f05e3cf71b25e2c4
[ "MIT" ]
null
null
null
docs/calliope/readme.md
JackGamesFTW/framework
d3856c4d4468bd74398abbf5f05e3cf71b25e2c4
[ "MIT" ]
null
null
null
# Model The model is at the hearth of this package. It boasts a lot of features, so they have been broken down into the following sections: - [Attributes](./attributes.md) - [Api calls](./api-calls.md) - [Query Building](./query-building.md) - [Relationships](./relationships.md) - [Timestamps](./timestamps.md) - [Additional Methods](#additional-methods) ## Creating Models To create a model, you should first define your model class and define the [getName](#getname) method: <code-group> <code-block title="Javascript"> ```js // User.js import { Model } from '@upfrontjs/framework'; export default class User extends Model { getName() { return 'User'; } } ``` </code-block> <code-block title="Typescript"> ```ts // User.ts import { Model } from '@upfrontjs/framework'; export default class User extends Model { public override getName(): string { return 'User'; } } ``` </code-block> </code-group> Then you can call your model in various way, for example ```js // myScript.js import User from '@Models/User'; User.find(1); // or User.create({ my: attributes }); // etc... ``` **If you're passing attributes to the model you have to use the [create](#create) method.** ::: tip TIP (Typescript) Typescript users may benefit from better typing support if they defined keys and their types on the models ```ts export default class User extends Model { public is_admin?: boolean; public age?: number; public name?: string; public override getName(): string { return 'User'; } } ``` This will typehint keys on the model when accessing the above keys like `user.age` and will get type hinted in various methods such as [getAttribute](./attributes.md#getattribute) where both the key, the default value and the return value will be type hinted. ::: ## Getters #### primaryKey The `primaryKey` is a getter of the attribute name which is used to identify your model. The default value is `'id'`. ```js // User.js import { Model } from '@upfrontjs/framework'; export default class User extends Model { getName() { return 'User'; } get primaryKey() { return 'id'; } } ``` #### keyType The `keyType` is a getter that identifies what the type is of your [primaryKey](#primarykey). Its value has to be either `'string'` or `'number'` with `'number'` being the default value. You should update this value to `'string'` if you're using a uuid or some custom string for the primary key as this is used when in the [Factory](../testing.md#factories) and [exists](#exists) logic. ```js // User.ts import { Model } from '@upfrontjs/framework'; export default class User extends Model { public override getName(): string { return 'User'; } public get primaryKey(): 'string' { return 'string'; } } ``` #### exists The `exists` property is a getter on the model that returns a boolean which can be used to assert that the model has been persisted. It takes the [primary key](#getkey), [timestamps](./timestamps.md#timestamps) and [soft deletes](./timestamps.md#soft-deletes) into account. ## Additional methods #### is The `is` method compares the given model with the current model based on the [getKey](#getkey) and [getName](#getname) ```js import User from '@Models/User'; import Shift from '@Models/Shift'; const user = User.create({ id: 1 }); const user2 = User.create({ id: 2 }); const shift = Shift.create({ id: 1 }); user.is(user); // true user.is(user2); // false user.is(shift); // false ``` #### isNot The `isNot` method is the inverse of the [is](#is) method. #### getKey The `getKey` method returns the value of the primary key from the model. #### getKeyName The `getKeyName` method returns the [primaryKey](#primarykey) of the model. #### getName The `getName` method expected to return the current class' name. For example in a class called `User` it should return `'User'`. ***Note: This is essential to add to every model as this is used throughout the framework.*** ::: danger This value cannot be `this.constructor.name` **if** you're minifying your code or in the minification options you haven't turned off the class rename or equivalent option. ::: ::: tip If you turn of the class renaming when the code gets mangled by the minifier or bundler of your choice, `this.constructor.name` would be an acceptable solution. This would allow you to have a base model you can extend from which can in turn implement the `getName` that returns `this.constructor.name`. Bundlers/minifier options examples: - [terser: keep_classnames](https://terser.org/docs/api-reference#minify-options) - [esbuild: keep-names](https://esbuild.github.io/api/#keep-names) - [babel-minify: keepClassName](https://babeljs.io/docs/en/babel-minify#node-api) ::: #### create <Badge text="static" type="warning"/> The `create` method instantiates your model while setting up attributes and relations. This will mass assign attributes to the model while respecting the [guarding](./attributes#guarding) settings. ```ts import User from '@Models/User'; const user = User.create({ name: 'User Name' }); // User ``` ::: warning Constructing a new class like `new User({...})` is **not** acceptable. This will not overwrite your class fields with default values if the same key has been passed in due to how JavaScript first constructs the parent class and only then the subclasses. However, you can still use it to call instance methods. Furthermore, it will not cause unexpected results if using it with the [setAttribute](./attributes.md#setattribute) method or call methods that under the hood uses the [setAttribute](./attributes.md#setattribute). ::: ::: tip When creating an instance and passing in another instance of the model: ```js import User from '@Models/User'; import Shift from '@Models/Shift'; const user = User.create({ name: 'John Doe' }); const newUser = User.create(user); ``` It will clone the [raw attributes](./attributes#getrawattributes) and the [relationships](./relationships.md#getrelations) of the model. ::: #### replicate The `replicate` method copies the instance into a non-existent instance. Meaning primary key and the timestamps won't be copied. ```js import User from '@Models/User'; const user = User.factory().create(); user.getKey(); // 1 user.name; // 'the name' user.getAttribute(user.getCreatedAtName()); // Date instance const userCopy = user.replicate(); userCopy.getKey(); // undefined userCopy.name; // 'the name' userCopy.getAttribute(userCopy.getCreatedAtName()); // undefined ``` #### clone The `clone` method clones the instance in its current state. Meaning all changes to the [query building](./query-building.md), [endpoint](./api-calls.md#endpoint-manipulation) and [attribute changes](./attributes.md#tracking-changes) will be copied along. The result will match the original model but nothing is copied by reference. ```js import User from '@Models/User'; const user = User.factory().create({ myKey: 1 }); const userClone = user.clone(); user.is(userClone); // true user.myKey = 2; userClone.myKey === 1; // true ``` #### factory <Badge text="static" type="warning"/> The `factory` is a method that returns a [Factory](../testing.md#factorybuilder) instance. Optionally it takes a number argument which is a shorthand for the [times](../testing.md#times) method. ```js import User from '@Models/User'; const user = User.factory().create(); // User const users = User.factory(2).create(); // ModelCollection ``` #### all <Badge text="static" type="warning"/><Badge text="async" type="warning"/> The `all` method will initiate a request that returns a [ModelCollection](./model-collection.md) from the underlying [get](./api-calls.md#get) method. ```js import User from '@Models/User'; const users = await User.all(); // ModelCollection[User, ...] ``` #### save <Badge text="async" type="warning"/> The `save` method will update or save your model based on whether the model [exists](#exists) or not. If the model exists it will send a `PATCH` request containing the changes, and the optionally passed in attributes. If the model does not exists it will send a `POST` request. The method returns the same current user updated with the response data if any. #### update <Badge text="async" type="warning"/> The `update` method sets the correct endpoint then initiates a [patch](./api-calls.md#patch) request. If the model does not [exists](#exists) it will throw an error. ```js import User from '@Models/User'; const user = User.factory.createOne(); await user.update({ optionalExtra: 'data' }); ``` #### find <Badge text="async" type="warning"/> The `find` method sends a `GET` request to the model [endpoint](./api-calls.md#getendpoint) supplemented with the given id. Available both static and non-statically. ```js import User from '@Models/User'; const user = await User.find('8934d792-4e4d-42a1-bb4b-45b34b1140b4'); ``` #### findMany <Badge text="async" type="warning"/> The `findMany` method similar to the [find](#find) method sends a `GET` request to the model [endpoint](./api-calls.md#getendpoint) but adds a [whereKey](./query-building.md#wherekey) constraint to the request, returning a [ModelCollection](./model-collection.md). Available both static and non-statically. ```js import User from '@Models/User'; const users = await User.findMany([1, 2]); // ModelCollection[User, User] ``` #### refresh <Badge text="async" type="warning"/> The `refresh` method updates all the attributes on the model by [selecting](./query-building.md#select) the present [attribute keys](./attributes.md#getattributekeys) and setting the attributes from the response. This will reset any attribute [changes](./attributes.md#getchanges). ```js import User from '@Models/User'; const user = await User.find(1); user.name = 'new name'; user.getChanges(); // { name: 'new name' } await user.refresh(); user.getChanges(); // {} ```
33.571895
524
0.691132
eng_Latn
0.975709
8798cae00b39c58fc722d741c60efd2bd181d5fa
2,289
md
Markdown
_posts/2018-11-21-Download-the-utility-of-force-the-art-of-war-in-the-modern-world-1st-vintage-books-edition.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
_posts/2018-11-21-Download-the-utility-of-force-the-art-of-war-in-the-modern-world-1st-vintage-books-edition.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
_posts/2018-11-21-Download-the-utility-of-force-the-art-of-war-in-the-modern-world-1st-vintage-books-edition.md
Jobby-Kjhy/27
ea48bae2a083b6de2c3f665443f18b1c8f241440
[ "MIT" ]
null
null
null
--- layout: post comments: true categories: Other --- ## Download The utility of force the art of war in the modern world 1st vintage books edition book From it something rose, purchase of Japanese, or not far away, did not even have an axis, and in fact the North-east Passage, he'd Again Nolan looked down at the girl who lay curled beside him on the bed. But then he saw how I fixtures. " The ball of sodden Kleenex was gripped so tightly in Junior's left hand that had its carbon content been higher, one of them would be president by now. The wise man and wise woman, Agnes noticed that they grew less naive, standing mountain. " "Yes, the enough. After a short pause she said, Islands in the Siberian Sea! If books could be brought together in one place. There'll be a next time? Without asking a thing, drew connecting lines through her constellations of coppery freckles, God hath it in His power to change a case from foul to fair, long after the execution of Josef Krepp, Mom and Dad and daughter think that she possessed the fortitude to endure the suffering of her innocent 253 discover that these behemoths were hosting a World Wrestling Federation beer party in his bungalow, and various other fluids, he thought, hung by the antlers jand the legs dangling down, too. LESLIE. "Disappointing, 1867 (Mittheil, Barry could feel the middle of his body turning outrageous. 7 of triumph over evil, Mr, and he the driver, the keystone of her soul. Spent the afternoon in a bookstore. " preternatural hush reputed to precede the biggest quakes. Dinosaur-loud, I didn't," he said, your-head not clean, he hath entered my house and lain down on my bed and I fear lest there be an intrigue between him and the woman? The murmur of their voices and their gentle laughter drifts back to him, doubtless in consequence of Martiniere's easy style. deg. " killed by the lance or knife, sweating spectators, inset with faceted old Clara, and whoso reareth the young of the serpent shall get of them nought but biting, to her the utility of force the art of war in the modern world 1st vintage books edition destination, Barry could feel the middle the utility of force the art of war in the modern world 1st vintage books edition his body turning outrageous. They were both grinning. Stairs?
254.333333
2,134
0.78768
eng_Latn
0.999841
8798f89d0eafaf4b9a662115c7c5df24fbe58e18
3,175
md
Markdown
overview.md
UrbanInstitute/git-tutorial
d2c577750ac5dd28d82c2daeff3438387678c84b
[ "MIT" ]
7
2015-01-22T07:30:20.000Z
2020-02-19T19:14:45.000Z
overview.md
UrbanInstitute/git-tutorial
d2c577750ac5dd28d82c2daeff3438387678c84b
[ "MIT" ]
1
2015-01-29T18:57:52.000Z
2015-01-29T20:06:36.000Z
overview.md
UrbanInstitute/git-tutorial
d2c577750ac5dd28d82c2daeff3438387678c84b
[ "MIT" ]
5
2015-01-29T18:24:06.000Z
2020-04-11T13:54:35.000Z
# An Overview ## What is git? Git is a piece of software that helps organize and track changes to files over time (usually text based files, like code), and helps multiple users work on the same projects, without stepping on each others' toes. Git is run from the command line (by default the [Terminal](http://en.wikipedia.org/wiki/Terminal_%28OS_X%29) on Macs and the [Command Prompt](http://en.wikipedia.org/wiki/Cmd.exe) on PCs), usually by running `git`, followed by another key word, then often followed by various options. ## What is github? Github is a website (you're on it now!) that provides a visual interface for git, as well as hosting the files that make up a project. *One* version of these files is hosted on github servers (called the *remote* version), but in almost every case there are many other versions on various other computers (called *clones*). For example, on my computer at work I have a version of these files, and another version on my computer at home. When you follow these tutorials, you'll have your own version(s). As we work on the files, at various times we will sync them up with the remote version that is on github, by **pushing** the changes we make to github and **pulling** down the changes other users make to the remote version. I go over a few nice github features in [another document](working.md), but most of what we'll talk about is specific to git, not github. Almost all of the functions and commands we talk about are still valid when using services such as [bitbucket](https://bitbucket.org/) or [gitlab](https://about.gitlab.com/). Many [open-source software projects](http://en.wikipedia.org/wiki/Free_and_open-source_software) are hosted on github (or bitbucket and other git hosting services). Personally, I am a passionate proponent for open-source software, working in the open, and the open data movement. I'd be happy to chat more about the whys and the hows of open-source development, and [many](https://github.com/18F/open-source-policy/blob/master/policy.md) [other](http://www.gnu.org/gnu/manifesto.html) [folks](https://okfn.org/opendata/why-open-data/) have written more eloquently and thorougly than I ever could. Most software developers write their code using git, sometimes in public, sometimes in private (charging for private hosting is one way github makes money). The more you work with git, the more invaluable it will become to you. ## Overview In the [main tutorial](working.md), there's a ton of information and steps to follow. Here, I'll just give a quick overview of the fundamental git workflow. Let's say you're working on some files, on your local clone, and want to sync up with the remote version on github: ### git add Flag certain file(s) as ready to be synced ### git commit Lump all your changes together, and write a message describing the changes ### git pull Retreive changes other users have sent to the remote repo ### git push Send your changes to the remote repo There's a lot of terminology, you'll get used to it! I also put together a [glossary](glossary.md) of the terms in this tutorial, which you can refer back to. ### NEXT: [Getting started](setup.md)
93.382353
726
0.767244
eng_Latn
0.999205
8799623123d088693f97da72475b9648fbcf6aba
7,023
md
Markdown
deps/oauth2/README.md
slaily/discuss
6f0eacd0f2c03d197f2cf9b6c27a03752c90e969
[ "MIT" ]
null
null
null
deps/oauth2/README.md
slaily/discuss
6f0eacd0f2c03d197f2cf9b6c27a03752c90e969
[ "MIT" ]
2
2018-09-07T04:52:56.000Z
2018-09-07T04:53:07.000Z
deps/oauth2/README.md
slaily/discuss
6f0eacd0f2c03d197f2cf9b6c27a03752c90e969
[ "MIT" ]
null
null
null
# OAuth2 (Client) > An Elixir OAuth2 Client [![Build Status](https://travis-ci.org/scrogson/oauth2.svg?branch=master)](https://travis-ci.org/scrogson/oauth2) [![Coverage Status](https://coveralls.io/repos/scrogson/oauth2/badge.svg?branch=master&service=github)](https://coveralls.io/github/scrogson/oauth2?branch=master) ## Install ```elixir # mix.exs def application do # Add the application to your list of applications. # This will ensure that it will be included in a release. [applications: [:logger, :oauth2]] end defp deps do # Add the dependency [{:oauth2, "~> 0.9"}] end ``` ## Configure a serializer This library can be configured to handle encoding and decoding requests and responses automatically. If you're using [Poison](https://hex.pm/packages/poison) for JSON in your application, this library is already pre-configured to use it for `"application/json"` request and response bodies. You will still need to include it as a dependency though. If you need to handle different MIME types, you can simply configure it like so: ```elixir # config/config.exs config :oauth2, serializers: %{ "application/vnd.api+json" => Poison, "application/xml" => MyApp.XmlParser, } ``` The `serializers` option is a map where the keys are MIME types and the values are modules. The modules are expected to export `encode!/1` and `decode!/1`. ```elixir defmodule MyApp.XmlParser do def encode!(data), do: # ... def decode!(binary), do: # ... end ``` ## Debug mode Some times its handy to see what's coming back from the response when getting a token. You can configure OAuth2 to output the response like so: ```elixir config :oauth2, debug: true ``` ## Usage Current implemented strategies: - Authorization Code - Password - Client Credentials ### Authorization Code Flow (AuthCode Strategy) ```elixir # Initialize a client with client_id, client_secret, site, and redirect_uri. # The strategy option is optional as it defaults to `OAuth2.Strategy.AuthCode`. client = OAuth2.Client.new([ strategy: OAuth2.Strategy.AuthCode, #default client_id: "client_id", client_secret: "abc123", site: "https://auth.example.com", redirect_uri: "https://example.com/auth/callback" ]) # Generate the authorization URL and redirect the user to the provider. OAuth2.Client.authorize_url!(client) # => "https://auth.example.com/oauth/authorize?client_id=client_id&redirect_uri=https%3A%2F%2Fexample.com%2Fauth%2Fcallback&response_type=code" # Use the authorization code returned from the provider to obtain an access token. client = OAuth2.Client.get_token!(client, code: "someauthcode") # Use the access token to make a request for resources resource = OAuth2.Client.get!(client, "/api/resource").body ``` ### Client Credentials Flow Getting an initial access token: ```elixir # Initializing a client with the strategy `OAuth2.Strategy.ClientCredentials` client = OAuth2.Client.new([ strategy: OAuth2.Strategy.ClientCredentials, client_id: "client_id", client_secret: "abc123", site: "https://auth.example.com" ]) # Request a token from with the newly created client # Token will be stored inside the `%OAuth2.Client{}` struct (client.token) client = OAuth2.Client.get_token!(client) # client.token contains the `%OAuth2.AccessToken{}` struct # raw access token access_token = client.token.access_token ``` Refreshing an access token: ```elixir # raw refresh token - use a client with `OAuth2.Strategy.Refresh` for refreshing the token refresh_token = client.token.refresh_token refresh_client = OAuth2.Client.new([ strategy: OAuth2.Strategy.Refresh, client_id: "client_id", client_secret: "abc123", site: "https://auth.example.com", params: %{"refresh_token" => refresh_token} ]) # refresh_client.token contains the `%OAuth2.AccessToken{}` struct again refresh_client = OAuth2.Client.get_token!(refresh_client) ``` ## Write Your Own Strategy Here's an example strategy for GitHub: ```elixir defmodule GitHub do use OAuth2.Strategy # Public API def client do OAuth2.Client.new([ strategy: __MODULE__, client_id: System.get_env("GITHUB_CLIENT_ID"), client_secret: System.get_env("GITHUB_CLIENT_SECRET"), redirect_uri: "http://myapp.com/auth/callback", site: "https://api.github.com", authorize_url: "https://github.com/login/oauth/authorize", token_url: "https://github.com/login/oauth/access_token" ]) end def authorize_url! do OAuth2.Client.authorize_url!(client(), scope: "user,public_repo") end # you can pass options to the underlying http library via `opts` parameter def get_token!(params \\ [], headers \\ [], opts \\ []) do OAuth2.Client.get_token!(client(), params, headers, opts) end # Strategy Callbacks def authorize_url(client, params) do OAuth2.Strategy.AuthCode.authorize_url(client, params) end def get_token(client, params, headers) do client |> put_param(:client_secret, client.client_secret) |> put_header("accept", "application/json") |> OAuth2.Strategy.AuthCode.get_token(params, headers) end end ``` Here's how you'd use the example GitHub strategy: Generate the authorize URL and redirect the client for authorization. ```elixir GitHub.authorize_url! ``` Capture the `code` in your callback route on your server and use it to obtain an access token. ```elixir client = GitHub.get_token!(code: code) ``` Use the access token to access desired resources. ```elixir user = OAuth2.Client.get!(client, "/user").body # Or case OAuth2.Client.get(client, "/user") do {:ok, %OAuth2.Response{body: user}} -> user {:error, %OAuth2.Response{status_code: 401, body: body}} -> Logger.error("Unauthorized token") {:error, %OAuth2.Error{reason: reason}} -> Logger.error("Error: #{inspect reason}") end ``` ## Examples - [Authenticate with Github (OAuth2/Phoenix)](https://github.com/scrogson/oauth2_example) ## License The MIT License (MIT) Copyright (c) 2015 Sonny Scroggin Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
28.782787
162
0.740282
eng_Latn
0.736978
8799694cf413f4daa5840cafbc5775d75e09a572
4,802
md
Markdown
docs/type-generation.md
darrylyoung/nexus
f78dc7b1b7122727030ad3f8f21806817e4a1480
[ "MIT" ]
null
null
null
docs/type-generation.md
darrylyoung/nexus
f78dc7b1b7122727030ad3f8f21806817e4a1480
[ "MIT" ]
null
null
null
docs/type-generation.md
darrylyoung/nexus
f78dc7b1b7122727030ad3f8f21806817e4a1480
[ "MIT" ]
null
null
null
--- id: type-generation title: Type Generation Details sidebar_label: Type Generation Details --- This is the most important piece to understand to get the most out of Nexus. It is relevant to JavaScript as well as TypeScript users, as tools like VSCode and `// @ts-check` can utilize these types to aid in autocomplete or type-checking. A core goal of Nexus is to have the best possible type coverage with the least possible manual type annotation. ## Overview Nexus was designed with TypeScript in mind. In order to fully typecheck our GraphQL objects, we need to generate a number of types that combine the schema, any type or field configuration provided, and the GraphQL resolution algorithm to create as much type-safety as possible without any additional work importing and assigning types throughout the codebase. ## Root Types A **root type** is a type representation of the value used to resolve the fields of an object type. It as the object that will be passed as the first argument of `resolve`. It can be a plain JS object, a database model, a mongoose document, a JS class, anything that fulfills the contract defined by the GraphQL object type, based on the field definitions. Scalars can also have backing types, representing the value they are parsed into. Sometimes GraphQL types are passthrough, and don't have a dedicated type backing them. One such case would be in the `Edge` of a Relay style pagination. In this case, Nexus will generate a type-definition which makes assumptions of the necessary value to fulfill the contract. If this is incorrect, you can always provide a concrete type for the object. ## Field Type A **field type** is the valid return value used to a field on an object type. In GraphQL, promises can be returned at every level of the type resolution, so we wrap the types in a `MaybePromiseDeep<T>` type to express this. ## Configuring our types The [Ghost Example](https://github.com/prisma/nexus/blob/develop/examples/ghost/src/ghost-schema.ts) is the best to look at for an example of how we're able to capture the types from existing runtime objects or definitions and merge them with our schema. The [makeSchema](api-makeSchema.md) takes an option `typegenAutoConfig` which helps us find the types we need to import into our generated schema. ```ts export interface TypegenAutoConfigOptions { /** * Any headers to prefix on the generated type file */ headers?: string[]; /** * Array of files to match for a type * * sources: [ * { source: 'typescript', alias: 'ts' }, * { source: path.join(__dirname, '../backingTypes'), alias: 'b' }, * ] */ sources: TypegenConfigSourceModule[]; /** * Typing for the context, referencing a type defined in the aliased module * provided in sources e.g. 'alias.Context' */ contextType?: string; /** * Types that should not be matched for a backing type, * * By default this is set to ['Query', 'Mutation', 'Subscription'] * * skipTypes: ['Query', 'Mutation', /(.*?)Edge/, /(.*?)Connection/] */ skipTypes?: (string | RegExp)[]; /** * If debug is set to true, this will log out info about all types * found, skipped, etc. for the type generation files. */ debug?: boolean; /** * If provided this will be used for the backing types rather than the auto-resolve * mechanism above. Useful as an override for one-off cases, or for scalar * backing types. */ backingTypeMap?: Record<string, string>; } export interface TypegenConfigSourceModule { /** * The module for where to look for the types. * This uses the node resolution algorithm via require.resolve, * so if this lives in node_modules, you can just provide the module name * otherwise you should provide the absolute path to the file. */ source: string; /** * When we import the module, we use 'import * as ____' to prevent * conflicts. This alias should be a name that doesn't conflict with any other * types, usually a short lowercase name. */ alias: string; /** * Provides a custom approach to matching for the type * * If not provided, the default implementation is: * * (type) => new RegExp('(?:interface|type|class)\s+(${type.name})\W') * */ typeMatch?: ( type: GraphQLNamedType, defaultRegex: RegExp ) => RegExp | RegExp[]; /** * A list of typesNames or regular expressions matching type names * that should be resolved by this import. Provide an empty array if you * wish to use the file for context and ensure no other types are matched. */ onlyTypes?: (string | RegExp)[]; /** * By default the import is configured 'import * as alias from', setting glob to false * will change this to 'import alias from' */ glob?: false; } ```
43.261261
359
0.714494
eng_Latn
0.997652
87996d4ffe0eab67b215bfbe4d87239c1ecdaaf8
900
md
Markdown
powerapps-docs/developer/model-driven-apps/clientapi/reference/events/presearch.md
noriji/powerapps-docs.ja-jp
0dc46ef4c9da7a8ab6d6e2a397f271170f7b9970
[ "CC-BY-4.0", "MIT" ]
null
null
null
powerapps-docs/developer/model-driven-apps/clientapi/reference/events/presearch.md
noriji/powerapps-docs.ja-jp
0dc46ef4c9da7a8ab6d6e2a397f271170f7b9970
[ "CC-BY-4.0", "MIT" ]
null
null
null
powerapps-docs/developer/model-driven-apps/clientapi/reference/events/presearch.md
noriji/powerapps-docs.ja-jp
0dc46ef4c9da7a8ab6d6e2a397f271170f7b9970
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: モデル駆動型アプリにおける 検索コントロール PreSearch イベント (Client API リファレンス) | Microsoft Docs ms.date: 10/31/2018 ms.service: crm-online ms.topic: reference applies_to: Dynamics 365 (online) ms.assetid: 89123cde-7c66-4c7d-94e4-e287285019f8 author: KumarVivek ms.author: kvivek manager: amyla search.audienceType: - developer search.app: - PowerApps - D365CE --- # <a name="lookup-control-presearch-event-client-api-reference"></a>検索コントロール PreSearch イベント (クライアント API 参照) このイベントは、レコードを検索するのに検索コントロールがダイアログを起動する直前に発生します。 このイベントに対するイベント ハンドラーに設定する UI はありません。 検索コントロールで [addPreSearch](../controls/addpresearch.md) と [removePreSearch](../controls/removepresearch.md) メソッドを使用して、このイベントのイベント ハンドラーの追加と削除をする必要があります。 このイベントを他の検索コントロール メソッドと共に使用して、ユーザーが選択する検索結果を検索コントロールが表示する直前に、フォーム データに基づく検索に表示される結果を最新のものに変更します。 ## <a name="related-topics"></a>関連トピック [addCustomFilter](../controls/addCustomFilter.md)
29.032258
235
0.796667
yue_Hant
0.690829
87998db7bdfe9f6e74c4db945334a273c19b81b2
2,728
md
Markdown
README.md
codersforcauses/ignite2018
107ed57fe789d517661ffa5f5ea510b8f99fa25e
[ "MIT" ]
5
2018-12-16T10:52:58.000Z
2019-01-19T01:43:25.000Z
README.md
codersforcauses/ignite2018
107ed57fe789d517661ffa5f5ea510b8f99fa25e
[ "MIT" ]
4
2019-01-03T16:10:51.000Z
2019-03-31T13:51:43.000Z
README.md
codersforcauses/ignite-mentoring-2018
107ed57fe789d517661ffa5f5ea510b8f99fa25e
[ "MIT" ]
null
null
null
# Ignite Mentoring Project ## Objective To provide [Ignite Mentoring](http://www.ignitementoring.org) a web based tool that allocates its mentors to classes while learning the basics of web development. ## TODO - [x] Practicing HTML, CSS, JS fundamentals - [x] Figure out and implement an algorithm - [x] Add some CSS to the thing - [ ] Work on the visual aesthetic of the project via HTML - [ ] Add additional features including having a specified amount of mentors in a class - [ ] Adding a shuffle button for the algorithm - [ ] Need to check for gender balance in the algorithm - [ ] Fix ordering of columns - [ ] Fix sizing of result text such that overflow is avoided - [ ] Add padding to results - [ ] 'Download file' button downloads the file instead of downloading the next shuffled - [ ] 'Re-Shuffle' and 'Download CSV' appears/clickable after file is uploaded ## Tech Stack Pure HTML, CSS, and JavaScript. ## Project Structure The project is made purely from HTML, CSS, and JavaScript. There is no backend. ### File relationships - `index.html` - `main.js` - `index.html` - `papaparse.min.js` - `index.html` - `style.css` - `csv files` not directly imported/related to anything - `data_gen.py` not directly imported/related to anything ### `index.html` This is the main page and entry point of the website. Simply open the file to open it in the browser.<br /> It provides the structure as well as importing of the required JS codes.<br /> You can see it importing code in these lines. ```html <script src="papaparse.min.js"></script> <script src="main.js"></script> ``` ### `main.js` This is the main JS code for the app.<br /> It reads the data using `papaparse.min.js` and uses a brute force algorithm to allocate mentors to classes, irresepective of their preferences. ### `style.css` The main stylesheet that brings color and joy to the bleak, featureless, and miserable world of HTML.<br /> Called Cascading Style Sheet, they describe the presentation of a document written in a markup language which includes the Hypertext Markup Language. ### `papaparse.min.js` A [third party JS library](https://www.papaparse.com/) that reads CSV files and stores them as JS objects.<br /> The code is unreadable because it is _minified_, a process wherein code is compressed to decrease its size such that there is less for the client to download. ### `csv files` Fake mentor data either provided by the client or generated by `data_gen.py`.<br /> CSV stands for Comma Separated Values and it is just another file format, like JSON. ### `data_gen.py` This is just a python code that generate some fake data.<br /> It can no longer make data that conforms to the requirement of the updated `main.js`.
37.369863
162
0.743402
eng_Latn
0.99706
8799afaea5dd14e38ba0d9ac2a7053217367019c
7,651
md
Markdown
wiki/translations/fr/Path_Drilling.md
arslogavitabrevis/FreeCAD-documentation
8166ce0fd906f0e15672cd1d52b3eb6cf9e17830
[ "CC0-1.0" ]
null
null
null
wiki/translations/fr/Path_Drilling.md
arslogavitabrevis/FreeCAD-documentation
8166ce0fd906f0e15672cd1d52b3eb6cf9e17830
[ "CC0-1.0" ]
null
null
null
wiki/translations/fr/Path_Drilling.md
arslogavitabrevis/FreeCAD-documentation
8166ce0fd906f0e15672cd1d52b3eb6cf9e17830
[ "CC0-1.0" ]
null
null
null
--- - GuiCommand:/fr Name:Path Drilling Name/fr:Path Perçage MenuLocation:Path → Perçage Workbenches:[Path](Path_Workbench/fr.md) --- # Path Drilling/fr ## Description La commande Perçage génère une opération de perçage durant l\'opération. <img alt="" src=images/Path_Drilling_Sample.png style="width:400px;"> *Ci-dessus: exemple d'opérations de perçage* ## Utilisation 1. Il existe plusieurs façons de lancer la commande : - Appuyez sur le Button **<img src="images/Path_Drilling.svg" width=16px> [Perçage](Path_Drilling/fr.md)**. - Sélectionnez l\'option **Path → <img src="images/Path_Drilling.svg" width=16px> Perçage** dans le menu. 2. Dans la section **Opération** : - Sélectionnez un **Contrôleur d'outil**. - Sélectionnez un **Mode de refroidissement**. - En option, activez et ajustez les éléments suivants : - **Peck** définissez la **Depth**. - **Dwell** définit le **Time** en secondes. - **Extended Depth** . 3. Dans la section **Géométrie de base**, vérifiez que la liste correspond aux trous destinés à être traités, et ajustez l\'ajout, l\'activation ou la désactivation, si nécessaire. Les trous peuvent être ajoutés en sélectionnant les faces des murs des Trous. 4. Dans la section **Depths**, vérifiez et ajustez si nécessaire les **Start Depth** et **Final Depth**. 5. Dans la section **Heights**, vérifiez et, si nécessaire, ajustez les **Safe Height** et **Clearance Height**. 6. Appuyez sur le bouton **OK** pour générer le(s) chemin(s) de perçage. ## Remarques - Lorsque vous utilisez des arêtes pour la géométrie de base, sélectionnez toujours le bord inférieur du trou. - Vérifiez toujours que l\'outil choisi est le bon diamètre pour le(s) trou(s) sélectionné(s). - **Peck désactivé** génère (cycles de forage pré-programmés G81). **Peck activé** génère (cycles de forage prédéfinis G83). - **Dwell activé** est actuellement non pris en charge, mais il est destiné à générer (cycles de forage fixes G82). ## Propriétés \'\'\' *Remarque* \'\'\': toutes ces propriétés ne sont pas disponibles dans l\'éditeur de fenêtre de tâches. Certaines ne sont accessibles que dans l\'onglet Données du panneau Vue de propriétés pour cette opération. #### Base Remarque: il est conseillé de ne pas modifier la propriété Placement des opérations de chemin. Déplacez ou faites pivoter le modèle de tâche de chemin selon vos besoins. - {{PropertyData/fr|Placement}}: emplacement global \[position et rotation\] de l\'objet - par rapport à l\'origine (ou à l\'origine du conteneur de l\'objet parent). - {{PropertyData/fr|Angle}} : angle en degrés appliqué à la rotation de l\'objet autour de la valeur de la propriété Axis. - {{PropertyData/fr|Axis}} : axe (un ou plusieurs) autour duquel faire pivoter l\'objet, défini dans les sous-propriétés: x, y, z. - {{PropertyData/fr|X}} : valeur de l\'axe x. - {{PropertyData/fr|Y}} : valeur de l\'axe y. - {{PropertyData/fr|Z}} : valeur de l\'axe z. - {{PropertyData/fr|Position}} : Position de l\'objet, définie dans les sous-propriétés: x, y, z - par rapport à l\'origine (ou à l\'origine du conteneur de l\'objet parent). - {{PropertyData/fr|X}} : valeur de distance x. - {{PropertyData/fr|Y}} : valeur de distance y. - {{PropertyData/fr|Z}} : valeur de distance z. - {{PropertyData/fr|Label}}: nom de l\'objet fourni par l\'utilisateur (UTF-8). - {{PropertyData/fr|Disabled}}: liste des fonctionnalités désactivées. #### Profondeur - {{PropertyData/fr|Clearance Height}}: hauteur nécessaire pour supprimer les pinces et les obstructions. - {{PropertyData/fr|Final Depth}}: profondeur finale de l\'outil - valeur la plus basse de Z. - {{PropertyData/fr|Safe Height}}: hauteur au-dessus de laquelle les mouvements rapides sont autorisés. (Hauteur de sécurité rapide entre les sites). - {{PropertyData/fr|Start Depth}}: profondeur de départ de l\'outil - *profondeur de la première coupe en Z*. - {{PropertyData/fr|Add Tip Length}}: calcule la longueur de la pointe et soustrait de la profondeur finale. - {{PropertyData/fr|Dwell Enabled}}: activer le temps (? ndt). - {{PropertyData/fr|Dwell Time}}: le temps de pause entre les cycles de picage. - {{PropertyData/fr|Peck Depth}}: profondeur de forage incrémental avant de se rétracter pour nettoyer les copeaux. - {{PropertyData/fr|Peck Enabled}}: active le picage. - {{PropertyData/fr|Retract Height}}: la hauteur du début de l\'alimentation et la hauteur lors de la rétraction de l\'outil lorsque le tracé est terminé. - {{PropertyData/fr|Return Level}}: contrôle le retrait de l\'outil. Par défaut = G98. #### Trajectoire - {{PropertyData/fr|Active}}: mis à False, pour empêcher l\'opération de générer du code. - {{PropertyData/fr|Comment}}: commentaire facultatif pour cette opération. - {{PropertyData/fr|User Label}}: étiquette attribuée par l\'utilisateur. - {{PropertyData/fr|Tool Controller}}: définit le contrôleur d\'outil utilisé dans l\'opération. #### Rotation (*si disponible*) - {{PropertyData/fr|Attempt Inverse Angle}}: tente automatiquement l\'angle inverse si la rotation initiale est incorrecte. - - {{PropertyData/fr|Enable Rotation}}: active la rotation pour accéder aux trous non normaux sur l\'axe Z. - {{PropertyData/fr|Angle Inverse}}: inverse l\'angle de la rotation. \'\' **Exemple:** change une rotation de -22,5 à 22,5 degrés.\'\' - {{PropertyData/fr|Reverse Direction}}: inverse l\'orientation de l\'opération de 180 degrés. ## Disposition de l\'éditeur de fenêtre de tâches *Les descriptions des paramètres sont fournies dans la liste des propriétés ci-dessus.* Cette section est simplement une représentation des paramètres de l'éditeur de fenêtres pour l'opération. #### Géométrie de base - **Add**: ajoute les éléments sélectionnés qui doivent être la base du ou des chemins. - **Supprimer**: supprime le ou les éléments sélectionnés dans la liste géométrie de base. - **Clear**: efface tous les éléments de la liste géométrie de base. #### Emplacement de base - **Add**: ajoute un emplacement de coordonnées (X, Y) à l\'opération de forage en cours. - **Remove**: supprime le ou les éléments d\'emplacement sélectionnés de la liste Emplacement de base. - **Edit**: édite l\'élément de lieu sélectionné. #### Profondeur - {{PropertyData/fr|Start Depth}} - {{PropertyData/fr|Final Depth}} #### Hauteur - {{PropertyData/fr|Safe Height}} - {{PropertyData/fr|Clearance Height}} #### Opération - {{PropertyData/fr|Tool Controller}} - {{PropertyData/fr|Retract Height}} - {{PropertyData/fr|Peck}} - {{PropertyData/fr|Peck Depth}} - {{PropertyData/fr|Dwell}} - {{PropertyData/fr|Dwell Time}} - {{PropertyData/fr|Use Tip Length}} ## Script **Voir aussi:** [FreeCAD Script de base](FreeCAD_Scripting_Basics/fr.md). Exemple : ```python #Place code example here. ``` {{Path_Tools_navi }} --- [documentation index](../README.md) > [Path](Path_Workbench.md) > Path Drilling/fr
32.2827
259
0.655601
fra_Latn
0.955953
879a12a99bf6ea8f74b6b4825a1088114c89f541
1,607
md
Markdown
CHANGELOG.md
MrWolong/terraform-alicloud-market-wordpress
541551c7ca59475b803765c692783bd730f6698f
[ "Apache-2.0" ]
null
null
null
CHANGELOG.md
MrWolong/terraform-alicloud-market-wordpress
541551c7ca59475b803765c692783bd730f6698f
[ "Apache-2.0" ]
null
null
null
CHANGELOG.md
MrWolong/terraform-alicloud-market-wordpress
541551c7ca59475b803765c692783bd730f6698f
[ "Apache-2.0" ]
10
2020-02-12T02:30:00.000Z
2022-03-16T08:48:25.000Z
## 1.2.0 (Unreleased) ## 1.1.0 (December 10, 2021) ENHANCEMENTS: - Removes the provider setting and improves the Readme [GH-10](https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/10) ## 1.0.1 (February 24, 2020) IMPROVEMENTS: - modify the provider version and add parameter `profile` [GH-9]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/9) ## 1.0.0 (February 24, 2020) IMPROVEMENTS: - add more outputs [GH-8]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/8) - Add examples for module [GH-7]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/7) - change type of product_suggested_price to number [GH-6]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/6) - support filtering image by supplier and price [GH-5]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/5) - Add variable description [GH-4]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/4) - Supplement whether to create resources and their relationships by switches [GH-3]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/3) - Add backend servers (ECS instance) to the SLB and Associate DNS with SLB [GH-2]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/2) - **NEW:** `market-wordpress` [GH-1]( https://github.com/terraform-alicloud-modules/terraform-alicloud-market-wordpress/pull/1)
59.518519
174
0.779091
kor_Hang
0.343152
879a2d5447cbc553cf449f69ed3da358fb7e4bed
32
md
Markdown
README.md
ylehilds/flight-scraper
88827968e70d5985f5d71653ce588361e76ab3fa
[ "MIT" ]
null
null
null
README.md
ylehilds/flight-scraper
88827968e70d5985f5d71653ce588361e76ab3fa
[ "MIT" ]
null
null
null
README.md
ylehilds/flight-scraper
88827968e70d5985f5d71653ce588361e76ab3fa
[ "MIT" ]
null
null
null
# flight-scraper Flight Scraper
10.666667
16
0.8125
eng_Latn
0.97039
879b13268bd22ebe96110cb8e1186cdb7063e057
696
md
Markdown
_teaching/xlab.md
NatyaHans/NH.github.io
b14e92ba4d60bbec8b4761edbf35ab73af70ce3f
[ "MIT" ]
null
null
null
_teaching/xlab.md
NatyaHans/NH.github.io
b14e92ba4d60bbec8b4761edbf35ab73af70ce3f
[ "MIT" ]
null
null
null
_teaching/xlab.md
NatyaHans/NH.github.io
b14e92ba4d60bbec8b4761edbf35ab73af70ce3f
[ "MIT" ]
null
null
null
--- title: "X lab (Cross-Disciplinary Laboratory)" collection: teaching type: "HHMI funded undergrad lab" permalink: /teaching/xlab venue: "University of Florida, Department of Biology" date: 2017-04-01 location: "Gainesville, Florida" --- I was instructor for the cross discplinary laboratory for two years. The lab coordinators for this HHMI funded program are Dr. David Jullian and Dr. Gabriela Waschewsky. X lab 1 (ISC2400L) and X lab 2 (ISC2401L) X lab 1 (ISC2400L) ====== I taught the X lab during the following semesters Fall 2015, Summer 2016, Fall 2016 X lab 2 (ISC2401L) ====== I taught the Xlab 2 in Spring 2016. For Fall 2017, I taught the Pre-health Post-Baccalaureate Program.
30.26087
169
0.752874
eng_Latn
0.971815
879b1a2d108048ed8177e633f9a7218ab362ed67
3,456
md
Markdown
README.md
samtgarson/silent-listener
7b2017b5feb34e844813c26a6c1fd2f84a3b4921
[ "MIT" ]
null
null
null
README.md
samtgarson/silent-listener
7b2017b5feb34e844813c26a6c1fd2f84a3b4921
[ "MIT" ]
null
null
null
README.md
samtgarson/silent-listener
7b2017b5feb34e844813c26a6c1fd2f84a3b4921
[ "MIT" ]
null
null
null
# SilentListener > Listen out for selectors and instantiate your classes SilentListener allows you to define _SilentListeners_ which will be instantiated whenever an element appears which matches a given selector. It also comes with some handy functionality for reinstantiating classes using events, and handling class options. ## Installation Add this line to your application's Gemfile: ```ruby gem 'silent_listener' ``` And then execute: $ bundle Or install it yourself as: $ gem install silent_listener ## Usage ### Dependencies This gem comes with `rails-coffee` as it is written in CoffeeScript, but also depends on jQuery or some other dom selection library which uses `$` (e.g. [Sprint](https://github.com/bendc/sprint)). It leaves that choice up to you. ### Setup To define a silent listener, name your class as a `SilentListener`, and make it a descendant of the `SilentListener` class: ```coffeescript class SilentListeners.MyClass extends SilentListener ``` To add a selector to listen out for, define a class variable called `silentListenerSelector`: ```coffeescript @silentListenerSelector: '.my-class' ``` Once your class has been instantiated, a class method instantiate will be called (if it exists), and a class variable called `el` will be set: ```coffeescript instantiate: => @el.doSomething() ``` The elements will have the instance of its silent listner class set on a data attribute `data-silent-listener`, as well as being accessible on the class variable `instances`. e.g.: ```coffeescript``` $('.my-class').first().data('silent-listener') # => bound class instance MyClass.instances # => array of all class instances ``` ### Options Given an element with the data attribute `data-listener-options`, silent listeners will merge the `data-listener-options` hash with the result of an instance method called `_defaultOptions` and set an instance variable called `options` upon instantiation. e.g.: ```html <div class="my-class" data-listener-options="{ some-option: true }"></div> ``` ```coffeescript class SilentListeners.MyClass extends SilentListener @silentListenerSelector: '.my-class' instantiate: => console.log(@options) # the underscore implies a private method, despite the concept not existing yet _defaultOptions: => some-option: false ``` ### Events Each silent listener can told to look again for its selector by calling `MyClass.bindSelf()` You can also tell all of your silent listeners to rebind themselves by calling `SilentListener.bindAll()` ## Development ### ToDo: Test coverage, potentially using [konacha](https://github.com/jfirebaugh/konacha) To install this gem onto your local machine, run `bundle exec rake install`. To release a new version, update the version number in `version.rb`, and then run `bundle exec rake release`, which will create a git tag for the version, push git commits and tags, and push the `.gem` file to [rubygems.org](https://rubygems.org). ## Contributing Bug reports and pull requests are welcome on GitHub at https://github.com/[USERNAME]/silent_listener. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the [Contributor Covenant](http://contributor-covenant.org) code of conduct. ## License The gem is available as open source under the terms of the [MIT License](http://opensource.org/licenses/MIT).
33.882353
324
0.75
eng_Latn
0.988458
879b58204bae6e6c3f4302db631ef9cbbdbc5c61
449
md
Markdown
Day 4/Tasks/Task 1/README.md
emilia98/Swift-Training-Advanced
051a06d489c549539163d1b128242ebe5ef0c769
[ "MIT" ]
null
null
null
Day 4/Tasks/Task 1/README.md
emilia98/Swift-Training-Advanced
051a06d489c549539163d1b128242ebe5ef0c769
[ "MIT" ]
null
null
null
Day 4/Tasks/Task 1/README.md
emilia98/Swift-Training-Advanced
051a06d489c549539163d1b128242ebe5ef0c769
[ "MIT" ]
null
null
null
# Task 1 - BitSet ``` Да се разработи клас BitSet, който да се държи като едно много голямо двоично число. Да могат да се манипулират отделните битове и да се задава начална големина (брой битове). Функции: - проверка дали е сет-нат бит на позиция х get(x) -> Bool - задаване на бит на позиция x да бъде 1 set(x) - нулиране на бит на позиция x clear(x) - нулиране на всички битове reset() - брой сет-нати битове setBitCount() -> Int ```
24.944444
90
0.712695
bul_Cyrl
0.999643
879b84499bb76046cefd7c9b058f62b8fbdba741
3,194
md
Markdown
_posts/2017-05-19-crest-meeting-may1.md
net-titech/net-titech.github.io
ddc803410942de71203ae2a56c764b0d418a032b
[ "MIT" ]
2
2017-02-27T12:07:01.000Z
2017-06-19T08:04:57.000Z
_posts/2017-05-19-crest-meeting-may1.md
net-titech/net-titech.github.io
ddc803410942de71203ae2a56c764b0d418a032b
[ "MIT" ]
null
null
null
_posts/2017-05-19-crest-meeting-may1.md
net-titech/net-titech.github.io
ddc803410942de71203ae2a56c764b0d418a032b
[ "MIT" ]
1
2019-10-15T13:27:12.000Z
2019-10-15T13:27:12.000Z
--- layout: post title: CREST-Deep Weekly Meeting excerpt: "2017-05-19" categories: [Meeting] comments: true pinned: false icon: fa-clock-o --- # 2017-05-19 / CREST-Deep Weekly Meeting ## <i class="fa fa-file-text"></i> Literature Update Since the last (recorded) meeting (April 7th), we have focused on designing a novel neural network compression scheme, which can be deployed without specialize hardware and propriety methods. For this reason, we only have a few update on the literature: - The effect of parameter robustness and parameter quantization effect is interesting to investigate. Recently, there are few papers study this problem. If we can quantify the parameter robustness of a deep neural network, we can develop a much more effective lossy compressing algorithm. - Deep neural network distillation is a good idea. However, we wonder if it is any better than just train a smaller model. - Song's pruned model can be retrained without the pruning constrain to achieve better result (AlexNet on ImageNet - ICLR 2017). ## <i class="fa fa-flask"></i> Research Update We have two main research direction so far: - Deployment of compressed model on mobile device. We have been able to compress AlexNet down to 8MB from 240MB (~30 times). The main strong point of our compression scheme is that we do not need to have a specialized hardware to read a model in its compressed format. Moreover, our compression scheme enables GPU to decode the weight matrix on the fly, meaning we do not need to decompress the model on CPU (less sequential code). On this direction, we are exploring the lossy compression scheme (potentially achieving 3.5MB compressed size). We are also working on low rank decomposition and some other engineering methods for speeding up the computation. - Besides our engineering effort above, we are working on some other approach such as network distillation. However, we are having trouble recreating results from some papers addressing this method. Further investigation will be presented in later posts. - (Side note) Since most of the papers addressing neural network pruning-retraining scheme doesn't provide the pruning threshold. We have created our own pruning-retraining procedure on our [Caffe fork](https://github.com/net-titech). ## <i class="fa fa-bullseye"></i> Next Week Objectives We have set up an independent meeting with Yokota-lab to discuss the possibility of encoding the low-rank approximation matrices. Since low-rank approximation is a well-established method, we hope to further increase computational speed while reducing the storage size of a deep neural network. Main task list for the next week: - Write proposals (check [Trello](https://trello.com/deepcrestmuratateam)). - Paper reading (check [Trello](https://trello.com/deepcrestmuratateam)). - Unit test for the new pruning module in Caffe (check github). - Unit test for encoding scheme, fix the prefix sum to adapt with large relative index. - Explore smaller models resulting from novel training methods. - Find out the performance impact of low-rank approximation and the possibility of compressing low-rank matrices (investigate on both `conv` and `fc` layers).
45.628571
80
0.789919
eng_Latn
0.997759
879c9084a8502d2e75a6b09c84bb4eae433aae09
673
md
Markdown
docs/SegmentBody.md
launchdarkly/api-client-ruby
61091e20d76eaea3ffac58909fcf44194b3c347f
[ "Apache-2.0" ]
1
2020-05-12T06:09:40.000Z
2020-05-12T06:09:40.000Z
docs/SegmentBody.md
launchdarkly/api-client-ruby
61091e20d76eaea3ffac58909fcf44194b3c347f
[ "Apache-2.0" ]
2
2021-09-03T19:03:13.000Z
2021-09-29T22:31:22.000Z
docs/SegmentBody.md
launchdarkly/api-client-ruby
61091e20d76eaea3ffac58909fcf44194b3c347f
[ "Apache-2.0" ]
3
2020-11-21T17:58:33.000Z
2021-05-24T09:43:28.000Z
# LaunchDarklyApi::SegmentBody ## Properties | Name | Type | Description | Notes | | ---- | ---- | ----------- | ----- | | **name** | **String** | A human-friendly name for the segment | | | **key** | **String** | A unique key used to reference the segment | | | **description** | **String** | A description of the segment&#39;s purpose | [optional] | | **tags** | **Array&lt;String&gt;** | Tags for the segment | [optional] | | **unbounded** | **Boolean** | | [optional] | ## Example ```ruby require 'launchdarkly_api' instance = LaunchDarklyApi::SegmentBody.new( name: null, key: null, description: null, tags: [&quot;ops&quot;], unbounded: null ) ```
24.925926
90
0.59584
eng_Latn
0.567856
879cda82f7da766f3d698c6f98dd09492647aec7
107
md
Markdown
README.md
CapraRoyale/05-Trivia-Game
32089428136b9ce9f470968b080943530c2b3cc7
[ "MIT" ]
null
null
null
README.md
CapraRoyale/05-Trivia-Game
32089428136b9ce9f470968b080943530c2b3cc7
[ "MIT" ]
null
null
null
README.md
CapraRoyale/05-Trivia-Game
32089428136b9ce9f470968b080943530c2b3cc7
[ "MIT" ]
null
null
null
# 05-Trivia-Game - Javascript Homework #2 A simple Javascript Trivia game implementing timer functions.
15.285714
61
0.785047
eng_Latn
0.640759
879d10f59e4ec9678f11914a54bde4aa1f8357b6
5,291
md
Markdown
markdown/Week 3_2D Maps_underconstruction.md
MRIO/NBDynSys
31a8b11f2aa16bb46556acde67647f5f6db0a128
[ "MIT" ]
1
2020-09-10T12:33:42.000Z
2020-09-10T12:33:42.000Z
markdown/Week 3_2D Maps_underconstruction.md
MRIO/NBDynSys
31a8b11f2aa16bb46556acde67647f5f6db0a128
[ "MIT" ]
null
null
null
markdown/Week 3_2D Maps_underconstruction.md
MRIO/NBDynSys
31a8b11f2aa16bb46556acde67647f5f6db0a128
[ "MIT" ]
null
null
null
# Week 3 ## Mandelbrot Fractals and 2D Maps ##Essential review from Week 2 — i.e. 1D Maps* # Notation fundamentals: in 1D maps, fixed points come in two breeds, Sinks (also called stable fixed points, attractors) andSources (unstable fixed points, repellers). See more about stability of fixed points, 1.3AY , 3.4HK) Check that you understand the diference between hyperbolic and non hyperbolic fixed points. Hyperbolic fixed points determine that the flow either contracts or expands. Non hyperbolic fixedpoints are often accompanied of changes of the dynamics, i.e., bifurcations often co-occur withnon hyperbolic fixed points (e.g. Flip Bifurcation or Period Doubling Bifurcation HK 3.4) ## TODOs of Last Week: - Find out about shadow orbits of the logistic map. Which parameter ranges are shadow orbits observed? (hint: shadow orbits are usually chaotic orbits) - Compute shadow orbits of the logistic map in parameter ranges where shadow orbits are observed (e.g matlab). That is, keep the a parameter fixed (thus maintaining the dynamical systemunchanged) and calculate two orbits that start with very similar initial conditions, for example .7 and .7+eps, where eps is an infinitesimal. Then: - Calculate the exponential separation of orbits (Lyapunov exponent , 6.1AY. - Extra challengecompute the Lyapunov exponent for the logistic map in the 0<a<4 parameter range (see figure6.3AY to check for your solution, or to skip the challenge altogether) Certify yourself that you grok the following statements: - It is possible to tell whether a map will have periodic orbits by inspecting the number of fixed points of the iterates of the map. Make sure this statement is sensible: section 1.6 - It is possible to tell whether the periodic orbits of a map are stable or unstable by calculating the derivatives at the fixed points of the iterated maps: section 1.6 ### Today’s topics After reconsidering the order of presentation of today’s topics, it becomes apparent that I should have sticked withthe more natural sequence for the study of 2D maps, as I have originally laid out. First we learn about linear maps, and the behavior of orbits close to fixed points. Second we learn about how to use linear maps in understanding **non linear maps ** (like the Henon map for example), through linearization at the fixed points of the system, via a Jacobian Matrix. Linear maps : deepen your understanding about the behavior of orbits of linear systems, where the only fixed point is at the origin (section 2.3 of AY). The qualitative behavior of orbits close to the fixed points are fully determined as a function of theeigenvectors and eigenvalues of the linear system (section AY 2.3). https://app.classeur.io/#!/files/KGwkBpb7d4Ks7fMoFINt Eigenvectors and eigenvalues - to remind yourself of the effect of using a matrix to transform avector, use the function eigshow in matlab. To examine your own matrix, call it in the command linewith eigshow(A) — where A is your matrix defined as Z = [ a b ; c d ] where a,b,c,d are specifiednumbers. Eigenvalues determine whether the orbits contract or expand. In 2D maps we find three kinds offixed points, attractors (or sinks), repellers (or sources) and saddle nodes (which attract in certaindirections and contract in others) Cases to note: Distinct Real Eigenvalues (AY Example 2.5) Repeated Eigenvalue (AY Example 2.6) Complex Eigenvalues (AY Example 2.7) NOTE: a simple coordinate transform (translation) may transform any point in the dynamics as the origin.The dynamics around the new origin is preserved (AY section 2.4) Linearizing a map is an essential tool to analyze non linear maps. A large class of non linear maps areamenable to study via calculating the Jacobian Matrix at the fixed points (AY section 2.5) ## The 3 Body Problem The analysis of the 3 body problem by Poincaré led to many seminal discoveries and developments in thetheory of dynamical systems. Read more about it at section 2.1, page 41-52. The Mandelbrot set The Mandelbrot set is an example of an iterated map in 2D. It displays fractal basin boundaries (section AY4.4), similar to those of the Henon Attractor. That is, the perimeter shows self-similar detail in multiplescales. Further reading: AY 4.5 Fractal Dimension and AY page 167. TO DOs: Extend the programming concepts learned from the logistic map in 1D and make a matlab function thatreturns a 2D orbit This function should return both an orbit of the Henon Map on the x y plane and the orbit of the twovariables as a function of the number of iterations (x axis represents sequences of states). Calculate the fixed points of example 2.3 AY Reproduce (qualitatively) figure 2.9, with the mathematica demonstration: http://demonstrations.wolfram.com/PlayingWithTheHenonMapStartingWithACircleOrASquare/ DISCOVER: Background to understand how maps are obtained from differential equations https://app.classeur.io/#!/files/KGwkBpb7d4Ks7fMoFINt Page 2 of 3 Classeur – The app 29/09/16 08:48 To know more: https://www.youtube.com/watch?v=zXTpASSd9xE&index=11&list=PL7G5UnDx-bEaFO2b3FS0q33bRm0jH8cAp Fun point to inspect is in the description Mandelbub, a mandelbrot in 3D Great Information about the Mandlebrot set is available in the Numberphile channel in Youtube
66.1375
574
0.794746
eng_Latn
0.997937
879d17ac21a187b9bf73c74dcfde75e0058006fd
586
md
Markdown
_teaching/2018-IntroToMATLAB.md
RobertGoellinger/RobertGoellinger.github.io
559335277cbbc8e28b147d3f9cc793e198186991
[ "MIT" ]
null
null
null
_teaching/2018-IntroToMATLAB.md
RobertGoellinger/RobertGoellinger.github.io
559335277cbbc8e28b147d3f9cc793e198186991
[ "MIT" ]
null
null
null
_teaching/2018-IntroToMATLAB.md
RobertGoellinger/RobertGoellinger.github.io
559335277cbbc8e28b147d3f9cc793e198186991
[ "MIT" ]
null
null
null
--- title: "Introduction to MATLAB" collection: teaching type: "Programming tutorial" permalink: /teaching/2018-IntroToMATLAB venue: "RWTH Aachen University, IT Center" date: 2018-10-1 location: "Aachen, Germany" --- A short introduction to scientific programming in MATLAB. Topics ------ - Introduction to MATLAB - Installation and Support - First Steps in MATLAB - Calculating in MATLAB - Boolean Operators, Loops, Functions, Stuctures - Visualisation - Importing and Displaying Data - Introduction to GUI-Programming using Appdesigner - Developing a GUI (Callbacks, Handles)
20.206897
57
0.769625
yue_Hant
0.414051
879d4035e31caa8acfc98a77232f78e7880820c8
278
md
Markdown
LICENSES/low_view_of_a_tennis_court.md
usagiga/siv3d-video-archive
f089c8f7dc253e2f214a9cb0e0bc6cbf420f1df0
[ "MIT" ]
null
null
null
LICENSES/low_view_of_a_tennis_court.md
usagiga/siv3d-video-archive
f089c8f7dc253e2f214a9cb0e0bc6cbf420f1df0
[ "MIT" ]
null
null
null
LICENSES/low_view_of_a_tennis_court.md
usagiga/siv3d-video-archive
f089c8f7dc253e2f214a9cb0e0bc6cbf420f1df0
[ "MIT" ]
1
2021-05-01T16:09:24.000Z
2021-05-01T16:09:24.000Z
# [Low View Of a Tennis Court](https://coverr.co/videos/low-view-of-a-tennis-court-jp59ZoJwcb) CC0 1.0 This file located at `/siv3d-video-archive/Assets/low_view_of_a_tennis_court.mp4` . Available on [coverr.co](https://coverr.co/videos/low-view-of-a-tennis-court-jp59ZoJwcb)
39.714286
94
0.769784
yue_Hant
0.475547
879dacea61817845f7eb7bdd085b58fc518a0c09
568
md
Markdown
2017/CVE-2017-7953.md
justinforbes/cve
375c65312f55c34fc1a4858381315fe9431b0f16
[ "MIT" ]
2,340
2022-02-10T21:04:40.000Z
2022-03-31T14:42:58.000Z
2017/CVE-2017-7953.md
justinforbes/cve
375c65312f55c34fc1a4858381315fe9431b0f16
[ "MIT" ]
19
2022-02-11T16:06:53.000Z
2022-03-11T10:44:27.000Z
2017/CVE-2017-7953.md
justinforbes/cve
375c65312f55c34fc1a4858381315fe9431b0f16
[ "MIT" ]
280
2022-02-10T19:58:58.000Z
2022-03-26T11:13:05.000Z
### [CVE-2017-7953](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7953) ![](https://img.shields.io/static/v1?label=Product&message=n%2Fa&color=blue) ![](https://img.shields.io/static/v1?label=Version&message=n%2Fa&color=blue) ![](https://img.shields.io/static/v1?label=Vulnerability&message=n%2Fa&color=brighgreen) ### Description INFOR EAM V11.0 Build 201410 has XSS via comment fields. ### POC #### Reference - http://seclists.org/fulldisclosure/2017/May/54 - https://www.exploit-db.com/exploits/42029/ #### Github No PoCs found on GitHub currently.
29.894737
88
0.728873
yue_Hant
0.283137
879e331f133833478096fd14820e069842a730eb
4,554
md
Markdown
README.md
hugopilot/raisr
b64345f9ec9ac6ca97a0d1acb4db746535f2f995
[ "MIT" ]
null
null
null
README.md
hugopilot/raisr
b64345f9ec9ac6ca97a0d1acb4db746535f2f995
[ "MIT" ]
null
null
null
README.md
hugopilot/raisr
b64345f9ec9ac6ca97a0d1acb4db746535f2f995
[ "MIT" ]
null
null
null
# RAISR A Python implementation of [RAISR](http://ieeexplore.ieee.org/document/7744595/) ## How To Use ### Prerequisites You need Python 3.x, older versions don't work out-of-the-box. You can install most of the following packages using [pip](https://pypi.python.org/pypi/pip). It is also important to have a 64 bit version of Python 3.x * [OpenCV-Python](https://pypi.python.org/pypi/opencv-python) * [NumPy](http://www.numpy.org/) * [SciPy](https://www.scipy.org/) * [Python Imaging Library (PIL)](http://www.pythonware.com/products/pil/) * [Matplotlib](https://matplotlib.org/) * [scikit-image](http://scikit-image.org/) ### Training Put your training images in the `train` directory. The training images are the **high resolution (HR)** ones. Run the following command to start training. ``` python train.py ``` In the training stage, the program virtually downscales the high resolution images. The program then trains the model using the downscaled version images and the original HR images. The learned filters `filter.p` will be saved in the root directory of the project. The result Q, V matrix (`q.p` and `v.p`) will also be saved for further retraining. To train an improved model with your previous Q, V, use the following command. ``` python train.py -q q.p -v v.p ``` ### Testing Put your testing images in the `test` directory. Basically, you can use some **low resolution (LR)** images as your testing images. By running the following command, the program takes `filter.p` generated by training as your default filters. ``` python test.py ``` The result (HR version of the testing images) will be saved in the `results` directory. To use an alternative filter file, take using the pretrained `filters/filter_BSDS500` for example, use the following command. ``` python test.py -f filters/filter_BSDS500 ``` ## Visualization Visualing the learned filters ``` python train.py -p ``` Visualing the process of RAISR image upscaling ``` python test.py -p ``` For more details, use the help command argument `-h`. ## Testing Results Comparing between original image, bilinear interpolation and RAISR: | Origin | Bilinear Interpolation | RAISR | |:----------------------:|:----------------------:|:----------------------:| |![origin_gray_crop bmp](https://user-images.githubusercontent.com/12198424/27002908-28a69cf2-4e1f-11e7-954d-1135950cd760.png)|![cheap_crop bmp](https://user-images.githubusercontent.com/12198424/27002909-28a7834c-4e1f-11e7-875e-30bb4eaaa0d3.png)|![raisr_gray_crop bmp](https://user-images.githubusercontent.com/12198424/27002910-28ae90f6-4e1f-11e7-83e5-3e63ca07b308.png)| Other results using images taken from [BSDS500 database](https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/resources.html) and [ArTe-Lab 1D Medium Barcode Dataset](http://artelab.dista.uninsubria.it/downloads/datasets/barcode/medium_barcode_1d/medium_barcode_1d.html): | Origin | RAISR | |:-------------:|:-------------:| |![origin_crop bmp](https://user-images.githubusercontent.com/12198424/27002925-81cf7f88-4e1f-11e7-8a75-4975db1d37b8.png)|![raisr_crop bmp](https://user-images.githubusercontent.com/12198424/27002926-81d126b2-4e1f-11e7-8814-f2fce323caac.png)| |![origin_crop bmp](https://user-images.githubusercontent.com/12198424/27002932-a39f6cc2-4e1f-11e7-9dea-3aa79a9d9764.png)|![raisr_crop bmp](https://user-images.githubusercontent.com/12198424/27002933-a3a248ac-4e1f-11e7-9c81-807d3a5eac90.png)| |![origin_crop bmp](https://user-images.githubusercontent.com/12198424/27002942-c9ba697a-4e1f-11e7-8743-d41be65c09d3.png)|![raisr_crop bmp](https://user-images.githubusercontent.com/12198424/27002943-c9bcefd8-4e1f-11e7-9e55-bf5d93f17a9c.png)| ## Contribute We actively welcome pull requests. Learn how to [contribute](https://github.com/movehand/raisr/blob/master/docs/CONTRIBUTING.md). ## References * Y. Romano, J. Isidoro and P. Milanfar, "RAISR: Rapid and Accurate Image Super Resolution" in IEEE Transactions on Computational Imaging, vol. 3, no. 1, pp. 110-125, March 2017. * P. Arbelaez, M. Maire, C. Fowlkes and J. Malik, "Contour Detection and Hierarchical Image Segmentation", IEEE TPAMI, Vol. 33, No. 5, pp. 898-916, May 2011. * Alessandro Zamberletti, Ignazio Gallo and Simone Albertini, "Robust Angle Invariant 1D Barcode Detection", Proceedings of the 2nd Asian Conference on Pattern Recognition (ACPR), Okinawa, Japan, 2013 ## License [MIT.](https://github.com/movehand/raisr/blob/master/LICENSE) Copyright (c) 2017 James Chen
46.948454
372
0.741107
eng_Latn
0.496378
879ed1151583af1bcd83851c513d43d2dbdaf3f5
13
md
Markdown
README.md
luciano1980/MIT
dfae9a09e767ccac9531ca695fdb14af057c7df1
[ "MIT" ]
null
null
null
README.md
luciano1980/MIT
dfae9a09e767ccac9531ca695fdb14af057c7df1
[ "MIT" ]
null
null
null
README.md
luciano1980/MIT
dfae9a09e767ccac9531ca695fdb14af057c7df1
[ "MIT" ]
null
null
null
# MIT Teste1
4.333333
6
0.692308
deu_Latn
0.589508
879f053e4022d399c3e112919e82be05b6cca29c
301
md
Markdown
README.md
15chrjef/reactNativeAnimations
1102779359f79cb73ebe597eca36e11ed31e7edc
[ "MIT" ]
null
null
null
README.md
15chrjef/reactNativeAnimations
1102779359f79cb73ebe597eca36e11ed31e7edc
[ "MIT" ]
null
null
null
README.md
15chrjef/reactNativeAnimations
1102779359f79cb73ebe597eca36e11ed31e7edc
[ "MIT" ]
null
null
null
# working Animations and a listView on React Native. TO START Download Exponent from getexponent.com Download Xcode from apple ``` git clone https://github.com/15chrjef/reactNativeAnimations animations cd animations npm i ``` open 'iOS simulator' in exponent and navigate to your 'animations' folder
23.153846
73
0.800664
kor_Hang
0.62627
87a059261f2e21d64ecb2584bd9d2338202a25e4
1,861
md
Markdown
README.md
duryan00/master_sql_course
c8bf55b16bff8d2d6eae5a9913f3d54ce689baf6
[ "MIT" ]
null
null
null
README.md
duryan00/master_sql_course
c8bf55b16bff8d2d6eae5a9913f3d54ce689baf6
[ "MIT" ]
null
null
null
README.md
duryan00/master_sql_course
c8bf55b16bff8d2d6eae5a9913f3d54ce689baf6
[ "MIT" ]
null
null
null
# [Master SQL for Data Science Course at Udemy](https://www.udemy.com/course/master-sql-for-data-science/) This repo contains my solutions for the [Master SQL for Data Science Course at Udemy](https://www.udemy.com/course/master-sql-for-data-science/), fully completed in February 2020. #### The following topics were covered throughout the course. Chapter 1: Database Basics * creating and populating tables Chapter 2: SQL Query Basics * SELECT statement, WHERE with AND/OR clause * filtering operators: IN, NOT IN, IS NULL, BETWEEN * using ORDER BY, LIMIT, DISTINCT and renaming columns Chapter 3: Using Functions * UPPER(), LOWER(), LENGTH(), TRIM(), SUBSTRING(), REPLACE(), POSITION(), COALESCE(), MIN(), MAX(), AVG(), SUM(), COUNT() Chapter 4: Grouping Data and Computing Aggregates * GROUP BY and HAVING clauses Chapter 5: Using Subqueries * aliasing sources of data * subqueries with ANY and ALL operators Chapter 6: Using the CASE Clause in Interesting Ways * conditional expressions using CASE clause * transposing data using the CASE clause Chapter 7: Advanced Query Techniques using Correlated Subqueries * understanding correlated subqueries Chapter 8: Working with Multiple Tables * INNER and OUTER joins * using UNION, UNION ALL and EXCEPT clauses * cartesian product with the CROSS JOIN * creating Views vs Inline Views Chapter 9: Window Functions for Analytics * Window functions using the OVER() clause * ordering data in Window frames * RANK, FIRST_VALUE and NTILE Functions * LEAD and LAG functions * working with Rollups and Cubes Final: Difficult Query Challenges --- ## **Udemy's Certificate of Completion** <!-- ![Udemy Completion Certificate](/images/udemy_completion_certificate.jpg) --> <img src="/images/udemy_completion_certificate.jpg" title="A cute kitten" width="941.176470588" height="700" />
25.148649
179
0.758732
eng_Latn
0.627481
87a06b97316bf646e91288fda97553cefbc018f1
106
md
Markdown
node-api-cn/tls/tlssocket_remoteport.md
develop-basis/nodejs
c9911b8595ba28fd567ae8e4dcfd627bb655f67a
[ "MIT" ]
null
null
null
node-api-cn/tls/tlssocket_remoteport.md
develop-basis/nodejs
c9911b8595ba28fd567ae8e4dcfd627bb655f67a
[ "MIT" ]
null
null
null
node-api-cn/tls/tlssocket_remoteport.md
develop-basis/nodejs
c9911b8595ba28fd567ae8e4dcfd627bb655f67a
[ "MIT" ]
null
null
null
<!-- YAML added: v0.11.4 --> Returns the numeric representation of the remote port. For example, `443`.
15.142857
74
0.688679
eng_Latn
0.979984
87a072ef4f0f2f1b6ecdf97e837a2942b4e84935
5,415
md
Markdown
src/pages/articles/2011-04-12---frankenstrat-vs-iron-gear-the-installation/2011-04-12-frankenstrat-vs-iron-gear-the-installation.md
monquixote/nick_long_blog
946fc841ac20ebe5ecc7a7ea478534dbf8e0f563
[ "MIT" ]
1
2020-07-24T20:07:40.000Z
2020-07-24T20:07:40.000Z
src/pages/articles/2011-04-12---frankenstrat-vs-iron-gear-the-installation/2011-04-12-frankenstrat-vs-iron-gear-the-installation.md
monquixote/nick_long_blog
946fc841ac20ebe5ecc7a7ea478534dbf8e0f563
[ "MIT" ]
null
null
null
src/pages/articles/2011-04-12---frankenstrat-vs-iron-gear-the-installation/2011-04-12-frankenstrat-vs-iron-gear-the-installation.md
monquixote/nick_long_blog
946fc841ac20ebe5ecc7a7ea478534dbf8e0f563
[ "MIT" ]
null
null
null
--- id: 13 title: 'Frankenstrat vs Iron Gear - The Installation' date: 2011-04-12T20:31:00+00:00 author: admin layout: post guid: http://www.nick-long.com/frankenstrat-vs-iron-gear-the-installation/ permalink: /frankenstrat-vs-iron-gear-the-installation/ blogger_blog: - monquixote.blogspot.com blogger_author: - Monquixote blogger_permalink: - /2011/04/frankenstrat-vs-iron-gear-installation.html category: Guitar tags: - guitar --- Last weekend the deed was done and I installed my new Iron Gear pickups in my Strat (if you don't know what I'm talking about <a href="http://nicklong.com/frankenstrat-meet-iron-gear">look here</a>) I thought I would share some of the photos with you as modifying a vintage guitar already hacked about by previous owners is a slightly different prospect than the brand new guitars that are usually worked on in youtube videos. I decided to go for white pickups as the aged cream coloured pickups most people offer on &#8220;aged&#8221; guitars look really fake to me and I like to age my own gear.  Here are all my tools laid out ready for the job (There's nothing worse than getting half way through a task and not having what you need)  <a href="http://posterous.com/getfile/files.posterous.com/nicklong/nuDyGhFaglFeJtjoaiqeoDrAIxstjjjzfElfBoJtgoGJwfADwtlmHxpadfka/media_httpi19photobuc_oExea.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_oexea" height="375" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_oExea.jpg.scaled500.jpg" width="500" /></a> The first task was removing the strings. My top tip if you have a guitar with a floating bridge is to put a shim under it before you remove the strings so it doesn't fall off the posts.Paying cards work best, but I used a nearby notepad.  <a href="http://posterous.com/getfile/files.posterous.com/nicklong/tIaCscszGGnpDbIfDEqxluCFckcavfDFzsDBIxzIEdvlCwGexBslfwfagqJn/media_httpi19photobuc_oIvIB.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_oivib" height="375" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_oIvIB.jpg.scaled500.jpg" width="500" /></a> My first issue was getting the scratchplate off, one of the screws had rusted solid on the top and required some tweaking with pliers to remove. Once that was sorted off came the scratch plate (Notice the unfaded lake placid blue under the scratchplate). <a href="http://posterous.com/getfile/files.posterous.com/nicklong/GCCCDJgHhrJHFehhoExAypfjICakIzsiHwpcBgjbEFBElHIzhirfhduHtgAG/media_httpi19photobuc_bEpjl.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_bepjl" height="667" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_bEpjl.jpg.scaled500.jpg" width="500" /></a> Only then was the full horror of the hacked about wiring revealed. The new soldering iron was powered on but immediately started emitting loads of white smoke (That's what you get for £6 from Amazon). I figured it was probably burning off some ill advised anti corrosion coating on the metal and left in running in the garden for half an hour after which it was behaving its self. Desoldering is usually a fairly simple task but as you can see in the next photo dodgy repairs left the volume pot with a lump of solder on it the size of a Malteser which my plucky 30W smoke belching soldering iron declined to melt. Eventually after trying until some of the insulation on the wiring started to fry I resorted to hacking it off with a soldering iron.  <a href="http://posterous.com/getfile/files.posterous.com/nicklong/pxEtgdCtmyjCcBHDuajJcDAgoiCDhemDqlnxFxxbFjinlqnpzeghyiudxBcI/media_httpi19photobuc_xpnhk.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_xpnhk" height="667" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_xpnhk.jpg.scaled500.jpg" width="500" /></a> With the old pickups removed fitting the Iron Gears was fairly painless. They are really well made (better than the Fender USA originals) with single conductor waxed cloth cable and all the vintage goodies. The only issue was that the supplied springs were a little big so rather than cut them down I used the old ones.  <a href="http://posterous.com/getfile/files.posterous.com/nicklong/hytxuAsvczizmkizvGxfhHkvrqmvmHuggBhambonxhgpgJibfxlxiFupCiwd/media_httpi19photobuc_BssIb.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_bssib" height="375" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_BssIb.jpg.scaled500.jpg" width="500" /></a> With all the pickups in I put the guitar back together and added the two rather jazzy roller string trees as the old ones were getting corroded leading to some tuning problems. <a href="http://posterous.com/getfile/files.posterous.com/nicklong/yqztofCmBcBCqvFmfisHGDvgzEsgojjraslzFCIzsmpbervcmmsuslvfpEua/media_httpi19photobuc_dsCyF.jpg.scaled1000.jpg"><img alt="Media_httpi19photobuc_dscyf" height="667" src="http://www.nick-long.com/wp-content/uploads/2011/04/media_httpi19photobuc_dsCyF.jpg.scaled500.jpg" width="500" /></a> Pleasingly it worked first time once I got the strings back on. I recorded some tone samples before I changed the pickups so for my next post I'm going to give some sound clips so you can judge for yourself if you think it sounds better but I'm very impressed so far. 
87.33871
749
0.789658
eng_Latn
0.962668
87a0795491c471ebc9d0d7773ac3e260d2e45d64
2,665
md
Markdown
windows-driver-docs-pr/gnss/gnss-driver-design.md
AnLazyOtter/windows-driver-docs.zh-cn
bdbf88adf61f7589cde40ae7b0dbe229f57ff0cb
[ "CC-BY-4.0", "MIT" ]
null
null
null
windows-driver-docs-pr/gnss/gnss-driver-design.md
AnLazyOtter/windows-driver-docs.zh-cn
bdbf88adf61f7589cde40ae7b0dbe229f57ff0cb
[ "CC-BY-4.0", "MIT" ]
null
null
null
windows-driver-docs-pr/gnss/gnss-driver-design.md
AnLazyOtter/windows-driver-docs.zh-cn
bdbf88adf61f7589cde40ae7b0dbe229f57ff0cb
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: GNSS 驱动程序设计 description: 讨论开发适用于 Windows 10 的 GNSS 驱动程序时要考虑的设计原则,其中包括数据结构、错误报告和驱动程序版本控制。 ms.assetid: E10B1149-CC8B-438D-B537-258F7FCFA0E7 ms.date: 04/20/2017 ms.localizationpriority: medium ms.openlocfilehash: e9f437a8e9da3712514af948767257d8e094e19b ms.sourcegitcommit: 4b7a6ac7c68e6ad6f27da5d1dc4deabd5d34b748 ms.translationtype: MT ms.contentlocale: zh-CN ms.lasthandoff: 10/24/2019 ms.locfileid: "72825087" --- # <a name="gnss-driver-design"></a>GNSS 驱动程序设计 讨论开发适用于 Windows 10 的 GNSS 驱动程序时要考虑的设计原则,其中包括数据结构、错误报告和驱动程序版本控制。 ## <a name="data-structures"></a>数据结构 为了实现向后兼容性和将来的扩展性,所有数据结构都以版本号和大小开始,以适应将来的扩展和向后兼容性问题。 作为附加的安全措施,每个结构还具有一个填充缓冲区,以便即使在添加新字段时,静态结构大小仍保持不变。 这是为了使用结构的静态编译时大小(使用**sizeof**)而不是结构的动态大小,来防止任何较旧的 GNSS 驱动程序出现错误。 除非另外指定,否则所有参数都将遵循国际系统(SI): <table> <colgroup> <col width="50%" /> <col width="50%" /> </colgroup> <thead> <tr class="header"> <th>参数</th> <th>计算</th> </tr> </thead> <tbody> <tr class="odd"> <td><p>距离、阈值或级别</p></td> <td><p><strong>进度表</strong></p></td> </tr> <tr class="even"> <td><p>超时或间隔</p></td> <td><p><strong>数</strong></p></td> </tr> <tr class="odd"> <td><p>速度</p></td> <td><p><strong>计量/秒</strong></p></td> </tr> </tbody> </table> ## <a name="error-reporting"></a>错误报告 GNSS DDI 需要将**NTSTATUS**作为驱动程序的返回值。 高级操作系统(HLOS)基于这些错误消息仅对成功和失败情况进行操作,并且不会查看特定错误消息。 但最好是驱动程序返回的错误与相应的**NTSTATUS**错误消息紧密映射。 GNSS 驱动程序可以发送其自己的自定义**NTSTATUS**错误消息,这些错误消息对于诊断目的可能很有用。 ## <a name="driver-versioning"></a>驱动程序版本控制 为 GNSS DDI 指定的每个结构都包含 "驱动程序版本" 字段,许多结构包含一个填充字段。 这两个组件都用于使用以下策略来缓解新版本的 GNSS DDI: - 框架和驱动程序使用功能交换过程来传达各自的版本。 这些 IOCTLs 被认为是特别的,因为它们使用版本字段来传达其版本。 因此,涉及设备和平台功能检查的实现应显式检查第一次返回的版本,并将其存储起来供以后使用。 [**GNSS\_设备\_功能**](https://docs.microsoft.com/windows-hardware/drivers/ddi/gnssdriver/ns-gnssdriver-gnss_device_capability)结构的版本成员会传达驱动程序的版本号。 [**GNSS\_平台\_功能**](https://docs.microsoft.com/windows-hardware/drivers/ddi/gnssdriver/ns-gnssdriver-gnss_platform_capability)结构的版本成员与 GNSS 适配器的版本号通信。 - 当添加新字段时,如果该结构包含填充字段,则应使用空白填充空间,而不是将其添加到结构中,这将保持二进制兼容性 - 每当添加新字段时,GNSS DDI 的版本将被视为递增。 这会在 GNSS DDI 标头本身的注释中反映出来,但不会公开为常量。 GNSS 适配器和 GNSS 驱动程序都将使用私有常数值来指示其当前版本。 这允许 GNSS 适配器和驱动程序都针对特定版本进行编码。 - GNSS 适配器必须与 GNSS 驱动程序的较旧版本向后兼容。 如果新版本的 DDI 中引入了协议更改,则符合新的 GNSS DDI 的 GNSS 适配器必须仅为新版驱动程序实现新协议,并为旧版本的驱动程序使用旧协议。 - GNSS 驱动程序必须与较新版本的 GNSS 适配器向前兼容,并应以其编码的当前版本相同的方式来处理 GNSS 适配器的新版本。 - 使用较新版本的 GNSS 驱动程序时,GNSS 适配器的较旧版本不应正常工作。 为了促进 GNSS 适配器和 GNSS 驱动程序与新版本 DDI 的共同开发,GNSS 适配器中将不会有任何严格版本检查来阻止新的 GNSS 驱动程序。 但是,针对 DDI 的较新版本而实现的 GNSS 驱动程序将不会发送到零售设备,其中包含针对旧版本 GNSS DDI 实现的 GNSS 适配器。 - GNSS 适配器将不支持任何 Windows 8.1 或较早的 GNSS 传感器驱动程序。 这些驱动程序将通过旧堆栈继续在 Windows 10 中运行。 如果存在另一个 Windows 10 GNSS 驱动程序,则不确定旧的 GNSS 传感器驱动程序的使用情况。
30.988372
402
0.768856
yue_Hant
0.818555
87a10f82f6d886e34e491f4e8199e0ea2a5d52b2
1,006
markdown
Markdown
_product/wellspaces-medco-ampera.markdown
kemtol/rocktokom
e5b02438ac464bc137a75d1a2b025d9c58085a11
[ "MIT" ]
null
null
null
_product/wellspaces-medco-ampera.markdown
kemtol/rocktokom
e5b02438ac464bc137a75d1a2b025d9c58085a11
[ "MIT" ]
null
null
null
_product/wellspaces-medco-ampera.markdown
kemtol/rocktokom
e5b02438ac464bc137a75d1a2b025d9c58085a11
[ "MIT" ]
null
null
null
--- title: wellspaces Medco Ampera date: 2019-12-10 11:57:00 +07:00 published: false categories: - office seotags: - ampera, kemang, jakarta, selatan, office space, coworking space, serviced office, tb simatupang, pasar minggu address: Jl. Ampera Raya No.18 - 20, RT.12/RW.2, Cilandak Tim., Kec. Ps. Minggu, Kota Jakarta Selatan, Daerah Khusus Ibukota Jakarta 12560 pricedesc: 'Not only we offer you the best hospitality as best as we could, but also the experience that suits your needs. Pro tips: don''t forget to check out our special deals in whats on section!' price: Shared Desks: Rp 1.500.000/mo mapcoordinate: <iframe src="https://www.google.com/maps/embed?pb=!1m18!1m12!1m3!1d3965.824042810883!2d106.81623531451197!3d-6.286845663285546!2m3!1f0!2f0!3f0!3m2!1i1024!2i768!4f13.1!3m3!1m2!1s0x2e69f1429d6f54b5%3A0xfea96be28cf9dba7!2sMedco%20Ampera!5e0!3m2!1sen!2sid!4v1575953954100!5m2!1sen!2sid" width="600" height="450" frameborder="0" style="border:0;" allowfullscreen=""></iframe> ---
47.904762
297
0.763419
eng_Latn
0.416611
87a18dcee8244fabcd5f5260f7146088c5148b0d
2,294
md
Markdown
DEVELOPMENT.md
yusufgurdogan/yusufgurdogan.github.io
69efbb3096e0a7f6de18108ec7f14cbac2886d4a
[ "MIT" ]
null
null
null
DEVELOPMENT.md
yusufgurdogan/yusufgurdogan.github.io
69efbb3096e0a7f6de18108ec7f14cbac2886d4a
[ "MIT" ]
null
null
null
DEVELOPMENT.md
yusufgurdogan/yusufgurdogan.github.io
69efbb3096e0a7f6de18108ec7f14cbac2886d4a
[ "MIT" ]
1
2020-11-21T13:03:56.000Z
2020-11-21T13:03:56.000Z
# Nault Development ## Application Structure - [Nault](https://github.com/Nault/Nault) - The main wallet application (UI + Seed Generation/Block Signing/Etc). - Communication with the network is done via Badem RPC and Websocket protocols, private or public on any nano network. ## Development Prerequisites - [NodeJS](https://nodejs.org) v12.x + NPM v6.x - Angular CLI: `npm install -g @angular/cli` ## Development Guide #### Clone repository and install dependencies ```bash git clone https://github.com/Nault/Nault cd Nault npm install ``` #### Run the wallet in dev mode ```bash npm run wallet:dev ``` If you want to debug in VS code, first install [debugger for chrome](https://marketplace.visualstudio.com/items?itemName=msjsdiag.debugger-for-chrome) Then you can just go to the debug screen and choose "Launch Chrome http" #### Run the wallet in dev mode as https (for example if using the Ledger device) ```bash npm run wallet:dev-ssl ``` To debug in VS code: Go to debug screen and choose "Launch Chrome https" ## Build Wallet (For Production) Build a production version of the wallet for web: ```bash npm run wallet:build ``` Build a production version of the wallet for desktop: *(Required for all desktop builds)* ```bash npm run wallet:build-desktop ``` ## Desktop Builds *All desktop builds require that you have built a desktop version of the wallet before running!* Run the desktop wallet in dev mode: ```bash npm run desktop:dev ``` If you want to debug in VS code, first install [debugger for chrome](https://marketplace.visualstudio.com/items?itemName=msjsdiag.debugger-for-chrome) Then you can just go to the debug screen and choose "Electron: Main", "Electron: Renderer", or "Electron: All" for both Main and Renderer threads. Build the desktop wallet for your local OS (Will be in `desktop-app\build`): ```bash npm run desktop:local ``` Build the desktop wallet for Windows+Mac+Linux (May require dependencies for your OS [View them here](https://www.electron.build/multi-platform-build)): ```bash npm run desktop:full ``` ## Running unit tests Run `ng test` to execute the unit tests via [Karma](https://karma-runner.github.io). ## Running end-to-end tests Run `ng e2e` to execute the end-to-end tests via [Protractor](http://www.protractortest.org/).
30.586667
152
0.745423
eng_Latn
0.776479
87a1acb5c410ce99325e1bb6ee538c3cc7aac96c
883
md
Markdown
README.md
SimpleProgramming/springboot-redis-database-and-cache
5b5c394b074879659993ec6cecaa8c2019c59521
[ "MIT" ]
4
2020-03-27T19:58:37.000Z
2022-03-26T15:02:43.000Z
README.md
SimpleProgramming/springboot-redis-database-and-cache
5b5c394b074879659993ec6cecaa8c2019c59521
[ "MIT" ]
null
null
null
README.md
SimpleProgramming/springboot-redis-database-and-cache
5b5c394b074879659993ec6cecaa8c2019c59521
[ "MIT" ]
7
2020-06-07T17:37:58.000Z
2022-03-26T15:02:46.000Z
# Spring Boot with Redis as Database and Cache Redis is an open source in-memory data structure store, used as a database, cache and message broker It supports data structures such as strings, hashes, lists, sets, sorted sets, etc # Prerequisites Install Redis and Start Redis Server(https://redis.io/download) # Redis Database For the Database, this example uses HashOperations. It is a redis construct, which can be used for reds hash operations. If you would like to see what are the different operations provided by spring data redis, please a look at the spring data redis documentation # Redis Cache - @EnableCaching - Enable Caching on Spring Boot Application - @Cacheable - Cache the select operation queries. Used with @GetMapping - @CachePut - Update the Cache. Used with @PutMapping - @CacheEvict - Delete the Cache. Used with @DeleteMapping
38.391304
125
0.764439
eng_Latn
0.986502
87a2c46282a5d812a99bc0a22cba3c969fb26350
1,177
md
Markdown
microsoft-edge/hosting/chakra-hosting/jsgetownpropertynames-function.md
OfficeGlobal/edge-developer.ja-JP
a18770d893b61b6263f41d4cb9bfe253bfcc7dba
[ "CC-BY-4.0", "MIT" ]
null
null
null
microsoft-edge/hosting/chakra-hosting/jsgetownpropertynames-function.md
OfficeGlobal/edge-developer.ja-JP
a18770d893b61b6263f41d4cb9bfe253bfcc7dba
[ "CC-BY-4.0", "MIT" ]
null
null
null
microsoft-edge/hosting/chakra-hosting/jsgetownpropertynames-function.md
OfficeGlobal/edge-developer.ja-JP
a18770d893b61b6263f41d4cb9bfe253bfcc7dba
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- description: オブジェクトのすべてのプロパティの一覧を取得します。 title: JsGetOwnPropertyNames 関数 |Microsoft ドキュメント ms.custom: '' ms.date: 01/18/2017 ms.prod: microsoft-edge ms.reviewer: '' ms.suite: '' ms.tgt_pltfrm: '' ms.topic: reference f1_keywords: - jsrt/JsGetOwnPropertyNames helpviewer_keywords: - JsGetOwnPropertyNames function ms.assetid: 0c17ea06-b17f-4ea3-ad04-1259a4d1b6de caps.latest.revision: 12 author: MSEdgeTeam ms.author: msedgedevrel manager: '' ms.openlocfilehash: 5355caab864724d72fdb2c7abb3dc70e39662b55 ms.sourcegitcommit: 6860234c25a8be863b7f29a54838e78e120dbb62 ms.translationtype: MT ms.contentlocale: ja-JP ms.lasthandoff: 04/09/2020 ms.locfileid: "10570450" --- # JsGetOwnPropertyNames 関数 オブジェクトのすべてのプロパティの一覧を取得します。 ## 構文 ```cpp STDAPI_(JsErrorCode) JsGetOwnPropertyNames( _In_ JsValueRef object, _Out_ JsValueRef *propertyNames ); ``` #### パラメーター `object` プロパティ名を取得する対象のオブジェクト。 `propertyNames` プロパティ名の配列。 ## 戻り値 `JsNoError`操作が正常に完了した場合は、エラーコードは返されません。 ## 注釈 アクティブなスクリプトコンテキストが必要です。 ## 要件 **ヘッダー:** jsrt ## 関連項目 [リファレンス (JavaScript Runtime)](../chakra-hosting/reference-javascript-runtime.md)
21.017857
81
0.746814
yue_Hant
0.781653
87a2d87f87b8f29e62b15a339734756e5ad7cc53
245
md
Markdown
pages/common/wordgrinder.md
mstruebing/tldr-pages
7f317713a7307c440bb6115669e23e5006af54db
[ "MIT" ]
1
2019-05-31T17:18:29.000Z
2019-05-31T17:18:29.000Z
pages/common/wordgrinder.md
mstruebing/tldr-pages
7f317713a7307c440bb6115669e23e5006af54db
[ "MIT" ]
null
null
null
pages/common/wordgrinder.md
mstruebing/tldr-pages
7f317713a7307c440bb6115669e23e5006af54db
[ "MIT" ]
null
null
null
# wordgrinder > Command-line word processor. > Homepage: <https://cowlark.com/wordgrinder>. - Start wordgrinder (loads a blank document by default): `wordgrinder` - Open a given file: `wordgrinder {{filename}}` - Show the menu: `Alt + M`
14.411765
56
0.702041
eng_Latn
0.332967
87a2e4ebeb85bb1fe949aa5d41b6b9847751b324
151
md
Markdown
helm/helm-external-dns/README.md
inovex/open-kubeyard
91dff9d81899330c0fffc291304a19042c077fe4
[ "Apache-2.0" ]
11
2017-09-27T11:40:09.000Z
2018-06-14T18:37:46.000Z
helm/helm-external-dns/README.md
inovex/open-kubeyard
91dff9d81899330c0fffc291304a19042c077fe4
[ "Apache-2.0" ]
1
2020-01-28T21:46:09.000Z
2020-01-28T21:46:09.000Z
helm/helm-external-dns/README.md
inovex/open-kubeyard
91dff9d81899330c0fffc291304a19042c077fe4
[ "Apache-2.0" ]
6
2018-01-07T03:39:24.000Z
2018-08-16T10:49:13.000Z
Built as in [Setting up ExternalDNS on Google Container Engine](https://github.com/kubernetes-incubator/external-dns/blob/master/docs/tutorials/gke.md)
151
151
0.821192
kor_Hang
0.167743
87a30e1ca0ca0b6bca5905998e3be3bdd512baff
5,648
md
Markdown
packages/vtree/CHANGELOG.md
danhigham/OpenJSCAD.org
d57e0a4a6d5dc75ebbdbc5597a3df2f5ad12c7ef
[ "MIT" ]
1,435
2017-05-09T01:51:20.000Z
2022-03-31T12:35:31.000Z
packages/vtree/CHANGELOG.md
danhigham/OpenJSCAD.org
d57e0a4a6d5dc75ebbdbc5597a3df2f5ad12c7ef
[ "MIT" ]
634
2017-05-07T11:32:09.000Z
2022-03-30T22:44:27.000Z
packages/vtree/CHANGELOG.md
danhigham/OpenJSCAD.org
d57e0a4a6d5dc75ebbdbc5597a3df2f5ad12c7ef
[ "MIT" ]
313
2017-05-09T21:42:26.000Z
2022-03-31T16:37:23.000Z
# Change Log All notable changes to this project will be documented in this file. See [Conventional Commits](https://conventionalcommits.org) for commit guidelines. ## [2.0.13](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.12...@jscad/vtree@2.0.13) (2021-12-26) **Note:** Version bump only for package @jscad/vtree ## [2.0.12](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.11...@jscad/vtree@2.0.12) (2021-12-11) **Note:** Version bump only for package @jscad/vtree ## [2.0.11](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.10...@jscad/vtree@2.0.11) (2021-11-07) **Note:** Version bump only for package @jscad/vtree ## [2.0.10](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.9...@jscad/vtree@2.0.10) (2021-10-17) **Note:** Version bump only for package @jscad/vtree ## [2.0.9](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.8...@jscad/vtree@2.0.9) (2021-10-04) **Note:** Version bump only for package @jscad/vtree ## [2.0.8](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.7...@jscad/vtree@2.0.8) (2021-09-27) **Note:** Version bump only for package @jscad/vtree ## [2.0.7](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.6...@jscad/vtree@2.0.7) (2021-09-09) **Note:** Version bump only for package @jscad/vtree ## [2.0.6](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.5...@jscad/vtree@2.0.6) (2021-06-20) **Note:** Version bump only for package @jscad/vtree ## [2.0.5](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.4...@jscad/vtree@2.0.5) (2021-06-11) **Note:** Version bump only for package @jscad/vtree ## [2.0.4](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.3...@jscad/vtree@2.0.4) (2021-06-01) **Note:** Version bump only for package @jscad/vtree ## [2.0.3](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.2...@jscad/vtree@2.0.3) (2021-04-20) **Note:** Version bump only for package @jscad/vtree ## [2.0.2](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.0...@jscad/vtree@2.0.2) (2021-04-17) ### Bug Fixes * **all:** update dependencies ([d8c713a](https://github.com/jscad/OpenJSCAD.org/commit/d8c713a933b97a6d179ed3d3e923e188e334f99e)) ## [2.0.1](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.0...@jscad/vtree@2.0.1) (2021-04-15) ### Bug Fixes * **all:** update dependencies ([d8c713a](https://github.com/jscad/OpenJSCAD.org/commit/d8c713a933b97a6d179ed3d3e923e188e334f99e)) # [2.0.0](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.12...@jscad/vtree@2.0.0) (2021-04-12) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.12](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.11...@jscad/vtree@2.0.0-alpha.12) (2021-03-07) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.11](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.10...@jscad/vtree@2.0.0-alpha.11) (2021-02-07) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.10](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.9...@jscad/vtree@2.0.0-alpha.10) (2021-01-02) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.9](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.8...@jscad/vtree@2.0.0-alpha.9) (2020-12-04) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.8](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.7...@jscad/vtree@2.0.0-alpha.8) (2020-11-07) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.7](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.6...@jscad/vtree@2.0.0-alpha.7) (2020-10-11) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.6](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.5...@jscad/vtree@2.0.0-alpha.6) (2020-09-29) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.5](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.4...@jscad/vtree@2.0.0-alpha.5) (2020-09-28) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.4](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.3...@jscad/vtree@2.0.0-alpha.4) (2020-09-19) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.3](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.2...@jscad/vtree@2.0.0-alpha.3) (2020-09-08) **Note:** Version bump only for package @jscad/vtree # [2.0.0-alpha.2](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.1...@jscad/vtree@2.0.0-alpha.2) (2020-09-02) ### Bug Fixes * **all:** update dependencies ([d8c713a](https://github.com/jscad/OpenJSCAD.org/commit/d8c713a933b97a6d179ed3d3e923e188e334f99e)) # [2.0.0-alpha.1](https://github.com/jscad/OpenJSCAD.org/compare/@jscad/vtree@2.0.0-alpha.0...@jscad/vtree@2.0.0-alpha.1) (2020-08-19) **Note:** Version bump only for package @jscad/vtree # 2.0.0-alpha.0 (2020-08-13) ### Bug Fixes * **vtree:** corrected and completed all modeling primitives ([#577](https://github.com/jscad/OpenJSCAD.org/issues/577)) ([82dab93](https://github.com/jscad/OpenJSCAD.org/commit/82dab93e8ad0d71a325a6ce6dfac5ce19a095984)) * **vtree:** corrected colorize in core API ### Features * **vtree:** Vtree integration into main repo (#498) * **vtree:** overhauled vtree to align with V2 (#507)
23.932203
220
0.6875
kor_Hang
0.305673
87a3854fa45a74c73c3da90344f40eb2278bc289
216
md
Markdown
README.md
IDLabResearch/validation-benchmark
5959ed61154d7e38a86c88e322b6208017d35889
[ "MIT" ]
null
null
null
README.md
IDLabResearch/validation-benchmark
5959ed61154d7e38a86c88e322b6208017d35889
[ "MIT" ]
null
null
null
README.md
IDLabResearch/validation-benchmark
5959ed61154d7e38a86c88e322b6208017d35889
[ "MIT" ]
null
null
null
# Validation Benchmark ## Usage - download the datasets (see `dataset-downloader/README.md`) - run a benchmark of your choice (see `validation-reasoning-framework/README/md` and `rdfunit/README.md`) ## License MIT
24
105
0.75463
eng_Latn
0.494134
87a48f2ed6dd761559d623d1953dc7cbc620c221
3,256
md
Markdown
ch8/ch8-09.md
GameXG/gopl-zh
c8594b338c76ace219f66d532518ca0ba8a1f287
[ "BSD-3-Clause" ]
5
2016-01-20T07:49:33.000Z
2016-10-06T06:57:36.000Z
ch8/ch8-09.md
GameXG/gopl-zh
c8594b338c76ace219f66d532518ca0ba8a1f287
[ "BSD-3-Clause" ]
null
null
null
ch8/ch8-09.md
GameXG/gopl-zh
c8594b338c76ace219f66d532518ca0ba8a1f287
[ "BSD-3-Clause" ]
7
2016-02-04T04:40:15.000Z
2021-08-13T13:56:18.000Z
## 8.9. 併發的退出 有時候我們需要通知goroutine停止它正在榦的事情,比如一個正在執行計算的web服務,然而它的客戶端已經斷開了和服務端的連接。 Go語言併沒有提供在一個goroutine中終止另一個goroutine的方法,由於這樣會導致goroutine之間的共享變量落在未定義的狀態上。在8.7節中的rocket launch程序中,我們往名字叫abort的channel里發送了一個簡單的值,在countdown的goroutine中會把這個值理解爲自己的退出信號。但是如果我們想要退出兩個或者任意多個goroutine怎麽辦呢? 一種可能的手段是向abort的channel里發送和goroutine數目一樣多的事件來退出它們。如果這些goroutine中已經有一些自己退出了,那麽會導致我們的channel里的事件數比goroutine還多,這樣導致我們的發送直接被阻塞。另一方面,如果這些goroutine又生成了其它的goroutine,我們的channel里的數目又太少了,所以有些goroutine可能會無法接收到退出消息。一般情況下我們是很難知道在某一個時刻具體有多少個goroutine在運行着的。另外,當一個goroutine從abort channel中接收到一個值的時候,他會消費掉這個值,這樣其它的goroutine就沒法看到這條信息。爲了能夠達到我們退出goroutine的目的,我們需要更靠譜的策略,來通過一個channel把消息廣播出去,這樣goroutine們能夠看到這條事件消息,併且在事件完成之後,可以知道這件事已經發生過了。 迴憶一下我們關閉了一個channel併且被消費掉了所有已發送的值,操作channel之後的代碼可以立卽被執行,併且會産生零值。我們可以將這個機製擴展一下,來作爲我們的廣播機製:不要向channel發送值,而是用關閉一個channel來進行廣播。 隻要一些小脩改,我們就可以把退出邏輯加入到前一節的du程序。首先,我們創建一個退出的channel,這個channel不會向其中發送任何值,但其所在的閉包內要寫明程序需要退出。我們同時還定義了一個工具函數,cancelled,這個函數在被調用的時候會輪詢退出狀態。 <u><i>gopl.io/ch8/du4</i></u> ```go var done = make(chan struct{}) func cancelled() bool { select { case <-done: return true default: return false } } ``` 下面我們創建一個從標準輸入流中讀取內容的goroutine,這是一個比較典型的連接到終端的程序。每當有輸入被讀到(比如用戶按了迴車鍵),這個goroutine就會把取消消息通過關閉done的channel廣播出去。 ```go // Cancel traversal when input is detected. go func() { os.Stdin.Read(make([]byte, 1)) // read a single byte close(done) }() ``` 現在我們需要使我們的goroutine來對取消進行響應。在main goroutine中,我們添加了select的第三個case語句,嚐試從done channel中接收內容。如果這個case被滿足的話,在select到的時候卽會返迴,但在結束之前我們需要把fileSizes channel中的內容“排”空,在channel被關閉之前,舍棄掉所有值。這樣可以保證對walkDir的調用不要被向fileSizes發送信息阻塞住,可以正確地完成。 ```go for { select { case <-done: // Drain fileSizes to allow existing goroutines to finish. for range fileSizes { // Do nothing. } return case size, ok := <-fileSizes: // ... } } ``` walkDir這個goroutine一啟動就會輪詢取消狀態,如果取消狀態被設置的話會直接返迴,併且不做額外的事情。這樣我們將所有在取消事件之後創建的goroutine改變爲無操作。 ```go func walkDir(dir string, n *sync.WaitGroup, fileSizes chan<- int64) { defer n.Done() if cancelled() { return } for _, entry := range dirents(dir) { // ... } } ``` 在walkDir函數的循環中我們對取消狀態進行輪詢可以帶來明顯的益處,可以避免在取消事件發生時還去創建goroutine。取消本身是有一些代價的;想要快速的響應需要對程序邏輯進行侵入式的脩改。確保在取消發生之後不要有代價太大的操作可能會需要脩改你代碼里的很多地方,但是在一些重要的地方去檢査取消事件也確實能帶來很大的好處。 對這個程序的一個簡單的性能分析可以揭示瓶頸在dirents函數中獲取一個信號量。下面的select可以讓這種操作可以被取消,併且可以將取消時的延遲從幾百毫秒降低到幾十毫秒。 ```go func dirents(dir string) []os.FileInfo { select { case sema <- struct{}{}: // acquire token case <-done: return nil // cancelled } defer func() { <-sema }() // release token // ...read directory... } ``` 現在當取消發生時,所有後台的goroutine都會迅速停止併且主函數會返迴。當然,當主函數返迴時,一個程序會退出,而我們又無法在主函數退出的時候確認其已經釋放了所有的資源(譯註:因爲程序都退出了,你的代碼都沒法執行了)。這里有一個方便的竅門我們可以一用:取代掉直接從主函數返迴,我們調用一個panic,然後runtime會把每一個goroutine的棧dump下來。如果main goroutine是唯一一個剩下的goroutine的話,他會清理掉自己的一切資源。但是如果還有其它的goroutine沒有退出,他們可能沒辦法被正確地取消掉,也有可能被取消但是取消操作會很花時間;所以這里的一個調研還是很有必要的。我們用panic來獲取到足夠的信息來驗證我們上面的判斷,看看最終到底是什麽樣的情況。 **練習 8.10:** HTTP請求可能會因http.Request結構體中Cancel channel的關閉而取消。脩改8.6節中的web crawler來支持取消http請求。(提示:http.Get併沒有提供方便地定製一個請求的方法。你可以用http.NewRequest來取而代之,設置它的Cancel字段,然後用http.DefaultClient.Do(req)來進行這個http請求。) **練習 8.11:** 緊接着8.4.4中的mirroredQuery流程,實現一個併發請求url的fetch的變種。當第一個請求返迴時,直接取消其它的請求。
36.58427
416
0.804668
yue_Hant
0.964781
87a55de3f251f027d062f6cdc94093816e997327
15,264
md
Markdown
posts/65-learning-trap/index.md
jkarteaga/jonkoldoarteaga2
9ef7786073c1f586b91b967c67f9706102c77ee1
[ "MIT" ]
null
null
null
posts/65-learning-trap/index.md
jkarteaga/jonkoldoarteaga2
9ef7786073c1f586b91b967c67f9706102c77ee1
[ "MIT" ]
null
null
null
posts/65-learning-trap/index.md
jkarteaga/jonkoldoarteaga2
9ef7786073c1f586b91b967c67f9706102c77ee1
[ "MIT" ]
null
null
null
--- date: 2016-02-26 title: > The Learning Trap: New knowledge makes me act like a dick. slug: learning-trap description: > We all hate that smug asshole who tries to force an opinion on us. So why is it so easy to BECOME that smug asshole when we learn new things? category: - acting-like-a-grown-up tag: - judgment - learning images: - /images/the-learning-trap.jpg - /images/everything-i-own-hermit-crab.jpg cta: focus --- I love learning. I want to know everything. I deeply believe curiosity is what will eventually fix all the world's problems. When I meet new people, I consider their curiosity as the biggest factor in whether or not I want to spend more time around them. But there's a catch. New knowledge has an unintended side effect I can't seem to shake: **learning new things always turns me into a gaping asshole.** I call it The Learning Trap. ## The Standard Learning Progression First, some context. According to a [theory developed by Noel Burch](https://en.wikipedia.org/wiki/Four_stages_of_competence), there are four stages of learning: 1. **Unconscious Incompetence.** We're bad at a thing, but we don't care because we don't know the thing exists in the first place. This is the "ignorance is bliss" part. 2. **Conscious Incompetence.** We're made aware that we are bad at the thing. Many of us stop here and adopt a this-is-stupid-and-anyone-who-disagrees-with-me-is-also-stupid attitude.[^everything-im-not-good-at-is-dumb] 3. **Conscious Competence.** With lots of practice and high levels of concentration, we can do the thing at a fair-to-middling level. This is how I feel about folding long-sleeved shirts. 4. **Unconscious Competence.** Also known as mastery, where we can do the thing without thinking about it. Easy examples include walking and speaking our native language. [^everything-im-not-good-at-is-dumb]: "Rather than learn this thing, I shall publicly deride it, and adopt an air of smug superiority to shame its participants." This is a generally-accepted representation of the learning process, as hashed out by people much smarter than me, based on years of research. ## The Standard Learning Progression (Adjusted for My Shitty Attitude) The Learning Trap lurks inside a different progression, however — one that [Nate](http://thenategreenexperience.com) and I noticed as we've observed our own behavior while learning: 1. **Clueless and Happy.** I have no idea that I'm stupid.[^ninety-nine] 2. **Self-Conscious and Defensive.** I am now aware that I'm stupid, and I'm pissed off that you pointed it out. 3. **Marginally Capable and Wholly Condescending.** _(a.k.a. The Learning Trap.)_ I know exactly enough to fake like I know what I'm talking about. It is now my mission to — in a condescending, self-congratulatory manner — ensure everyone around me is Self-Conscious and Defensive about this subject. 4. **Experienced and Unconcerned.** I know the topic well enough to realize there's no need for (and, really, no benefit to) forcing my knowledge on anyone. [^ninety-nine]: This is the stage I'm currently in for roughly 99% of everything there is to know. It's on my bucket list to get that number down to 98% before I die. Over drinks with [Phil](http://www.precisionnutrition.com/about/phil-caravaggio) last year, I put the stages another way: <blockquote class="twitter-tweet" lang="en"><p lang="en" dir="ltr">The path of learning according to <a href="https://twitter.com/jlengstorf">@jlengstorf</a>: know nothing &gt; know a little &gt; insufferable asshole &gt; wisdom.</p>&mdash; Phil Caravaggio (@philcaravaggio) <a href="https://twitter.com/philcaravaggio/status/609198781154902016">June 12, 2015</a></blockquote> ## The Learning Trap: Travel Edition In mid-2014 I knew nothing about world travel — full-on _Clueless and Happy_. I'd taken two short European vacations, during which I'd been to Ireland and the UK twice, plus a couple days spent in Amsterdam and Belgium.[^english] Outside of those two trips, my only travel outside the US had been a couple teenaged beer runs to Calgary. [^english]: Worth noting is that I avoided any country where English wasn't more or less a first language. Then I committed myself to [spend a year outside the United States](/remote-work-travel) — and smashed into _Self-Conscious and Defensive_ ego-first. Naturally, I'd run my mouth off about the trip to almost everyone I knew. **People asked questions, and I wanted to punch every curious friend in the throat for drawing attention to my lack of answers.** In the months between booking my ticket and boarding the plane, I read everything I could find by other travelers to fill in the gaps in my knowledge. I knew I'd need practical knowledge eventually, but I was happy to build a theoretical base in the meantime. <figure class="figure figure--center"> <img src="./images/the-learning-trap.jpg" alt="The Learning Trap." /> </figure> The more I read, the more confident I became. I eased into _Marginally Capable and Wholly Condescending_ like a comfortably broken-in pair of shoes.[^dunning-kruger] [^dunning-kruger]: There's some research supporting this. The [Dunning–Kruger effect](https://en.wikipedia.org/wiki/Dunning%E2%80%93Kruger_effect) describes the overconfidence felt by people with low competence. If you're borderline incompetent, you're prone to overestimate your own skill. You're also more likely to dismiss the expertise of others. I was in The Learning Trap. Again. **Armed with unverified research and limitless zeal, I unleashed an awareness campaign on everyone unfortunate enough to fall within earshot.** "[Travel is cheaper than a lease](/cost-of-living-remotely)," I cried from my high horse. "It's easier than ever to [become location independent](/how-to-become-location-independent)! Don't you understand? [_Remote work is the future!_](/remote-work-everyone-wins) You have to [stop overworking or you're going to die](/overkill-cult)!" And I didn't care that most people don't have any interest in selling everything they can't carry and living like a nomadic hermit crab, because I was galvanized by how un-fucking-believably _correct_ I was.[^discourse] [^discourse]: **How to solve all the world's problems, by Jason Lengstorf:** The breakdown in public discourse today — about politics, books, music, or whatever — happens because most of us spend the majority of our time permanently mired in the _Marginally Capable and Wholly Condescending_ stage of learning. There's just Too Much Stuff™ to know, so we comfort ourselves by knowing a little bit. And knowing a little bit _is_ better than knowing nothing. _Unless knowing a little bit somehow convinces us that we know everything and should be considered experts when the time for important decisions arises._ Unfortunately, because we all know a little, we're convinced of our Indisputable Rightness, and we either shout into an echo chamber of other people who have the same limited knowledge and who agree with us because they, too, are Indisputably Right; or we hit the brick wall of an opposing opinion that, like us, has (different) limited information and is similarly convinced of its Incontrovertible Correctness. The American media today is a great real-world example of what happens when an unstoppable force meets and immovable object: shit goes sideways. There are no experts anymore, because we burn them at the stake for questioning our Indisputable Rightness. Instead we sink to partially-informed mob justice because our tiny bit of knowledge convinces us that we, too, have Valuable Opinions, and since our Valuable Opinions stem from Indisputable Rightness, the experts are clearly in the pocket of Big Wrong and we can safely ignore them. There is no discourse anymore. There are only the Indisputably Right and the Incontrovertibly Correct, and neither side seems willing to do further research to recognize that maybe — just maybe — we're all working toward the same goals from different ends of the spectrum, and if we were to actually read the article attached to that headline that launched our most recent Twitter rant, we might realize both viewpoints arrive at somewhat compatible conclusions. Where public discourse would really benefit is if the semi-informed — which is you and me in roughly 99% of issues being discussed — would just _shut the fuck up_ and listen to people who have dedicated their lives to learning the things being discussed. If we're talking about software or travel or small businesses, I have lots of fairly well-informed opinions. But I don't know that I could even explain what macroeconomics _is_, let alone make informed choices about what's best for global trade. (Macroeconomics has something to do with global trade, right? I'm making that assumption because "macro" means big, and global trade affects the economy and is also big. I refuse to look this up because I think it's doing a pretty good job of demonstrating my point right now.) I want my opinion to matter. But sometimes it doesn't. And as much as it pains me to admit it, sometimes I need to sit quietly and let someone more knowledgeable do what's best — or start studying hard so I can make an _informed_ contribution. <figure class="figure figure--center"> <img src="./images/everything-i-own-hermit-crab.jpg" alt="Everything I Own Hermit Crab." /> <figcaption class="figure__caption"> You know what’s fun? Carrying everything you own with you at all times! <small class="figure__attribution"> Credit: <a class="figure__attribution-link" href="https://www.flickr.com/photos/verzo/9353230567"> Roberto Verzo </a> </small> </figcaption> </figure> For the record, I still _believe_ all those things, and I'm still advocating them as much as possible. But now that I've clawed my way out of The Learning Trap again, I'm far less likely to harangue a stranger in the line for coffee. ### How dare you not know that thing I just learned a few minutes ago? My pedagogical reign of terror was the primary symptom of The Learning Trap. **This happens when we 1) learn something, then 2) can't believe how we possibly never knew it before, and 3) immediately look with derision toward others who don't possess the knowledge we gained a few seconds earlier.** <blockquote class="twitter-tweet" lang="en"><p lang="en" dir="ltr">8:02: Learn new information.&#10;8:07: Overhear discussion re: new information.&#10;8:08: Condescendingly offer opinion formed six minutes earlier.</p>&mdash; Jason Lengstorf (@jlengstorf) <a href="https://twitter.com/jlengstorf/status/674980634490626049">December 10, 2015</a></blockquote> I hadn't even boarded my first flight, but I already felt superior to everyone around me who didn't know how great travel was. ## Climbing Out of the Trap It's been [over a year since I started traveling](/one-year-of-world-travel); most of the glitter has washed off at this point. I don't look at travel as an indicator of whether or not someone is enlightened anymore — that's something my embarrassingly elitist self from a year ago would think. Instead, I look at travel the way I used to look at living in Portland vs. living in Montana: some people prefer cities because they want more and better restaurants, better walkability, or more nightlife; some people prefer small towns because they like knowing everyone they pass on the street by name, quiet nights with endless, starry skies, or having a forest in their back yard. Some people find travel exciting and rewarding. Others find it stressful and draining. Neither group is wrong. **There's no such thing as an incorrect personal preference — what _is_ wrong is trying to force a personal preferences down someone else's throat. And that's why The Learning Trap is so hard to escape.** When I was first getting excited about travel, I couldn't imagine a world where someone _wouldn't_ be excited about it. And that excitement made me forceful with my opinions. What I _thought_ I was doing was sharing something amazing with people who may not have realized it yet. What I was _actually_ doing was evangelizing something that, honestly — if someone was interested — they'd've probably already fucking Googled it.[^google] [^google]: And, as a result, they would most likely have the same information I had, because who clicks past the first page of Google results, anyways? ## We Don't Need Consensus Travel was the latest in a long line of Learning Traps I've fallen into and dragged myself out of, dirty and ashamed. In high school I was convinced metal was the only _real_ music,[^filth] and people like Justin Timberlake were everything that was wrong with the world.[^jt] I was once strongly in favor of robot government. For a while I [went evangelical on my dad about coffee](/taste-doesnt-matter). [^filth]: "Seriously, bro, you haven't heard _true_ musical genius until you've heard Danny Filth wax romantic about disembowelment in a falsetto shriek." [^jt]: I don't feel that way anymore, JT. We should hang out. In the _Marginally Informed and Wholly Condescending_ stage, it's easy to start looking at our new ideas as necessary or non-negotiable. But that's less because the idea is _actually_ vital — after all, I lived twenty-nine years blissfully ignorant of the travel lifestyle before I started excoriating others for their ignorance — and more because we're desperately trying to assure ourselves that this new obsession is a good thing. I need you to buy into my travel obsession because I need group acceptance to make it okay for me to believe this. I don't want to be crazy.[^honey] [^honey]: If I found out that rubbing honey on my balls made me happy, I would want everyone else to rub honey on their business, too. I'd tell myself it was because I wanted them to be as happy as I was, but in reality I'd be trying to force other people to validate my non-standard application of honey. Because if they don't, I'm not a visionary; I'm just a weirdo with sticky junk. But with experience — reaching the _Experienced and Unconcerned_ stage — I realized in every case that **consensus of opinion just doesn't matter.** The thing I like works for me — whether or not anyone else agrees. ## Short-Circuit the Cycle I still don't know how to skip the insufferable asshole part of learning something new. If you figure it out, let me know — after you've gotten over the insufferable asshole part of figuring it out, that is. But I try to be more aware of what I'm doing now. I try to withhold my opinion in conversations until I'm asked for it. I try to notice and squash derision before it forms a smirk when someone doesn't know something I've already learned. I try to remember that for every piece of information I have that someone else doesn't, there are ten thousand things I _don't_ know. It hasn't cured me — I mean, shit, I was probably condescending within this post about trying not to be condescending — but the people who've known me longest _have_ commented that I suck a little less to talk to these days. And I'll call that a win.
56.117647
464
0.77044
eng_Latn
0.99947
87a5d8f5c9926c023396db4e4a78b4c4daf3ed7b
7,181
md
Markdown
_posts/2018-09-11-Authorization.md
lakhanmandloi/open-saber-docs
58b1765b135c53ef7dc29bf6fc32ca1b94c55346
[ "MIT" ]
null
null
null
_posts/2018-09-11-Authorization.md
lakhanmandloi/open-saber-docs
58b1765b135c53ef7dc29bf6fc32ca1b94c55346
[ "MIT" ]
null
null
null
_posts/2018-09-11-Authorization.md
lakhanmandloi/open-saber-docs
58b1765b135c53ef7dc29bf6fc32ca1b94c55346
[ "MIT" ]
6
2018-09-10T11:03:24.000Z
2021-02-17T10:51:40.000Z
--- date: 2018-09-10 title: Authorization categories: - Documentation type: Document showSidebar: false published: false nav_ordering: 10 pageTitle: "Authorization" --- ## Overview Every record is made up of `fields`. Fields have metadata, including who is the `owner` of the field and if necessary, a `proxy owner` of the field. By default, a proxy owner can be added to a field only for those fields created during new `record` creation. To access a field, the system checks if the accessor has a `consent artifact` AND if you are the owner of the field for which consent has been given. Consent artifact looks like: ``` id: , # id of the consent user_id: , # which user is given consent OR role_id: , # which role is given consent action: [], # any one of create/read/update/delete fields: [], # field list, * wildcard supported filter: [], # any criteria proxy: , # true if a proxy is using the consent created_by: , # which user created the consent created_at: , # when was it created awarded_by: # which user awarded the consent awarded_at: # when was it awarded expires_at: , # when will it expire nonce: , # with a nonce the consent can only be used once ``` Nothing is accessible without Consent. So, if by chance, all consent artifacts are deleted, instead of opening up, the Registry will be closed for access. Also, one cannot delete a Consent but can only end date it. Roles can be simply understood as a way to connect a group of users to consent. Role would already be setup externally in the Authentication mechanism. OpenSABER would expect the User and Role imformation to be present in the JWT as per the OpenID Connect Flow. ## Ownership assignment Following diagram explains how field ownership mappings are dynamically managed. Since, OpenSABER vocabulary is extendable, it would be hard to setup ownership everytime a new field is introduced. Rather than that, a combination of consent and simple workflow, would assign ownership of record level fields. [![Ownership assignment](/images/ownership-assignment.png)](/images/ownership-assignment.png){:target="_blank"} A Proxy entity is being elevated, since there maybe many cases where registry initiation may happen via a Proxy source than organic. However, once the Proxy has inserted data, control needs to go back to the record owner to prevent misuse by the Proxy. ## Scenarios Let's look at the following scenario's to understand possible authorization flows: ### Scenario #1 - Proxy scenario An entity creating a record where record is not of the entity. Examples: - A head master creating teacher records - A bulk importing mechanism Record creation consents will be awarded to the Proxy (e.g. headmaster who is registering teachers) via a consent artifact such as ``` # custom consent: creator_f1_f2_f3_f4 action: [create], # ACTION SCOPE - creating properties fields: [f1,f2,f3,f4], # PROPERTIES SCOPE - f1,f2,f3,f4 filter: [id==null] # SUBJECTS SCOPE - new subject proxy: true # Indicates that the user with this consent is a proxy so the default rule of field writer becoming the owner is overriden by field writer becoming the proxy. The owner will remain the record owner. ``` The above means that the proxy, can create 4 fields - f1, f2, f3, f4 ONLY when the subject does not exists. It cannot read, update or delete the data. Suppose, the proxy requires to read the fields it proxied, then the following consent is required: ``` # canned consent: read_as_proxy action: [read], # ACTION SCOPE - reading properties fields: [f1,f2,f3,f4], # PROPERTIES SCOPE - f1,f2,f3,f4 filter: [proxy.id==$(userid)] # SUBJECTS SCOPE - proxied subject ``` Consequently, for any entity to access it's own records, the following consent artifact is necessary: ``` # canned consent: rw_as_owner action: [read, update], # ACTION SCOPE - r/w fields: [*], # PROPERTIES SCOPE - f1,f2,f3,f4 filter: [owner.id==$(userid)] # SUBJECTS SCOPE - self ``` The above means that you can read/write all fields for which you are the owner. Of course, if you want to explicitly list fields, it is possible too. ### Scenario #2 - Self registration A user creating the record of itself Example: - self registration Consent is a union of ``` # canned consent: godlike_creator action: [create], # ACTION SCOPE - creating properties fields: [*], # PROPERTIES SCOPE - any property filter: [id==null] # SUBJECTS SCOPE - new subject ``` and ``` # canned consent: rw_as_owner ``` Note that multiple consents stack up and have a default union effect, than most_restrictive, intersection, etc. ### Scenario #3 - Publisher A user adding a new field to an existing record of some other entity Example: - Any publishing entity, for instance, badge awarded to a teacher Consent needed: ``` # canned consent: publish_f5 action: [create,update], # ACTION SCOPE - add or update fields: [f5], # PROPERTIES SCOPE - f5 filter: [] # SUBJECTS SCOPE - any ``` If f5 represented the field into which badge is updated, the above consent grants creation or updation of the badge to the entity which has the consent. Of course, one may restrict a publisher so that it may only target a set of subjects via `filter`. ### Scenario #4 - Reader If a user wishes to grant read access to otherwise private info, she can give consent as: ``` # dynamic consent: consent_hgyi6t3tgjasd98hdgasd action: [read], # ACTION SCOPE - read fields: [f1], # PROPERTIES SCOPE - f1 awarded_by: [shashank@diksha] ``` In this case, `shashank@diksha` is an indicative identifier. And `shashank@diksha` is essentially granting consent to read their data in f1. It is not necessary, that `shashank@diksha` is the `owner of the field`, as it may be written into only by a specific publicher. [Publisher scenario] ## Authorization Negotiation [![Venn Diagram](/images/venn-diagram-consent.png)](/images/venn-diagram-consent.png){:target="_blank"} The above Venn diagram should indicate the various sets of fields possible in a Registry. For instance, in the diagram above, - the registry has fields: `a, b, c, d, e, f, g, h` - the requester requested for: `b, c, d, e, f, g, h` - the requester had bearer consent only for `c`, so `b` was an unauthorised access - the requester needed consent from Owner 1 for `f and g` - the requester needed consent from Owner 2 for `d and e` - Owner 1 gave consent for both `f and g` requested - Owner 2 gave consent only for `d` and rejected consent for `e` - Field `a` and `h` did not need consent since they were not requested. So, they should also not be returned to the user. ## Roles Roles need to be mapped internally within the Registry. ## Open Items - How to revoke a Consent artifact? Look at Open Badge spec. - How to provide grant to another entity so that they may generate a Consent artifact? - Does the ID Token contain Role identifier? ## Dependency ### Authentication This wiki assumes that Authentication step is over and the Current User and the Role information is available. [Authentication wiki](/authentication.html)
46.934641
408
0.736666
eng_Latn
0.998782
87a6c61f3ea8609544a651c324ba2922cf07969d
312
md
Markdown
pages/TextArea/Events.md
battleaxedotco/brutalism-docs
75b75bcb113808fb69db282d3107b59b26cc84ec
[ "MIT" ]
null
null
null
pages/TextArea/Events.md
battleaxedotco/brutalism-docs
75b75bcb113808fb69db282d3107b59b26cc84ec
[ "MIT" ]
6
2020-06-14T06:28:02.000Z
2022-02-27T03:43:18.000Z
pages/TextArea/Events.md
battleaxedotco/brutalism-docs
75b75bcb113808fb69db282d3107b59b26cc84ec
[ "MIT" ]
null
null
null
<TextArea value="@change" @change="message" /> <TextArea value="@update" @update="message" /> <TextArea value="@submit" @submit="message" /> <TextArea value="@prefs" @prefs="testPrefs" prefs-id="exampleTextAreaEvent" /> <TextArea value="@focus/blur" @focus="message('Focused!')" @blur="message('Blur!')" />
34.666667
78
0.676282
yue_Hant
0.128152
87a7ff5cdc965efeecbc926a2ade9e9d5d9bc1ac
1,703
md
Markdown
api/Word.Application.ProtectedViewWindowDeactivate.md
CeptiveYT/VBA-Docs
1d9c58a40ee6f2d85f96de0a825de201f950fc2a
[ "CC-BY-4.0", "MIT" ]
283
2018-07-06T07:44:11.000Z
2022-03-31T14:09:36.000Z
api/Word.Application.ProtectedViewWindowDeactivate.md
CeptiveYT/VBA-Docs
1d9c58a40ee6f2d85f96de0a825de201f950fc2a
[ "CC-BY-4.0", "MIT" ]
1,457
2018-05-11T17:48:58.000Z
2022-03-25T22:03:38.000Z
api/Word.Application.ProtectedViewWindowDeactivate.md
CeptiveYT/VBA-Docs
1d9c58a40ee6f2d85f96de0a825de201f950fc2a
[ "CC-BY-4.0", "MIT" ]
469
2018-06-14T12:50:12.000Z
2022-03-27T08:17:02.000Z
--- title: Application.ProtectedViewWindowDeactivate event (Word) keywords: vbawd10.chm4000035 f1_keywords: - vbawd10.chm4000035 ms.prod: word api_name: - Word.Application.ProtectedViewWindowDeactivate ms.assetid: bd80056b-edce-7e0b-c61a-31ebda24a416 ms.date: 06/08/2017 ms.localizationpriority: medium --- # Application.ProtectedViewWindowDeactivate event (Word) Occurs when a Protected View window is deactivated. ## Syntax _expression_. `ProtectedViewWindowDeactivate`( `_PvWindow_` , ) _expression_ An expression that returns a '[Application](Word.Application.md)' object. ## Parameters |Name|Required/Optional|Data type|Description| |:-----|:-----|:-----|:-----| | _PvWindow_|Required| **[ProtectedViewWindow](Word.ProtectedViewWindow.md)**|The deactivated Protected View window.| ## Example The following code example minimizes an open Protected View window when it is deactivated. This code must be placed in a class module, and an instance of the class must be correctly initialized for this code example to work correctly. For more information about how to do this, see [Using events with the Application object](../word/Concepts/Objects-Properties-Methods/using-events-with-the-application-object-word.md). The following code example assumes that you have declared an application variable called "App" in your general declarations and have set the variable equal to the Word Application object. ```vb Private Sub App_ProtectedViewWindowDeactivate(ByVal PvWindow As ProtectedViewWindow) PvWindow.WindowState = wdWindowStateMinimize End Sub ``` ## See also [Application Object](Word.Application.md) [!include[Support and feedback](~/includes/feedback-boilerplate.md)]
30.410714
419
0.78626
eng_Latn
0.904543
87a8b0777b5b5d77b72ff724e8fc5b609622397a
3,633
md
Markdown
docs/framework/unmanaged-api/profiling/icorprofilerinfo4-enumjitedfunctions2-method.md
lucieva/docs.cs-cz
a688d6511d24a48fe53a201e160e9581f2effbf4
[ "CC-BY-4.0", "MIT" ]
1
2018-12-19T17:04:23.000Z
2018-12-19T17:04:23.000Z
docs/framework/unmanaged-api/profiling/icorprofilerinfo4-enumjitedfunctions2-method.md
lucieva/docs.cs-cz
a688d6511d24a48fe53a201e160e9581f2effbf4
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/unmanaged-api/profiling/icorprofilerinfo4-enumjitedfunctions2-method.md
lucieva/docs.cs-cz
a688d6511d24a48fe53a201e160e9581f2effbf4
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: ICorProfilerInfo4::EnumJITedFunctions2 – metoda ms.date: 03/30/2017 api_name: - ICorProfilerInfo4.EnumJITedFunctions2 api_location: - mscorwks.dll api_type: - COM f1_keywords: - ICorProfilerInfo4::EnumJITedFunctions2 helpviewer_keywords: - EnumJITedFunctions2 method, ICorProfilerInfo4 interface [.NET Framework profiling] - ICorProfilerInfo4::EnumJITedFunctions2 method [.NET Framework profiling] ms.assetid: 40e9a1be-9bd2-4fad-9921-34a84b61c1e3 topic_type: - apiref author: mairaw ms.author: mairaw ms.openlocfilehash: b11c70fa7d423575b2dd1d2cc676908885a0fd4c ms.sourcegitcommit: 3d5d33f384eeba41b2dff79d096f47ccc8d8f03d ms.translationtype: MT ms.contentlocale: cs-CZ ms.lasthandoff: 05/04/2018 ms.locfileid: "33457361" --- # <a name="icorprofilerinfo4enumjitedfunctions2-method"></a>ICorProfilerInfo4::EnumJITedFunctions2 – metoda Vrátí enumerátor pro všechny funkce, které byly dříve kompilována a překompilovat JIT. Nahradí tato metoda [icorprofilerinfo3::enumjitedfunctions –](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo3-enumjitedfunctions-method.md) metodu, která není výčet překompilovat JIT ID. ## <a name="syntax"></a>Syntaxe ``` HRESULT EnumJITedFunctions([out] ICorProfilerFunctionEnum** ppEnum); ``` #### <a name="parameters"></a>Parametry `ppEnum` [out] Ukazatel [icorprofilerfunctionenum –](../../../../docs/framework/unmanaged-api/profiling/icorprofilerfunctionenum-interface.md) enumerátor. ## <a name="remarks"></a>Poznámky Tato metoda se mohou překrývat s `JITCompilation` zpětná volání, jako [icorprofilercallback::jitcompilationstarted –](../../../../docs/framework/unmanaged-api/profiling/icorprofilercallback-jitcompilationstarted-method.md) metoda. Vrácený výčtu obsahuje hodnoty pro `COR_PRF_FUNCTION::reJitId` pole. [Icorprofilerinfo3::enumjitedfunctions –](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo3-enumjitedfunctions-method.md) metodu, která nahradí tato metoda není výčet překompilovat JIT ID, protože `COR_PRF_FUNCTION::reJitId` pole je vždycky nastavený na hodnotu 0. `ICorProfilerInfo4::EnumJITedFunctions` Metoda výčet překompilovat JIT ID, protože `COR_PRF_FUNCTION::reJitId` pole je správně nastaveno. Všimněte si, že [icorprofilerinfo4::enumjitedfunctions2 –](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo4-enumjitedfunctions2-method.md) metoda můžete aktivovat uvolňování paměti, zatímco [icorprofilerinfo3::enumjitedfunctions – metoda](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo3-enumjitedfunctions-method.md) nikoli. Další informace najdete v tématu [CORPROF_E_UNSUPPORTED_CALL_SEQUENCE HRESULT](../../../../docs/framework/unmanaged-api/profiling/corprof-e-unsupported-call-sequence-hresult.md). ## <a name="requirements"></a>Požadavky **Platformy:** najdete v části [požadavky na systém](../../../../docs/framework/get-started/system-requirements.md). **Záhlaví:** CorProf.idl, CorProf.h **Knihovna:** CorGuids.lib **Verze rozhraní .NET framework:** [!INCLUDE[net_current_v45plus](../../../../includes/net-current-v45plus-md.md)] ## <a name="see-also"></a>Viz také [EnumJITedFunctions – metoda](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo3-enumjitedfunctions-method.md) [ICorProfilerInfo4 – rozhraní](../../../../docs/framework/unmanaged-api/profiling/icorprofilerinfo4-interface.md) [Rozhraní pro profilaci](../../../../docs/framework/unmanaged-api/profiling/profiling-interfaces.md) [Profilace](../../../../docs/framework/unmanaged-api/profiling/index.md)
63.736842
1,270
0.765483
ces_Latn
0.54461
87a93f06785f78d6cfb0a78cecda8e63d2cc8ad4
17
md
Markdown
2017/06/16/commit.md
ITWONDER753/ffmpeg-test
770f0c607bd9db5538632b30b02ae99c39ae1583
[ "MIT" ]
null
null
null
2017/06/16/commit.md
ITWONDER753/ffmpeg-test
770f0c607bd9db5538632b30b02ae99c39ae1583
[ "MIT" ]
null
null
null
2017/06/16/commit.md
ITWONDER753/ffmpeg-test
770f0c607bd9db5538632b30b02ae99c39ae1583
[ "MIT" ]
null
null
null
07 on 06/16/2017
8.5
16
0.705882
fin_Latn
0.611984
87aa83a5b808c12a7f5d2dc5b148c0a4ae15768f
2,429
md
Markdown
articles/supply-chain/service-management/service-template.md
saurabhsurana/dynamics-365-unified-operations-public
f59df61799915f6a79aec7e3e8664c02df6597da
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/supply-chain/service-management/service-template.md
saurabhsurana/dynamics-365-unified-operations-public
f59df61799915f6a79aec7e3e8664c02df6597da
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/supply-chain/service-management/service-template.md
saurabhsurana/dynamics-365-unified-operations-public
f59df61799915f6a79aec7e3e8664c02df6597da
[ "CC-BY-4.0", "MIT" ]
1
2021-01-28T06:03:43.000Z
2021-01-28T06:03:43.000Z
--- # required metadata title: Service templates description: You can define a service agreement as a template and copy the lines of the template later into another service agreement or into a service order. author: ShylaThompson manager: tfehr ms.date: 02/19/2018 ms.topic: article ms.prod: ms.service: dynamics-ax-applications ms.technology: # optional metadata ms.search.form: SMAAgreementTable # ROBOTS: audience: Application User # ms.devlang: ms.reviewer: kamaybac # ms.tgt_pltfrm: ms.custom: ms.assetid: ms.search.region: Global # ms.search.industry: ms.author: kamaybac ms.search.validFrom: 2016-02-28 ms.dyn365.ops.version: AX 7.0.0 --- # Service templates [!include [banner](../includes/banner.md)] You can define a service agreement as a template and copy the lines of the template later into another service agreement or into a service order. ## Example A customer for whom you provide service has identical service elevators at five different locations. You want to set up five service agreements for the customer, one for each site. To limit repetitive setup work, and to make sure that the setup information in the service agreements is identical, you create a service agreement and specify it as a template for the service work on the elevators. You can now copy the template lines into the five new service agreements, so that each service agreement is populated with the lines from the template. ## Create a template When you create a service template, you create a service agreement, create the required lines on it, and attach the service agreement to a service-template group. > [!NOTE] > A service agreement can be defined as a template only if it has no service orders attached to it. Also, no service orders can be generated from a service agreement that is defined as a template. ## Copy template lines You can copy template lines from a service template into another service agreement or into a service order. When you copy template lines into your service orders or service agreements, your template groups are displayed in a tree view in which each group can be expanded. This display lets you view each template and template line. It is easier to determine the service-template lines that you want to copy if you have grouped the templates under names that reflect the use of the templates. ## Related topics [Copy service templates lines](copy-service-template-lines.md)
31.545455
158
0.787567
eng_Latn
0.99836
87aaa484f66e9e549d23ea889cc46e1d4c6049c9
3,465
md
Markdown
articles/cognitive-services/Bing-Autosuggest/concepts/get-suggestions.md
changeworld/azure-docs.sv-se
6234acf8ae0166219b27a9daa33f6f62a2ee45ab
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Bing-Autosuggest/concepts/get-suggestions.md
changeworld/azure-docs.sv-se
6234acf8ae0166219b27a9daa33f6f62a2ee45ab
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/cognitive-services/Bing-Autosuggest/concepts/get-suggestions.md
changeworld/azure-docs.sv-se
6234acf8ae0166219b27a9daa33f6f62a2ee45ab
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Föreslå söktermer med API:et för automatiska förslag i Bing titleSuffix: Azure Cognitive Services description: I den här artikeln beskrivs begreppet förslag på frågetermer med hjälp av API:et för automatiska förslag på Bing och hur frågelängden påverkar relevansen. services: cognitive-services author: aahill manager: nitinme ms.service: cognitive-services ms.subservice: bing-autosuggest ms.topic: conceptual ms.date: 02/20/2019 ms.author: aahi ms.openlocfilehash: 060dbd29ee4ddb78e8ae9b2ed4e7814da3c4eebf ms.sourcegitcommit: 2ec4b3d0bad7dc0071400c2a2264399e4fe34897 ms.translationtype: MT ms.contentlocale: sv-SE ms.lasthandoff: 03/28/2020 ms.locfileid: "74072897" --- # <a name="suggesting-query-terms"></a>Föreslå frågeord Vanligtvis anropar du API:et för automatiska förslag i Bing varje gång en användare matar in ett nytt tecken i programmets sökruta. Frågesträngens fullständighet påverkar relevansen för de föreslagna frågetermer som API:et returnerar. Ju mer fullständig frågesträngen är, desto mer relevant blir listan över föreslagna frågetermer. Till exempel är de förslag som API:et kan returnera för `s` förmodligen mindre relevanta än de frågor det returnerar för `sailing dinghies`. ## <a name="example-request"></a>Exempelbegäran I följande exempel visas en begäran som returnerar de föreslagna frågesträngarna för *sail* (segla). Kom ihåg att URL-koda användarens partiella frågeterm när du anger frågeparametern [q](https://docs.microsoft.com/rest/api/cognitiveservices-bingsearch/bing-autosuggest-api-v7-reference#query). Om användaren till exempel anger *sailing les* ställer du in `q` till `sailing+les` eller `sailing%20les`. ```http GET https://api.cognitive.microsoft.com/bing/v7.0/suggestions?q=sail&mkt=en-us HTTP/1.1 Ocp-Apim-Subscription-Key: 123456789ABCDE X-MSEdge-ClientIP: 999.999.999.999 X-Search-Location: lat:47.60357;long:-122.3295;re:100 X-MSEdge-ClientID: <blobFromPriorResponseGoesHere> Host: api.cognitive.microsoft.com ``` Följande svar innehåller en lista över [SearchAction](https://docs.microsoft.com/rest/api/cognitiveservices-bingsearch/bing-autosuggest-api-v7-reference#searchaction)-objekt som innehåller de föreslagna frågetermerna. ```json { "url" : "https:\/\/www.bing.com\/search?q=sailing+lessons+seattle&FORM=USBAPI", "displayText" : "sailing lessons seattle", "query" : "sailing lessons seattle", "searchKind" : "WebSearch" }, ... ``` ## <a name="using-suggested-query-terms"></a>Använda föreslagna söktermer Varje förslag innehåller fälten `displayText`, `query` och `url`. Fältet `displayText` innehåller den föreslagna fråga som du använder för att fylla i sökrutans listruta. Du måste visa alla förslag som svaret innehåller, och i den angivna ordningen. I följande exempel visas en nedrullningsbar sökruta med föreslagna sökord från API för automatiska förslag för Bing. ![Lista med automatiska förslag i sökrutans listruta](../media/cognitive-services-bing-autosuggest-api/bing-autosuggest-drop-down-list.PNG) Om användaren väljer en föreslagen fråga från listrutan använder du frågetermen i fältet `query` för att anropa [API för webbsökning i Bing](../../bing-web-search/search-the-web.md) och visa resultaten själv. Alternativt kan du använda URL:en i fältet `url` för att dirigera användaren till sidan för Bing-sökresultat i stället. ## <a name="next-steps"></a>Nästa steg * [Vad är API:et för automatiska förslag i Bing?](../get-suggested-search-terms.md)
57.75
472
0.792785
swe_Latn
0.997452
87aaa86245ffde426d63d41c1b2d3ed9a9f4fe0e
896
md
Markdown
README.md
mazgi-sandbox/ansible-playbook
b68e921213a68616050a0851c92e39cb8c98f813
[ "MIT" ]
null
null
null
README.md
mazgi-sandbox/ansible-playbook
b68e921213a68616050a0851c92e39cb8c98f813
[ "MIT" ]
null
null
null
README.md
mazgi-sandbox/ansible-playbook
b68e921213a68616050a0851c92e39cb8c98f813
[ "MIT" ]
null
null
null
# Ansible Playbook ## How To Use ### How To install ```shellsession $ brew install https://raw.githubusercontent.com/kadwanev/bigboybrew/master/Library/Formula/sshpass.rb $ sudo easy_install pip $ sudo pip install virtualenv $ export PIP_RESPECT_VIRTUALENV=true $ export PATH="${HOME}/.local/bin:${PATH}" $ export PYTHONPATH="${HOME}/.local/lib64/python2.7/site-packages:${PYTHONPATH}" $ git clone https://github.com/mazgi/ansible-playbook.git # cd ansible-playbook $ virtualenv --no-site-packages --distribute --python=$(which python2.7) . $ source bin/activate (envdir)$ pip install --requirement requirements.txt ``` ### How To Play ```shellsession $ export PIP_RESPECT_VIRTUALENV=true $ export PATH="${HOME}/.local/bin:${PATH}" $ export PYTHONPATH="${HOME}/.local/lib64/python2.7/site-packages:${PYTHONPATH}" $ source bin/activate (envdir)$ ansible-playbook -i 10.11.12.13, site.yml ```
29.866667
102
0.737723
yue_Hant
0.255883
87ab33ae71c1f3d1c2b269773eeaa253698cb920
185
md
Markdown
content/lua/bt/lycanthrope.md
xackery/eqquestapi
e3bb4d58651c7c2bb1ced94deb59115946eed3c5
[ "MIT" ]
null
null
null
content/lua/bt/lycanthrope.md
xackery/eqquestapi
e3bb4d58651c7c2bb1ced94deb59115946eed3c5
[ "MIT" ]
1
2020-09-08T17:21:08.000Z
2020-09-08T17:21:08.000Z
content/lua/bt/lycanthrope.md
xackery/eqquestapi
e3bb4d58651c7c2bb1ced94deb59115946eed3c5
[ "MIT" ]
1
2020-08-29T00:49:26.000Z
2020-08-29T00:49:26.000Z
--- title: BT Lycanthrope searchTitle: Lua BT Lycanthrope weight: 1 hidden: true menuTitle: BT Lycanthrope --- ## Lycanthrope This represents a Lycanthrope BT ```lua BT.Lycanthrope ```
14.230769
32
0.751351
eng_Latn
0.938988
87abf66789d18a46109ce07dd2dbd40ca4a6cbd0
1,209
md
Markdown
README.en-US.md
inkuxuan/BNFCartesianProduct
59a017b5a62e7c85b03cdb79f55ef9ab7c45cbfa
[ "WTFPL" ]
2
2020-10-16T07:52:04.000Z
2021-01-01T03:26:35.000Z
README.en-US.md
inkuxuan/BNFCartesianProduct
59a017b5a62e7c85b03cdb79f55ef9ab7c45cbfa
[ "WTFPL" ]
null
null
null
README.en-US.md
inkuxuan/BNFCartesianProduct
59a017b5a62e7c85b03cdb79f55ef9ab7c45cbfa
[ "WTFPL" ]
1
2021-01-01T03:26:37.000Z
2021-01-01T03:26:37.000Z
# Simple BNF Defined Plain Text's Cartesian Production Generator [中文(简体)](README.md) ## Grammar * bracket including `'(', ')', '[', ']', '<', '>'` with`()` and `<>`'s meaning identical, `[]` meaning optional * the OR operator `|` e.g. input ```$xslt I <wanna|want to|need to> have <lunch|dinner>. ``` output ```$xslt I wanna have lunch. I wanna have dinner. I want to have lunch. I want to have dinner. I need to have lunch. I need to have dinner. ``` ## Usage invoke class Main, parameters are as follows `[-o <output-file>] [-append] <input-file> [input-file]...` ### parameters * `-o` specify an output file, the default one is `output.txt` * `-append` append to the output file rather than overwriting it * multiple input file can be provided at once Note: output encoding depends on the encoding of the first input file # Dependencies depend on `com.mozilla.universalchardet` to detect file encoding PLEASE TELL ME IF THERE IS A BETTER ONE TO USE [地址在此处](https://code.google.com/archive/p/juniversalchardet/) # LICENSE com.mozilla.universalchardet uses [Mozilla Public License](https://www.mozilla.org/en-US/MPL/) The rest of this project uses [WTFPL](http://www.wtfpl.net/txt/copying)
23.25
111
0.704715
eng_Latn
0.843447
87ad0fce0ed8e41347121d89ed447688d8c5009f
38
md
Markdown
posts/fun.md
colingao77/colingao77.github.io
f1ae65e154b5c1a9b7262434f61d1a13c745cb34
[ "Apache-2.0" ]
null
null
null
posts/fun.md
colingao77/colingao77.github.io
f1ae65e154b5c1a9b7262434f61d1a13c745cb34
[ "Apache-2.0" ]
null
null
null
posts/fun.md
colingao77/colingao77.github.io
f1ae65e154b5c1a9b7262434f61d1a13c745cb34
[ "Apache-2.0" ]
null
null
null
--- layout: category category: fun ---
9.5
16
0.657895
eng_Latn
0.473572
87add04a5b35beaaa4c04049ec0d6d7d74edfb98
1,155
md
Markdown
README.md
robbiet480/winston-newrelic
58b87671aee8a6a87acf301939069fbd4a0629bc
[ "MIT" ]
null
null
null
README.md
robbiet480/winston-newrelic
58b87671aee8a6a87acf301939069fbd4a0629bc
[ "MIT" ]
null
null
null
README.md
robbiet480/winston-newrelic
58b87671aee8a6a87acf301939069fbd4a0629bc
[ "MIT" ]
null
null
null
# winston-newrelic [![Build Status](https://secure.travis-ci.org/ranm8/winston-newrelic.png?branch=master)](http://travis-ci.org/ranm8/winston-newrelic) New Relic trasporter for [winston](https://github.com/flatiron/winston). Logs winston caught exceptions and logs to New Relic. ## Installation npm install winston-newrelic ## Usage ```javascript // Top of your boot file require('newrelic'); var winston = require('winston'), WinstonNewrelic = require('winston-newrelic'); // Add newrelic logger as transporter winston.add(winston.transports.newrelic, {}); ``` ## Error logging support It is now possible to log errors by sending them as the meta-object to the logger. ```javascript winston.log('error', 'message', new Error(message)); // An errorish is any object with a stack property var errorish = { stack: 'stacktrace'}; winston.log('error', 'message', errorish); // Will log {message: 'message', stack: 'stacktrace'} var errorish = { stack: 'stacktrace', message: 'error-message'}; winston.log('error', 'message', errorish); // Will log {message: 'error-message', stack: 'stacktrace'} // meta.message overrides message param ```
27.5
152
0.726407
eng_Latn
0.707548
87aefbe2e78687a3909e589f518a8c447147671f
3,818
md
Markdown
articles/connections/passwordless/spa-sms.md
alejofernandez/docs
6cebd53a07374f57a9366a5dd008236523a9624c
[ "MIT" ]
1
2019-02-17T02:22:26.000Z
2019-02-17T02:22:26.000Z
articles/connections/passwordless/spa-sms.md
alejofernandez/docs
6cebd53a07374f57a9366a5dd008236523a9624c
[ "MIT" ]
null
null
null
articles/connections/passwordless/spa-sms.md
alejofernandez/docs
6cebd53a07374f57a9366a5dd008236523a9624c
[ "MIT" ]
4
2017-05-18T15:51:41.000Z
2021-01-06T13:26:43.000Z
--- title: Using Passwordless Authentication in SPA with SMS --- # Authenticate users with a one-time code via SMS in a SPA <%= include('./_introduction-sms', { isMobile: true }) %> ## Setup <%= include('./_setup-sms-twilio') %> <%= include('./_setup-cors') %> ## Implementation ### Use Auth0 UI widget (Lock) <%= include('../../_includes/_package', { pkgRepo: 'auth0-jquery-passwordless-sample', pkgBranch: 'master', pkgPath: null, pkgFilePath: null, pkgType: 'js' }) %> <%= include('./_init-passwordless-lock') %> You can then trigger the login widget with the following code: ```html <script src="${lock_passwordless_url}"></script> <script type="text/javascript"> function login(){ // Initialize Passwordless Lock instance var lock = new Auth0LockPasswordless('${account.clientId}', '${account.namespace}'); // Open the lock in Email Code mode with the ability to handle // the authentication in page var appearanceOpts = { autoclose: true }; // Open the lock in SMS mode with the ability to handle the authentication in page lock.sms(appearanceOpts,function (error, profile, id_token, access_token, state, refresh_token) { if (!error) { //usually save profile and id_token } }); } </script> <a href="javascript:login()">Login</a> ``` This will open a dialog that asks the user for their phone number. ![](/media/articles/connections/passwordless/passwordless-sms-enter-phone-web.png) Then Auth0 will use Twilio to send to the user an SMS containing the one-time code: ![](/media/articles/connections/passwordless/passwordless-sms-receive-code-web.png) Lock will ask for the code that has been sent via SMS to the provided number. The code can then be used as a one-time password to log in: ![](/media/articles/connections/passwordless/passwordless-sms-enter-code-web.png) If the code is correct, the user will be authenticated. This will call the callback of the `lock.sms` function where the `id_token`, `refresh_token` and user profile are typically stored. Then the user will be allowed to continue to the authenticated part of the application. ### Use your own UI <%= include('../../_includes/_package', { pkgRepo: 'auth0-jquery-passwordless-sample', pkgBranch: 'master', pkgPath: null, pkgFilePath: null, pkgType: 'js' }) %> You can perform passwordless authentication in your SPA with your own custom UI using the [Auth0 JavaScript client library](/libraries/auth0js). <%= include('./_init-auth0js', {withCallbackURL:false} ) %> You must provide a way for the user to enter a phone number to which the SMS will be sent. Then you can begin the passwordless authentication as follows: ```js function sendSMS(){ var phoneNumber = $('input.phone-number').val(); auth0.requestSMSCode({ phoneNumber:phoneNumber }, function(err) { if (err) { alert('error sending SMS: '+ err.error_description); return; } // the request was successful and you should // receive the passcode to the specified phone $('.enter-phone').hide(); $('.enter-code').show(); }); } ``` This will send an SMS to the provided phone number. The user must now enter the code into your custom UI. Then you can continue with the login as follows: ```js function login(){ var phone = $('input.phone-number').val(); var code = $('input.code').val(); //submit the passcode to authenticate the phone auth0.verifySMSCode({ phoneNumber: phone, code: code }, function (err, profile, id_token, access_token) { if (err){ alert('Couldn\'t login ' + err.message); } else { // the authentication was successful // you can use profile, id_token and access_token localStorage.setItem('userToken', id_token); alert('Welcome ' + profile.name ); } }); }; ```
32.355932
275
0.693557
eng_Latn
0.964756
87af591b25f193ff86a70f997ccf6c2fdc6596be
574
md
Markdown
docs/grunt-json-htmltemplate.md
michaelBenin/grunt-json-htmltemplate
b03abe659b824f8bfe8065fa0463d90b8f3977d6
[ "MIT" ]
1
2015-11-07T12:37:07.000Z
2015-11-07T12:37:07.000Z
docs/grunt-json-htmltemplate.md
michaelBenin/grunt-json-htmltemplate
b03abe659b824f8bfe8065fa0463d90b8f3977d6
[ "MIT" ]
null
null
null
docs/grunt-json-htmltemplate.md
michaelBenin/grunt-json-htmltemplate
b03abe659b824f8bfe8065fa0463d90b8f3977d6
[ "MIT" ]
null
null
null
## Example config ```javascript grunt.initConfig({ json_template_html: { src: { src: 'src/js/template/src/templates.json', templates: 'src/templates/compress/*', dest: 'src/js/template/built/templates.json' } } }); grunt.registerTask('default', ['json_template_html']); // reference the templates name in grunt.jsTemplate { "share_partial": <% print(grunt.jsTemplate["_share_partial"]); %>, "home_partial": <% print(grunt.jsTemplate["_home_partial"]); %> } ```
19.133333
70
0.576655
yue_Hant
0.118328
87af798e2cda2521f7b52dc439992948716b908f
1,871
md
Markdown
README.md
jeffcrouse/nola300-client
a45366da3ed7c2fa9d03fc181ee05adadb6cc79c
[ "MIT" ]
null
null
null
README.md
jeffcrouse/nola300-client
a45366da3ed7c2fa9d03fc181ee05adadb6cc79c
[ "MIT" ]
null
null
null
README.md
jeffcrouse/nola300-client
a45366da3ed7c2fa9d03fc181ee05adadb6cc79c
[ "MIT" ]
null
null
null
# nola300 code repo for the Nola300 project #install instructions - Install SublimeText and ``ln -s "/Applications/Sublime Text 2.app/Contents/SharedSupport/bin/subl" /usr/local/bin/subl`` - install git https://git-scm.com/download/mac - setup git names ``` git config --global user.name "Jeff Crouse" git config --global user.email jeff@seethroughlab.com ``` - ``mkdir Developer && cd Developer`` - ``git clone https://github.com/jeffcrouse/nola300-client.git`` - ``cd ~/Developer/nola300-client && npm install`` - Move the 2 frameworks in Frameworks folder into /Library/Frameworks - Add .env file - Make sure to enable auto-startup options - Install https://robomongo.org - Turn on auto-login - Turn off energy settings - install Node https://nodejs.org/en/ - ``npm install pm2 -g` - Make aliases of video apps on desktop ### Homebrew stuff ```/usr/bin/ruby -e "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/master/install)"``` ``sudo chown -R `whoami` /usr/local`` ``brew update`` ``brew install mongodb@3.4`` ``` brew install ffmpeg \ --with-tools \ --with-fdk-aac \ --with-freetype \ --with-fontconfig \ --with-libass \ --with-libvorbis \ --with-libvpx \ --with-opus \ --with-x265 ``` ``` brew install lame brew install sox --with-lame ``` ## Make folders - Music Folder: ~/Music/NOLA_MUSIC (used by RearProjector) - Textures Folder: ~/Movies/NOLA_TEXTURES (used by FrontProjector) - Videos Folder: ~/Movies/NOLA_VIDEOS (used by nola300-client) - Storage Folder: ~/Documents/NOLA_STORAGE (used by nola300-client) ## Copy Files - To Music Folder - To Textures Folder - To Videos Folder ## Make launch_server.command ``` cd ~/Developer/nola300-client npm start ``` chmod a+x launch_server.command ## Run Video Sync Script ``` cd ~/Developer/nola300-client node scripts/video_tag_sync.js ```
22.011765
122
0.703902
kor_Hang
0.351798
87b00f70c233294fbcc9e47dd4227d0bb393a7ad
862
md
Markdown
content/lectures/lecture17/index.md
tteenathankachan/2021-CS109A
9349749539f439deff280ceb25621fc61a337253
[ "MIT" ]
19
2021-08-29T21:23:48.000Z
2022-03-16T14:38:25.000Z
content/lectures/lecture17/index.md
SBalas/2021-CS109A
0f57c3d80b7cef99d660f6a77c0166cffc1253e8
[ "MIT" ]
null
null
null
content/lectures/lecture17/index.md
SBalas/2021-CS109A
0f57c3d80b7cef99d660f6a77c0166cffc1253e8
[ "MIT" ]
22
2021-09-01T13:03:05.000Z
2022-03-31T14:34:36.000Z
Title: Lecture 17: Bagging Category: lectures Slug: lecture17 Author: Pavlos Protopapas and Natesh Pillai Date: 2021-11-03 Tags: Decision Trees, Regression Trees, Stopping Conditions, Pruning, Bagging, Overfitting ## Slides - [Lecture 17 : Decision Trees - B (PDF)]({attach}presentation/Decision_Trees_B.pdf) - [Lecture 17 : Decision Trees - C (PDF)]({attach}presentation/Decision_Trees_C.pdf) - [Lecture 17 : Decision Trees - Bagging (PDF)]({attach}presentation/Bagging.pdf) ## Exercises - [Lecture 17: Exercise - Classification using Decision Tree [Notebook]]({filename}notebook/dt_classifier_scaffold.ipynb) - [Lecture 17: Exercise - Bagging Classification with Decision Boundary [Notebook]]({filename}notebook/bagging_classification_scaffold.ipynb) - [Lecture 17: Exercise - Regression with Bagging [Notebook]]({filename}notebook/overfitting_bagging.ipynb)
53.875
141
0.786543
eng_Latn
0.278017
87b117928e6b93fb94ff782ef6b32c59d35f2e4e
3,361
md
Markdown
_posts/2019-02-17-Download-sony-ic-recorder-icd-ux71-manual-espanol.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
2
2019-02-28T03:47:33.000Z
2020-04-06T07:49:53.000Z
_posts/2019-02-17-Download-sony-ic-recorder-icd-ux71-manual-espanol.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
null
null
null
_posts/2019-02-17-Download-sony-ic-recorder-icd-ux71-manual-espanol.md
Anja-Allende/Anja-Allende
4acf09e3f38033a4abc7f31f37c778359d8e1493
[ "MIT" ]
null
null
null
--- layout: post comments: true categories: Other --- ## Download Sony ic recorder icd ux71 manual espanol book you can be his Ford Country Squire out of the garage and drove to the nursery, as he can. "We made an arrest over at the shuttle base-just before midnight, "But I sony ic recorder icd ux71 manual espanol buying the English," she said firmly, and it had been only a matter of minutes before lift-off when one of the flight-crew noticed that suddenly they weren't there-any of them. same extent that a stone-serious fan of Star Trek III: The Search for Spock could recite its dialogue quickly? " with European harpoons, which were made of a single tree stem, as we discussed. " had been a burden to him in his youth, Morris Hazeldorf, if he is not too old or has not recently small collection of fine wines. " fingers. "You're not a healer?" It was an accusation! He's color-blind. "What I hope I found there was direction, Jean emitted an audible sigh of relief. You're like Eve and Jerry? If you Tuesday, which was little, and one of them lurched against a table behind, and the loss of her will leave a hole in his heart for the rest of his time in probing questions about his marital status. 19, brush-cut man in black slacks and a gray herringbone sports jacket, with a tumbleweed bush of red hair; her face isn't so much pretty as it is intense, Junior was some might hope to defend against a wrongful-death suit, gathering wizards to work together at the In fact. Let me check. And what is a mesk! 193 when frozen in--The nature of the neighbouring country--The _Vega_ losing those he loved. " Petersburg. "Not just Oregon. it passed. " I took a deep breath and lied with a straight face. The overlapping swish-and-lug of seven were telltales that none of these professionals would overlook. just as the smile curved to completion, Bobby said, but. " He moved away from view? But the colony needs it We've all sony ic recorder icd ux71 manual espanol it: long-term consequences, a _simovie_ on the bank of the Yenisej in newfound fragile hopefulness represented progress, the irony that is the mother-of-all in human sony ic recorder icd ux71 manual espanol, went in to the trooper; but she had foregone him. She started walking toward the sony ic recorder icd ux71 manual espanol again, but that, and Micky wished this would service-station pumps and barricades of parked vehicles to reach him. Unfortunately, drawn by V, and we travel all around W Some of the scaffolding was still in place along the wall of the sixth stage. " them a part of the House that will be all their own, but a would-be stage magician eventually needed a mentor to campsites with power-and-water hookups to motor homes and travel trailers, LEHMANN the geologist. Sometimes he wanted darkness for the made a baby with me, seated himself at his head and bathed his face with rose-water! He didn't look at the license till he was out on the street Stapled to the back of it was a printed was shade from the hot sun four or five women sat spinning by a well. might be figments of his imagination rather than real presences perceived untaught knowledge of at least some words of the Language of the Making. " He closed the ring box. dressed in all manner of styles and colors and reflecting the various races of Earth in more or less even proportions, Colman reflected.
373.444444
3,247
0.785778
eng_Latn
0.999923
87b136d1126adc5359b47174c3cb3b3e9d3df6d5
1,110
md
Markdown
Changelog.md
lvcabral/PoPSpritesConverter
a78f85d7a8bd48a8548950d87ac60bb9aa28c3ce
[ "MIT" ]
2
2017-04-20T00:03:25.000Z
2018-08-01T22:10:24.000Z
Changelog.md
lvcabral/PoPSpritesConverter
a78f85d7a8bd48a8548950d87ac60bb9aa28c3ce
[ "MIT" ]
6
2016-08-20T22:22:17.000Z
2016-08-29T06:35:18.000Z
Changelog.md
lvcabral/PoPSpritesConverter
a78f85d7a8bd48a8548950d87ac60bb9aa28c3ce
[ "MIT" ]
2
2018-10-01T10:59:24.000Z
2021-11-02T02:24:43.000Z
# Changelog ##### v1.1.2 - 23-Ago-2016 - Tiles Palettes, Doors, Slicer and Potions * Add: Support for different palettes on tiles (DOS 1.3 & 1.4) * Fix: Door frames painting is wrong * Fix: Stars frames to scenes are not included on sprite map * Fix: Title messages should not be transparent (except game-name) * Fix: Slicer painting is wrong * Fix: Palace potions are getting dungeon graphics * Fix: Dungeon should only have two potion foreground tiles ##### v1.1.1 - 14-Ago-2016 - Mirror and Balcony * Fix: Mirror tile is not complete (missing right part of the floor) * Fix: Balcony tile has wall marks over it ##### v1.1.0 - 13-Ago-2016 - WDA and Bubbles * Add: Support for WDA for palace tiles (replaced the old -pwm parameter) * Add: Support for bubble frames with any number of colors palette (the last color is replaced) * Fix: The generation of tiles does not handle correctly custom tiles with different sizes * Fix: Guards generation is crashing when the guard palette file is missing ##### v1.0.0 - 04-Ago-2016 - Initial Release * Add: Reversal Potion (big green): Flip screen upside down (L9)
46.25
95
0.737838
eng_Latn
0.98952
87b1857aeab34dc60a92e08c1134310a1707c282
446
md
Markdown
_posts/2021-07-29-gitblog-1.md
zzangyeon/zzangyeon.github.io
34149390e3eb9ef1b3bd561c8a027786abfe7656
[ "MIT" ]
null
null
null
_posts/2021-07-29-gitblog-1.md
zzangyeon/zzangyeon.github.io
34149390e3eb9ef1b3bd561c8a027786abfe7656
[ "MIT" ]
null
null
null
_posts/2021-07-29-gitblog-1.md
zzangyeon/zzangyeon.github.io
34149390e3eb9ef1b3bd561c8a027786abfe7656
[ "MIT" ]
null
null
null
--- layout: post title: 깃허브 블로그 만들기 date: 2021-07-29 00:00 +0800 # last_modified_at: 2020-10-01 01:08:25 +0900 tags: github toc: true --- ## 1. 자꾸 까먹는 깃허브 블로그 명령어 [**지킬을 이용한 정적페이지 생성시 초기 셋팅**] ``` bundle install ``` - 폴더 내에 Gemfile이 있어야 한다. gemfile이 없는 경우는 오류가 난다. ``` bundle exec jekyll serve ``` - 로컬 서버를 통해 수정한 부분이 바로바로 적용이 돼어 실시간으로 모니터링 할 수 있다. - 일일이 깃허브에 commit push하고 새로고침하여 반영되었는지 확인하는 번거로운 일을 하지 않아도 된다. 최종본만 commit push.
13.9375
80
0.659193
kor_Hang
1.00001
87b27a0613389e63f8abc7cfd5369f40391be2b6
8,946
md
Markdown
src/com/practice/sheet/linkedlist/ReadMe.md
lakshayjawa/dsapractice
4c5ea7dceec52db7efa2913c9059b501cfd4a176
[ "MIT" ]
3
2021-03-23T18:08:51.000Z
2021-10-18T08:15:00.000Z
src/com/practice/sheet/linkedlist/ReadMe.md
lakshayjawa/dsapractice
4c5ea7dceec52db7efa2913c9059b501cfd4a176
[ "MIT" ]
null
null
null
src/com/practice/sheet/linkedlist/ReadMe.md
lakshayjawa/dsapractice
4c5ea7dceec52db7efa2913c9059b501cfd4a176
[ "MIT" ]
null
null
null
## Linked List Questions for DSA - Legend - Difficulty - :green_circle: Easy - :orange_circle: Medium - :red_circle: Hard - :star: Important - :white_check_mark: Done - :negative_squared_cross_mark: Not done | Difficulty | Done | Solution | Problem | | ------------- | ------------------------------ | -------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ | | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Write a Program to reverse the Linked List. (Both Iterative and recursive)](https://www.geeksforgeeks.org/reverse-a-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Reverse a Linked List in group of Given Size. \[Very Imp\]](https://practice.geeksforgeeks.org/problems/reverse-a-linked-list-in-groups-of-given-size/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Write a program to Detect loop in a linked list.](https://practice.geeksforgeeks.org/problems/detect-loop-in-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Write a program to Delete loop in a linked list.](https://practice.geeksforgeeks.org/problems/remove-loop-in-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Find the starting point of the loop. ](https://www.geeksforgeeks.org/find-first-node-of-loop-in-a-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Remove Duplicates in a sorted Linked List.](https://practice.geeksforgeeks.org/problems/remove-duplicate-element-from-sorted-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Remove Duplicates in a Un-sorted Linked List.](https://practice.geeksforgeeks.org/problems/remove-duplicates-from-an-unsorted-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Write a Program to Move the last element to Front in a Linked List.](https://www.geeksforgeeks.org/move-last-element-to-front-of-a-given-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Add “1” to a number represented as a Linked List.](https://practice.geeksforgeeks.org/problems/add-1-to-a-number-represented-as-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Add two numbers represented by linked lists.](https://practice.geeksforgeeks.org/problems/add-two-numbers-represented-by-linked-lists/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Intersection of two Sorted Linked List.](https://practice.geeksforgeeks.org/problems/intersection-of-two-sorted-linked-lists/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Intersection Point of two Linked Lists.](https://practice.geeksforgeeks.org/problems/intersection-point-in-y-shapped-linked-lists/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Merge Sort For Linked lists.\[Very Important\]](https://practice.geeksforgeeks.org/problems/sort-a-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Quicksort for Linked Lists.\[Very Important\]](https://practice.geeksforgeeks.org/problems/quick-sort-on-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Find the middle Element of a linked list.](https://leetcode.com/problems/middle-of-the-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Check if a linked list is a circular linked list.](https://practice.geeksforgeeks.org/problems/circular-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Split a Circular linked list into two halves.](https://practice.geeksforgeeks.org/problems/split-a-circular-linked-list-into-two-halves/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Write a Program to check whether the Singly Linked list is a palindrome or not.](https://practice.geeksforgeeks.org/problems/check-if-linked-list-is-pallindrome/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Deletion from a Circular Linked List.](https://www.geeksforgeeks.org/deletion-circular-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Reverse a Doubly Linked list.](https://practice.geeksforgeeks.org/problems/reverse-a-doubly-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Find pairs with a given sum in a DLL.](https://www.geeksforgeeks.org/find-pairs-given-sum-doubly-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Count triplets in a sorted DLL whose sum is equal to given value “X”.](https://www.geeksforgeeks.org/count-triplets-sorted-doubly-linked-list-whose-sum-equal-given-value-x/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Sort a “k”sorted Doubly Linked list.\[Very IMP\]](https://www.geeksforgeeks.org/sort-k-sorted-doubly-linked-list/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Rotate DoublyLinked list by N nodes.](https://www.geeksforgeeks.org/rotate-doubly-linked-list-n-nodes/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Rotate a Doubly Linked list in group of Given Size.\[Very IMP\]](https://www.geeksforgeeks.org/reverse-doubly-linked-list-groups-given-size/) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | Can we reverse a linked list in less than O(n) ? | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | Why Quicksort is preferred for. Arrays and Merge Sort for LinkedLists ? | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Flatten a Linked List](https://practice.geeksforgeeks.org/problems/flattening-a-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Sort a LL of 0's, 1's and 2's](https://practice.geeksforgeeks.org/problems/given-a-linked-list-of-0s-1s-and-2s-sort-it/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Clone a linked list with next and random pointer](https://practice.geeksforgeeks.org/problems/clone-a-linked-list-with-next-and-random-pointer/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Merge K sorted Linked list](https://practice.geeksforgeeks.org/problems/merge-k-sorted-linked-lists/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Multiply 2 no. represented by LL](https://practice.geeksforgeeks.org/problems/multiply-two-linked-lists/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Delete nodes which have a greater value on right side](https://practice.geeksforgeeks.org/problems/delete-nodes-having-greater-value-on-right/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Segregate even and odd nodes in a Linked List](https://practice.geeksforgeeks.org/problems/segregate-even-and-odd-nodes-in-a-linked-list/0) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Program for n’th node from the end of a Linked List](https://practice.geeksforgeeks.org/problems/nth-node-from-end-of-linked-list/1) | :green_circle: | :negative_squared_cross_mark: | [Solution](ReverseAnArray.java) | [Find the first non-repeating character from a stream of characters](https://practice.geeksforgeeks.org/problems/first-non-repeating-character-in-a-stream/0)
172.038462
397
0.680416
eng_Latn
0.265477
87b2b09aa516e205b1fa2bf57b606e0edea132a4
15,256
md
Markdown
translations/pt-BR/content/rest/overview/api-previews.md
Ruth536/docs
e341012b6c9b1079b187d08fb0508152fe9d2606
[ "CC-BY-4.0", "MIT" ]
9
2021-04-02T15:46:42.000Z
2022-03-06T17:21:45.000Z
translations/pt-BR/content/rest/overview/api-previews.md
Ruth536/docs
e341012b6c9b1079b187d08fb0508152fe9d2606
[ "CC-BY-4.0", "MIT" ]
52
2022-01-28T19:09:04.000Z
2022-03-30T18:16:41.000Z
translations/pt-BR/content/rest/overview/api-previews.md
erik-888/docs
2e78dc37bc875b4df09986b152a56129ace3d62d
[ "CC-BY-4.0", "MIT" ]
1
2021-03-15T15:02:46.000Z
2021-03-15T15:02:46.000Z
--- title: Pré-visualizações da API intro: Você pode usar pré-visualizações da API para testar novos recursos e fornecer feedback antes que estes recursos se tornem oficiais. redirect_from: - /v3/previews versions: free-pro-team: '*' enterprise-server: '*' github-ae: '*' --- Pré-visualizações da API permitem que você experimente novas APIs e alterações nos métodos de API existentes antes de se tornarem parte da API oficial do GitHub. Durante o período de pré-visualização, poderemos alterar alguns recursos com base no feedback do desenvolvedor. Se fizermos alterações, iremos anunciá-las no [blogue do desenvolvedor](https://developer.github.com/changes/) sem aviso prévio. Para acessar uma pré-visualização da API, você precisará fornecer um [tipo de mídia](/rest/overview/media-types) personalizado no cabeçalho `Aceitar` para suas solicitações. A documentação dos recursos para cada pré-visualização especifica qual tipo de mídia personalizado deve ser fornecido. {% if currentVersion == "free-pro-team@latest" %} ### Migrações Permite que você faça o download de repositórios da conta do usuário ou da organização do GitHub para revisar, fazer backup e [fazer a migração dos dados](/rest/reference/migrations) para {% data variables.product.prodname_ghe_server %}. **Tipo de mídia personalizada:** `wyandotte-preview` **Anunciado em:** [2018-05-24](https://developer.github.com/changes/2018-05-24-user-migration-api/) {% endif %} ### Implementações aprimoradas Exerça um maior controle sobre as [implantações](/rest/reference/repos#deployments) com mais informações e uma granularidade mais precisa. **Tipo de mídia personalizada:** `ant-man-preview` **Anunciado em:** [2016-04-06](https://developer.github.com/changes/2016-04-06-deployment-and-deployment-status-enhancements/) ### Reações Gerencie as [reações](/rest/reference/reactions) de commits, problemas e comentários. **Tipo de mídia personalizado:** `squirrel-girl-preview` **Anunciado em:** [2016-05-12](https://developer.github.com/changes/2016-05-12-reactions-api-preview/) **Atualização em:** [ 2016-07](https://developer.github.com/changes/2016-06-07-reactions-api-update/)</p> ### Linha do tempo Obter uma [lista de eventos](/rest/reference/issues#timeline) para um problema ou pull request. **Tipo de mídia personalizada:** `mockingbird-preview` **Anunciado em:** [2016-05-23](https://developer.github.com/changes/2016-05-23-timeline-preview-api/) {% if enterpriseServerVersions contains currentVersion %} ### Ambientes pre-receive Cria, lista, atualiza e exclui ambientes para hooks pre-receive. **Tipo de mídia personalizada:** `eye-scream-preview` **Anunciado em:** [2015-07-29](/rest/reference/enterprise-admin#pre-receive-environments) {% endif %} {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.22" %} ### Integrações Gerencie as [integrações](/v3/integrations) através da API. **Tipo de mídia personalizada:** `machine-man-preview` **Anunciado em:** [2016-09-14](https://developer.github.com/changes/2016-09-14-Integrations-Early-Access/) {% endif %} ### Projetos Gerencie [projetos](/rest/reference/projects). **Tipo de mídia personalizada:** `inertia-preview` **Anunciado em:** [2016-09-14](https://developer.github.com/changes/2016-09-14-projects-api/) **Atualização em:** [ 2016-10-27](https://developer.github.com/changes/2016-10-27-changes-to-projects-api/)</p> ### Pesquisa de commit [Pesquisa commits](/rest/reference/search). **Tipo de mídia personalizada:** `cloak-preview` **Anunciado em:** [2017-01-05](https://developer.github.com/changes/2017-01-05-commit-search-api/) ### Tópicos do repositório Ver uma lista dos [tópicos do repositório](/articles/about-topics/) em [chamadas](/rest/reference/repos) que retornam resultados do repositório. **Tipo de mídia personalizada:** `mercy-preview` **Anunciado em:** [2017-01-31](https://github.com/blog/2309-introducing-topics) ### Códigos de conduta Veja todos os [códigos de conduta](/rest/reference/codes-of-conduct) ou obtenha qual código de conduta um repositório tem atualmente. **Tipo de mídia personalizado:** `scarlet-witch-preview` {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.20" %} ### Equipes aninhadas Inclua o conteúdo aninhado das cargas da [equipe](/rest/reference/teams). **Tipo de mídia personalizada:** `hellcat-preview` **Anunciado em:** [2017-09-01](https://developer.github.com/changes/2017-08-30-preview-nested-teams) {% endif %} {% if currentVersion == "github-ae@latest" or enterpriseServerVersions contains currentVersion %} ### Webhooks globais Habilita [webhooks globais](/rest/reference/enterprise-admin#global-webhooks/) para [organizações](/webhooks/event-payloads/#organization) e tipos de evento do [usuário](/webhooks/event-payloads/#user). Esta visualização da API só está disponível para {% data variables.product.prodname_ghe_server %}. **Tipo de mídia personalizada:** `superpro-preview` **Anunciado em:** [2017-12-12](/rest/reference/enterprise-admin#global-webhooks) {% endif %} {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.20" %} ### Transferência de repositório Transfira um [repositório](/rest/reference/repos) para uma organização ou usuário. **Tipo de mídia personalizada:** `nightshade-preview` **Anunciado em:** [2017-11-09](https://developer.github.com/changes/2017-11-09-repository-transfer-api-preview) {% endif %} {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.22" %} ### Adicionar motivo do bloqueio Agora você pode adicionar um motivo ao[bloquear um problema](/rest/reference/issues#lock-an-issue). **Tipo de mídia personalizada:** `sailor-v-preview` **Anunciado em:** [2018-01-10](https://developer.github.com/changes/2018-01-10-lock-reason-api-preview) {% endif %} ### Exigir commits assinados Agora você pode usar a API para gerenciar a configuração para [exigir commits assinados em branches protegidos](/rest/reference/repos#branches). **Tipo de mídia personalizada:** `zzzax-preview` **Anunciado em:** [2018-02-22](https://developer.github.com/changes/2018-02-22-protected-branches-required-signatures) ### Exigir múltiplas revisões de aprovação Agora você pode [exigir múltiplas revisões de aprovação](/rest/reference/repos#branches) para um pull request usando a API. **Tipo de mídia personalizada:** `luke-cage-preview` **Anunciado em:** [2018-03-16](https://developer.github.com/changes/2018-03-16-protected-branches-required-approving-reviews) {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.19" %} ### Recuperar informações do hovercard Recuperar informações do [hovercard de alguém](/rest/reference/users#get-contextual-information-for-a-user). **Tipo de mídia personalizada:** `hagar-preview` **Anunciado:** [2018-03-21](https://developer.github.com/changes/2018-03-21-hovercard-api-preview) {% endif %} {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.23" %} ### Verificar execuções e a API de conjuntos de verificações Permite que um aplicativo GitHub execute verificações externas no código de um repositório. Veja as [execuções de verificação](/rest/reference/checks#runs) e [Conjuntos de verificação](/rest/reference/checks#suites) das APIs para obter mais informações. **Tipo de mídia personalizada:** `antiope-preview` **Anunciado:** [2018-05-07](https://developer.github.com/changes/2018-05-07-new-checks-api-public-beta/) {% endif %} {% if currentVersion == "github-ae@latest" or enterpriseServerVersions contains currentVersion %} ### Acesso de Git anônimo aos repositórios Quando uma instância do {% data variables.product.prodname_ghe_server %} estiver em modo privado, os administradores do site e do repositório podem habilitar o acesso anônimo ao Git para um repositório público. **Tipo de mídia personalizada:** `x ray-preview` **Anunciado:** [2018-07-12](https://blog.github.com/2018-07-12-introducing-enterprise-2-14/) {% endif %} ### Detalhes do cartão de projeto As respostas da API REST para [eventos de problemas](/rest/reference/issues#events) e [eventos da linha do tempo de problemas](/rest/reference/issues#timeline) agora retornam o campo `project_card` para eventos relacionados ao projeto. **Tipo de mídia personalizada:** `starfox-preview` **Anunciado:** [2018-09-05](https://developer.github.com/changes/2018-09-05-project-card-events) {% if currentVersion == "free-pro-team@latest" %} ### Manifestoes do aplicativo GitHub Os manifestos do aplicativo GitHub permitem que pessoas criem aplicativos GitHub pré-configurados. Veja "[Criar aplicativos GitHub a partir de um manifesto](/apps/building-github-apps/creating-github-apps-from-a-manifest/)" para obter mais inoformações. **Tipo de mídia personalizada:** `fury-preview` {% endif %} ### Status da implantação Agora você pode atualizar o ambiente `` de um [status de implantação](/rest/reference/repos#create-a-deployment-status) e usar os estados `in_progress` e `na fila`. Ao criar o status da implantação, agora você pode usar o parâmetro `auto_inactive` para marcar implantações de `produção` antigas como `inativa`. **Tipo de mídia personalizada:** `flash-preview` **Anunciado:** [2018-10-16](https://developer.github.com/changes/2018-10-16-deployments-environments-states-and-auto-inactive-updates/) ### Permissões de criação de repositório Agora você pode configurar se os integrantes da organização podem criar repositórios e que tipos de repositórios podem criar. Consulte "[Atualizar uma organização](/rest/reference/orgs#update-an-organization)" para obter mais informações. **Tipos de mídia personalizada:** `surtur-preview` **Anunciado:** [2019-12-03](https://developer.github.com/changes/2019-12-03-internal-visibility-changes/) ### Anexos de conteúdo Agora você pode fornecer mais informações no GitHub para URLs vinculadas a domínios registrados usando a API de {% data variables.product.prodname_unfurls %}. Consulte "[Usar anexos de conteúdo](/apps/using-content-attachments/)" para obter mais informações. **Tipos de mídia personalizada:** `corsair-preview` **Anunciado:** [2018-12-10](https://developer.github.com/changes/2018-12-10-content-attachments-api/) {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.21" %} ### Pull requests de rascunho Você pode usar a API do Pull Requests de rascunho e seus pontos de extremidade de [pull request](/rest/reference/pulls) para ver se um pull request está em estado rascunho. Para saber mais sobre pull requests, consulte "[Sobre pull requests](/articles/about-pull-requests/)". **Tipo de mídia personalizada:** `shadow-cat-preview` **Anunciado:** [2019-02-14](https://developer.github.com/changes/2019-02-14-draft-pull-requests/) {% endif %} ### Habilitar e desabilitar páginas Você pode usar os novos pontos de extremidade no [API de páginas](/rest/reference/repos#pages) para habilitar ou desabilitar páginas. Para saber mais sobre páginas, consulte "[Princípios básicos do GitHub Pages](/categories/github-pages-basics)". **Tipos de mídia personalizada:** `switcheroo-preview` **Anunciado:** [2019-03-14](https://developer.github.com/changes/2019-03-14-enabling-disabling-pages/) ### Listar branches ou pull requests para um commit Você pode usar dois novos pontos de extremidade na [API de commits](/rest/reference/repos#commits) para listar branches ou pull requests para um commit. **Tipos de mídia personalizada:** `groot-preview` **Anunciado:** [2019-04-11](https://developer.github.com/changes/2019-04-11-pulls-branches-for-commit/) {% if enterpriseServerVersions contains currentVersion and currentVersion ver_lt "enterprise-server@2.21" %} ### Desinstalar um aplicativo do GitHub Agora os proprietários dos aplicativos GitHub podem desinstalar um aplicativo usando a [API de aplicativos](/rest/reference/apps#delete-an-installation-for-the-authenticated-app). **Tipos de mídia personalizada:** `gambit-preview` {% endif %} ### Habilitar ou desabilitar alertas de vulnerabilidade para um repositório Você pode usar dois novos pontos de extremidade na [API de Repositórios](/rest/reference/repos) para habilitar ou desabilitar os alertas de vulnerabilidade. **Tipos de mídia personalizada:** `dorian-preview` **Anunciado:** [2019-04-24](https://developer.github.com/changes/2019-04-24-vulnerability-alerts/) ### Atualizar um branch de pull request Você pode usar um novo ponto de extremidade para [atualizar um branch de pull request](/rest/reference/pulls#update-a-pull-request-branch) com alterações do HEAD do branch upstream. **Tipos de mídia personalizada:** `lidian-preview` **Anunciado:** [2019-05-29](https://developer.github.com/changes/2019-05-29-update-branch-api/) {% if currentVersion == "free-pro-team@latest" %} ### Habilitar ou desabilitar correções de segurança automatizadas Você pode usar um novo conjunto de pontos de extremidade para [habilitar e desabilitar as correções de segurança automatizadas](/rest/reference/repos#enable-automated-security-fixes). **Tipo de mídia personalizada:** `london-preview` **Anunciado:** [2019-06-04](https://developer.github.com/changes/2019-06-04-automated-security-fixes/) {% endif %} ### Criar e usar modelos de repositório Você pode usar um novo ponto de extremidade para [Criar um repositório usando um modelo](/rest/reference/repos#create-a-repository-using-a-template) e [Criar um repositório para o usuário autenticado](/rest/reference/repos#create-a-repository-for-the-authenticated-user) que é um repositório de modelo, definindo o parâmetro `is_template` como `verdadeiro`. [Obter um repositório](/rest/reference/repos#get-a-repository) para verificar se ele é definido como um repositório de modelo usando a chave `is_template`. **Tipos de mídia personalizada:** `baptiste-preview` **Anunciado:** [2019-07-05](https://developer.github.com/changes/2019-07-16-repository-templates-api/) {% if currentVersion == "enterprise-server@2.20" %} ### Novos pontos de extremidade da API de aplicativos OAuth Você pode gerenciar os tokens de forma mais segura para aplicativos OAuth usando os tokens OAuth como parâmetros de entrada em vez dos parâmetros de caminho com os novos pontos de extremidade da [API dos aplicativos OAuth](/rest/reference/apps#oauth-applications). **Tipos de mídia personalizada:** `doutor-strange-preview` **Anunciado:** [2019-11-05](https://developer.github.com/changes/2019-11-05-deprecated-passwords-and-authorizations-api/) {% endif %} {% if currentVersion == "free-pro-team@latest" or currentVersion ver_gt "enterprise-server@2.19" or currentVersion == "github-ae@latest" %} ### Novo parâmetro de visibilidade para a API de repositórios Você pode definir e recuperar a visibilidade de um repositório na [API de repositórios](/rest/reference/repos). **Tipos de mídia personalizada:** `nebula-preview` **Anunciado:** [2019-11-25](https://developer.github.com/changes/2019-12-03-internal-visibility-changes/) {% endif %}
45.540299
513
0.765535
por_Latn
0.967081
87b30092353df1835f986a39368362a96dd5d8fb
650
md
Markdown
README.md
vgaurav3011/Average-Screen-Time-Calculation-of-Actors-Cartoon-Characters
0dc724f2e2d5ed927aa2cac2c05ee311b694c8f7
[ "MIT" ]
2
2018-09-19T16:27:48.000Z
2021-03-05T16:37:41.000Z
README.md
vgaurav3011/Average-Screen-Time-Calculation-of-Actors-Cartoon-Characters
0dc724f2e2d5ed927aa2cac2c05ee311b694c8f7
[ "MIT" ]
null
null
null
README.md
vgaurav3011/Average-Screen-Time-Calculation-of-Actors-Cartoon-Characters
0dc724f2e2d5ed927aa2cac2c05ee311b694c8f7
[ "MIT" ]
null
null
null
# Average-Screen-Time-Calculation-of-Actors-Cartoon-Characters <br/> ## Artificial Intelligence based project which computes the average screen duration of characters in a video through analysing the video frames <br/> Based on Sequential Model of keras, the program creates frames of the video played into various images of training and testing sets. <br/> It identifies the characters in the videos and predicts the screen space of the same in another video featuring the same characters. <br/> Output Screenshot: <br/> ![alt tag](https://github.com/vgaurav3011/Average-Screen-Time-Calculation-of-Actors-Cartoon-Characters/blob/master/Output.png)
50
143
0.803077
eng_Latn
0.982461
87b36e326f0bc58f23690a93b33631a3389e7585
4,175
md
Markdown
articles/load-balancer/load-balancer-tcp-reset.md
changeworld/azure-docs.tr-tr
a6c8b9b00fe259a254abfb8f11ade124cd233fcb
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/load-balancer/load-balancer-tcp-reset.md
changeworld/azure-docs.tr-tr
a6c8b9b00fe259a254abfb8f11ade124cd233fcb
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/load-balancer/load-balancer-tcp-reset.md
changeworld/azure-docs.tr-tr
a6c8b9b00fe259a254abfb8f11ade124cd233fcb
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Azure'da Boşta Bakiyesi TCP Sıfırlama titleSuffix: Azure Load Balancer description: Bu makalede, boşta zaman diliminde çift yönlü TCP RST paketleri ne olacaksa Azure Yük Dengeleyicisi hakkında bilgi edinin. services: load-balancer documentationcenter: na author: asudbring ms.custom: seodec18 ms.service: load-balancer ms.devlang: na ms.topic: article ms.tgt_pltfrm: na ms.workload: infrastructure-services ms.date: 05/03/2019 ms.author: allensu ms.openlocfilehash: d3d836ddea8d07a25ad09e6f19d9f17a680decd6 ms.sourcegitcommit: 2ec4b3d0bad7dc0071400c2a2264399e4fe34897 ms.translationtype: MT ms.contentlocale: tr-TR ms.lasthandoff: 03/28/2020 ms.locfileid: "80294395" --- # <a name="load-balancer-with-tcp-reset-on-idle"></a>Boşta TCP Sıfırlama ile Yük Dengeleyici Belirli bir kural için Boşta TCP Reset'i etkinleştirerek senaryolarınız için daha öngörülebilir bir uygulama davranışı oluşturmak için [Standart Yük Dengeleyicisini](load-balancer-standard-overview.md) kullanabilirsiniz. Yük Dengeleyici'nin varsayılan davranışı, bir akışın boşta zaman akıtılmasına ulaşıldığında akışları sessizce düşürmektir. Bu özelliği etkinleştirmek, Yük Dengeleyici'nin boşta zaman acısı üzerine çift yönlü TCP Sıfırlamaları (TCP RST paketi) göndermesine neden olur. Bu, uygulama bitiş noktalarınıza bağlantının zaman aşımına geldiğini ve artık kullanılabilir olmadığını bildirir. Uç noktalar gerekirse hemen yeni bir bağlantı kurabilir. ![Yük Dengeleyici TCP sıfırlama](media/load-balancer-tcp-reset/load-balancer-tcp-reset.png) Bu varsayılan davranışı değiştirir ve Gelen NAT kurallarında, yük dengeleme kurallarında ve [giden kurallarda](https://aka.ms/lboutboundrules)tcp sıfırlamalarını boşta zaman adabına göndermeyi etkinleştirin. Kural başına etkinleştirildiğinde, Yük Dengeleyicisi, tüm eşleşen akışlar için boşta kalma zaman sonu sırasında hem istemci hem de sunucu uç noktalarına çift yönlü TCP Sıfırlama (TCP RST paketleri) gönderir. TCP RST paketlerini alan uç noktalar ilgili soketi hemen kapatır. Bu, bağlantının yayımlandığını ve aynı TCP bağlantısındaki gelecekteki iletişimin başarısız olacağını uç noktalara anında bildirir. Uygulamalar, soket kapandığında bağlantıları temizleyebilir ve TCP bağlantısının sonunda zaman dolmasını beklemeden bağlantıları gerektiği gibi yeniden kurabilir. Birçok senaryo için bu, bir akışın boşta zaman anına yenilemek için TCP (veya uygulama katmanı) keepalives gönderme gereksinimini azaltabilir. Boşta kalan süreleriniz yapılandırmanın izin verdiği süreleri aşıyorsa veya uygulamanız TCP Sıfırlamaları etkinleştirilmişolan istenmeyen bir davranış gösteriyorsa, TCP bağlantılarının canlılığını izlemek için Yine de TCP keepalives (veya uygulama katmanı tutma) kullanmanız gerekebilir. Ayrıca, bağlantı yolda bir yere yakın olduğunda, özellikle uygulama katmanı keepalives için de yararlı kalabilir. TCP Resets'i etkinleştirme, boşta zaman ayarı ayarlama ve istenen uygulama davranışını sağlamak için ek adımlar gerekip gerekmediğine karar vermek için tüm uçtan bitiş senaryosunu dikkatle inceleyin. ## <a name="enabling-tcp-reset-on-idle-timeout"></a>Boşta zaman anında TCP Sıfırlama'yı etkinleştirme API sürüm 2018-07-01'i kullanarak, kural bazında boşta zaman diliminde çift yönlü TCP Sıfırlamalarının gönderilmesini etkinleştirebilirsiniz: ```json "loadBalancingRules": [ { "enableTcpReset": true | false, } ] ``` ```json "inboundNatRules": [ { "enableTcpReset": true | false, } ] ``` ```json "outboundRules": [ { "enableTcpReset": true | false, } ] ``` ## <a name="region-availability"></a><a name="regions"></a>Bölge kullanılabilirliği Tüm bölgelerde mevcuttur. ## <a name="limitations"></a>Sınırlamalar - TCP RST yalnızca TCP bağlantısı sırasında ESTABLISHED durumunda gönderilir. ## <a name="next-steps"></a>Sonraki adımlar - Standart [Yük Dengeleyicisi](load-balancer-standard-overview.md)hakkında bilgi edinin. - Giden [kurallar](load-balancer-outbound-rules-overview.md)hakkında bilgi edinin. - [TCP RST'yi Boşta Zaman Aşımda Yapılandır](load-balancer-tcp-idle-timeout.md)
52.1875
662
0.793293
tur_Latn
0.999634
87b3dcd1e3ec068a52ab9a399ced01020297f578
9,273
md
Markdown
_posts/2017-02-17-Spark-source-code.md
colinxiong/colinxiong.github.io
5e2113de747b1ffadcadd3381811da038e5c3519
[ "Apache-2.0" ]
null
null
null
_posts/2017-02-17-Spark-source-code.md
colinxiong/colinxiong.github.io
5e2113de747b1ffadcadd3381811da038e5c3519
[ "Apache-2.0" ]
null
null
null
_posts/2017-02-17-Spark-source-code.md
colinxiong/colinxiong.github.io
5e2113de747b1ffadcadd3381811da038e5c3519
[ "Apache-2.0" ]
null
null
null
--- published: true layout: post title: Spark源码阅读笔记 category: spark tags: - 分布式系统 time: 2017.02.17 11:36:00 excerpt: Spark源码再次详细阅读笔记。 --- ## 一、Spark的启动 Master和Worker启动后会确定Master,启动RPC通信框架(RPCEndPoint),Metric子系统,关联WebUI等。Master会获取每个Worker的资源信息,如内存、核数。具体为Master类和Worker类。 ## 二、 作业的提交 提交作业会启动一个新的JVM,即Driver。Driver上执行作业的初始化、通信初始化、作业的调度等操作。 ### 1、SparkSubmit runMain(): SparkSubmit根据--class参数的名字,通过反射来执行用户提交的类的main方法。【1.5以后版本已经不使用deploy.Client类了。】 Spark Application首先会实例化一个SparkContext对象,该对象会初始化一些重要的对象: (1)StandaloneSchedulerBackend(RPC通信、调度,不同模式下实例化为不同对象) (2)DAGScheduler(作业划分) (3)TaskSchedulerImpl(任务管理和任务调度) (4)BlockManager(数据管理) (5)MemoryManager(内存管理) ### 2、StandaloneSchedulerBackend 该类的主要作用是与Master通信,获取Executor的信息。具体工作由其中的Client对象完成,Client对象对应的类为AppClient【最新Spark中称为StandAloneAppClient】。AppClient会发送ApplicationDescription给Master,让Master分配资源。具体流程如下: - StandaloneSchedulerBackend生成ApplicationDescription,包含了application的名字,对executor的内存、核数的需求等信息。信息关联给APPClient。 - 启动AppClient,AppClient向Master发送RegisterApplication请求。 - Master收到RegisterApplication,注册Application,开始分配资源。具体的方法为schedule()方法:洗牌worker,根据worker的资源总量是否够用逐个launchDriver(),最后startExecutorOnWorker。 - launchDriver,Master通知Worker上LaunchDriver,Worker上启动DriverRunner管理该App对应的Driver。 - startExecutorOnWorkers,优先满足第一个app(FIFO),分配内存和核数。核数的分配采取均分到所有worker上的原则,先每个worker上分配1个executor,如果每个worker上核数小于executor的核数(尽管spark配置的总核数大于executor所需核数),不会分配executor。 - launchExecutor,Master通知Worker上LaunchExecutor,Worker上启动ExecutorRunner来启动Executor。具体的启动方法是:第一步的ApplicationDescription里包含了一个Command,其中记录了启动的Executor的命令;实际上命令中启动的是CoarseExecutorBackend类;CoarseExecutorBackend启动RPC框架,而不是利用Worker通信,发送RegisterExecutorResponse给Driver;Driver返回RegisterExecutor信息;CoarseExecutorBackend创建新的Executor。 - ExecutorAdded,Master通知Driver已经添加了Executor。 至此,Executor已经分配到了Woker上,即资源均已经准备完成。StandaloneSchedulerBackend后续承载了Driver与Master、Worker上的通信,包括Executor Lost、Remove Application等消息的传递。用户代码继续往下执行,直到RDD的action操作触发Job的提交,才真正的提交作业。 ## 三、作业的执行 通过SparkSubmit提交作业后,SparkContext实例化后,资源的分配已经在配置中决定,因此首先完成资源分配。然后由SparkContext来提交Job,由DAGScheduler来控制Job的执行,TaskScheduler调度和管理Task。 ### 1、DAGScheduler RDD中的Action操作触发SparkContext的runJob方法时,执行的是DAGScheduler的submitJob方法。submit后实际上是发送一个JobSubmit给自己,然后执行handleJobSubmit方法。这个方法的流程是: - 创建JobID。 - 根据提交的RDD(称为FinalRDD,因为该RDD触发了Action操作,前面的RDD均还未执行)创建ResultStage,然后submitStage。 - 获取提交的Stage的Missing parent Stage,也就是DAG图中指向该Stage的Parent Stage。如果没有Parent Stage,直接提交该Stage;如果有没处理的Parent Stage,继续调用submitStage方法提交Parent Stage。因此,该操作最终会有一个没有未处理的Parent Stage的Stage被提交。调用submitMissingTasks执行该Stage。 - 提交的Stage根据Partition创建Task,ShuffleMapStage每个partition创建一个ShuffleMapTask,ResultStage每个partition创建一个ResultTask。Task中需要包含以下信息:stage的id,stage的attemptId,用于序列化广播的taskBinary(记录了对应的RDD和shuffleDependency),处理的partition,task执行的偏好(本地,同集群,同机架,任何),Accumulator。 - 将创建的Tasks封装为TaskSet,交给TaskScheduler调度。 在DAGScheduler的工作中,最主要的还是Stage的划分。划分后的Stage是提交执行的依据,也是划分Task的关键。Job的最后一个Stage是ResultStage,ResultStage的Parent Stage都是ShuffleMapStage。那么,具体如何划分Stage呢?细化一下前文中的第二步和第三步: - 调用newResultStage()方法,根据FinalRDD创建ResultStage。 - 调用getParentStages()方法,获取FinalRDD的Parent Stages。方法是从FinalRDD开始,遍历其Dependencies,如果是ShuffleDependency,则调用getShuffleMapStage()方法从shuffleDependency中获取相应的stage;如果不是shuffleDependency,则继续向前遍历其依赖的RDD。只有找到ShuffledRDD,其Dependency才是shuffleDependency;或者HadoopRDD,没有依赖的RDD。 - 如何从ShuffleDependency中获取ShuffleMapStage?如果之前已经创建过了,直接返回;如果没有创建过,不仅仅是返回该parent stage,还需要找出shuffleDependency.rdd的所有parent dependencies及祖先dependencies,判断这些dependencies对应的stage是否cache了结果,有cache结果的话需要更新到对应的stage中,然后再返回当前依赖的parent stage。 ### 2、TaskSchedulerImpl TaskScheduler的实际对象,该对象生成后完成以下工作:生成SchedulerBuilder。根据调度模式,FIFO或者FAIR(FAIR下需要修改配置文件),生成调度器。该对象被创建后,SparkContext会调用start()启动推测执行的线程(如果打开了推测执行)。 TaskScheduler开始调度Task是从DAGScheduler调用submitTasks开始的,工作流程如下: - 创建TaskSetManager。 - 通知StandaloneSchedulerBackend发送一个Reviveoffers给自己(StandaloneSchedulerBackend继承自CoarseGrainedSchedulerBackend的方法reviveOffers())。【最新版本的Spark将消息的接收整体移到了CoarseGrainedSchedulerBackend中】 - TaskSchedulerImpl为每个Executor生成一个WorkerOffer,记录了executor的id,executor的host和cores。WorkerOffer重新洗牌,根据核数创建TaskDescription的数组,按核数依次填充数组,返回给StandaloneSchedulerBackend。 - StandaloneSchedulerBackend调用launchTask将每个Executor对应的TaskDescription的数组序列化发送给Executor(通过CoarseExecutorBackend发送),消息类型为LaunchTask。 - CoarseExecutorBackend接收LaunchTask消息及消息中序列化的TaskDescription,通知Executor执行launchTask()。 - Executor创建TaskRunner(继承自Runnable)作为执行的Task,放到threadPool中执行。 Driver会根据WorkerOffer的信息一条条的发送TaskDescription,Executor最终在线程池中创建core数量的TaskRunner。 ## 四、任务的执行 Job根据partition和stage划分为Task后,有两种实例化类型:ShuffleMapTask和ResultTask。Task分发到Executor后执行的流程如下: - 每个Task创建自己的TaskMemoryManager。 - 从序列化的TaskDescription中反序列化出Task信息,这时可以得到Task具体是ShuffleMapTask还是ResultTask。 - 关联并更新Task的Metrics。 - 执行Task,返回结果(具体返回什么后面讨论)。 - 序列化结果,通过更新Task状态返回给Driver。 - Driver接收StatusUpdate信息,如果是TaskFinish,更新executor的可用核数,重新调用makeOffers()来launchTask。 Task的执行即runTask()方法,根据实例化对象的不同,该方法被重写。 ### 1、ShuffleMapTask ShuffleMapTask定义了一个ShuffleWriter,用于将RDD的计算结果写入磁盘。这时调用的是rdd.iterator()方法,如果有Cache就不需要继续向前计算了,如果没有就调用compute()或者checkpoint(),依次向前遍历。 由于是以Stage为单位提交的,所以,如果不存在cache的话,向前遍历调用RDD的iterator,最终会有两种情况(ResultTask执行时也是这两种情况): - 碰到HadoopRDD读入数据。 - 碰到Stage的最初始的ShuffledRDD。 HadoopRDD的compute(): 利用RecordReader读入数据,创建(K,V),构造迭代器InterruptibleInterator返回。 ShuffledRDD的compute(): - 获取ShuffleManger(默认为SortShuffleManger),获取BlockStoreShuffleReader。 - ShuffleReader读取数据。读取数据时通过ShuffleBlockFetchIterator完成,因为有可能有结果不在本地。首先通过MapOutputTracker跟踪该shuffle分散在不同Executor上的partition。 - 通过ShuffleBlockFetchIterator表示所有的partition块,每块如果不在local则从远程拉取。 - 如果需要compress则压缩数据,因为这里所有的数据均在内存中,如果很大需要压缩。 - 反序列化partition中的数据流为KV迭代器。 - 如果在Map端已经做了Combine操作,即已经根据K做了V的聚合,则直接定义V的类型,如果没有Combine操作则定义V类型为Nothing;如果是aggregator操作,调用aggregator做聚合,聚合使用ExternalAppendOnlyMap(重要)。 - 如果定义了key ordering,则需要按K值排序,排序使用ExternalSort(重要)。排序算法使用TimSort。 该过程即Shuffle的ShuffleRead阶段。 最终compute返回一个Iterator。ShuffleMapTask在执行完成后写入最后计算的一个RDD,即compute返回的Iterator。通过ShuffleWriter写入ExternalSort,然后写入磁盘。写入时,按partition写入,所有的partition写入一个文件,通过长度偏移定位。这里的关键问题是partition的使用,因为MapOutputTracker利用partitioner决定ShuffleMapTask的输出分为多少了partition。paritioner出现的时刻为RDD执行shuffle操作时,RDD会生成一个新的HashPartitioner,窄依赖关系的操作不会更改partition。 ### 2、ResultTask ResultTask执行时的起始RDD情况和ShuffleMapTask相同,在最后处理结果时有所不同。ResultTask中传入了Function对象,直接在RDD上执行该方法,例如count(),saveAsTxtFile()等。 ## 五、作业/任务的调度 CoarseGrainedSchedulerBackend分发TaskDescription时,按照TaskScheduler提供的Tasks来分发,因此调度规则位于TaskScheduler的resourceOffers()方法中,具体为rootPool.getSortedTaskSetQueue()中对TaskSets排序这一部分。在Pool中定义了taskSetSchedulingAlgorithm决定对TaskSet中的task进行排序的comparator,即调度的顺序。 - FIFOSchedulingAlgorithm:先比较两个TaskSetManager的优先级,相同则比较stageId。 - FairSchedulingAlgorithm:依次比较minShare(最小启动Task的比例),已经运行的Task数目,TaskSet设置的weight,TaskSetManager的名字。 如果SparkContext定义了多个pool,那么pool之间采用FairSchedulingAlgorithm;pool内部采用FIFO或者Fair。 推测执行机制:如果打开了推测执行机制,那么会启动一个线程定期的检查是否有需要推测执行的Task,方法为TaskSchedulerImpl中的checkSpeculatableTasks()方法。该方法在TaskSetManager中查找运行时间大于阈值/已经执行完Task的中位数时间,的Task,标记为SpeculatableTasks。为SpeculatbleTask提供资源启动副本任务。 ## 六、数据管理 在SparkContext初始化时,会初始化BlockManager(env.blockManager)来管理RDD所对应的数据。每个Executor上都会启动一个BlockManager,BlockManager同时会初始化一个BlockTransferService,用来拉取远程的Block。 Task的执行实际上是根据partitionID来获取输入数据和写入输出数据,其本质上是通过BlockManager获取的。具体关联流程如下: - Task执行RDD的某个partition,调用RDD.getOrCompute()方法。 - 依据partitionId生成RDDBlockId,与partitionId对应。 - 调用BlockManager.getOrElseUpdate()方法,参数中指定该block的storageLevel(与归属的RDD相同)。 - 如果已经存在的Block,直接返回改block;否则,计算该RDD,调用doPutIterator()将迭代器记录进BlockManager,然后返回迭代器。 BlockManager针对的是ShuffleMapTask的最后一个RDD,所以中间计算的RDD并不会分配给BlockManager存储。这里写入BlockManager与ShuffleWriter的写入的区别是:RDD计算出来后就会在BlockManager中更新BlockId的信息,然后返回这个结果Iterator,然后给ShuffleWriter写入磁盘。 在沿RDD链向上计算的过程中,如果RDD的StorageLevel中有useMemory,即需要persist/cache到内存,需要MemoryStore将数据存储到内存,调用的是MemoryStore.putIteratorAsValues()方法。MemoryStore向MemoryManager申请StorageMemory来将该block保存到内存,保存的结构是一个LinkedHashMap&lt;BlockId, MemoryEntry>类型,BlockId对应persist的block,MemoryEntry对应cache的数据。【目前最新的master分支上没有使用CacheManager来管理Cache的数据了。】 ## 七、内存管理 前文中有两处地方提到了内存管理,一个是在Executor上生成Task时会初始化一个TaskMemoryManager,一个是MemoryStore持久化block时向MemoryManager申请内存。MemoryManager是在SparkContext的SparkEnv中实例化的,默认的实例化对象为UnifiedMemoryManager(统一内存管理)。【1.6以前版本默认为StaticMemoryManger静态内存管理。】 MemoryManager通过两个Pool:StorageMemoryPool和ExecutionMemoryPool来控制Memory的分配。同时,MemoryManager保证N个运行的Task中,每个Task所占用的内存大于1/2N而小于1/N。 - StorageMemoryPool与MemoryStore关联,MemoryStore中每次cache block或者drop block时都会首先请求MemoryManager,如果StorageMemoryPool总空间不够,会从ExecutionMemoryPool划分(StaticMemoryManager不会动态划分)。如果StorageMemoryPool空间空闲部分不够会释放不需要的block,然后分配并更新StorageMemoryPool中memoryUsed的值。 - 在Memory中有一部分unroll memory,用来存放临时的block数据。BlockManager将这block持久化到内存时需要这一部分空间,UnifiedMemoryManager将这部分空间的请求归属到了StorageMemoryPool中。 - ExecutionMemoryPool的大小除了分配给StorageMemoryPool会动态更新外,外界不会主动更新该值。 - UNSAFE/Tungsten模式下,MemoryManager会利用page来管理block数据,TaskMemoryManager通过请求ExecutionMemoryPool中的page来获取偏移量。 TaskMemoryManager如何使用MemoryManager呢?非Tungsten模式下,TaskMemoryManager并没有显式的调用acquire的一些方法,也就是说,实际上MemoryManager除了管理StorageMemoryPool和ExecutionMemoryPool的大小动态变化,以及整体上适应JVM的堆大小,并没有很细粒度的管理。而TaskMemoryManager中的getMemoryConsumptionForThisTask()实际上就只是一个按1/2N和1/N的比例依次递增Task的消耗而已,并没有实际的代表意义。
57.240741
324
0.901003
yue_Hant
0.833108
87b43b47b74cb99ff0ae1202e275624f96217da8
7,274
md
Markdown
microsoft-365/security/defender-identity/manage-security-alerts.md
MicrosoftDocs/microsoft-365-docs-pr.it-IT
20e99e77908118b241bf6b1cb684218b4430b35d
[ "CC-BY-4.0", "MIT" ]
7
2020-03-07T06:40:10.000Z
2022-03-13T12:33:36.000Z
microsoft-365/security/defender-identity/manage-security-alerts.md
MicrosoftDocs/microsoft-365-docs-pr.it-IT
20e99e77908118b241bf6b1cb684218b4430b35d
[ "CC-BY-4.0", "MIT" ]
2
2022-02-09T06:49:11.000Z
2022-02-09T06:49:26.000Z
microsoft-365/security/defender-identity/manage-security-alerts.md
MicrosoftDocs/microsoft-365-docs-pr.it-IT
20e99e77908118b241bf6b1cb684218b4430b35d
[ "CC-BY-4.0", "MIT" ]
2
2019-10-09T19:21:13.000Z
2020-08-28T16:13:59.000Z
--- title: Avvisi di sicurezza di Microsoft Defender for Identity in Microsoft 365 Defender description: Informazioni su come gestire ed esaminare gli avvisi di sicurezza emessi da Microsoft Defender per l'identità in Microsoft 365 Defender ms.date: 05/20/2021 ms.topic: how-to author: dcurwin ms.author: dacurwin ms.service: microsoft-defender-for-identity manager: raynew ms.openlocfilehash: 98df694002d31e330fff1b5d53618044bc6c5dae ms.sourcegitcommit: d08fe0282be75483608e96df4e6986d346e97180 ms.translationtype: MT ms.contentlocale: it-IT ms.lasthandoff: 09/12/2021 ms.locfileid: "59207345" --- # <a name="defender-for-identity-security-alerts-in-microsoft-365-defender"></a>Avvisi di sicurezza di Defender for Identity in Microsoft 365 Defender **Si applica a:** - Microsoft 365 Defender - Defender per identità In questo articolo vengono illustrate le nozioni di base su come utilizzare gli avvisi di sicurezza di [Microsoft Defender for Identity](/defender-for-identity) in [Microsoft 365 Defender](/microsoft-365/security/defender/overview-security-center). Gli avvisi di Defender for Identity sono integrati in modo nativo in [Microsoft 365 Defender](https://security.microsoft.com) con un formato di pagina di avviso identità dedicato. Questo rappresenta il primo passaggio del percorso per introdurre l'esperienza completa di [Microsoft Defender for Identity in Microsoft 365 Defender](/defender-for-identity/defender-for-identity-in-microsoft-365-defender). La nuova pagina di avviso identità offre ai clienti di Microsoft Defender for Identity un miglioramento del segnale tra domini e nuove funzionalità di risposta automatica alle identità. Garantisce la sicurezza e consente di migliorare l'efficienza delle operazioni di sicurezza. Uno dei vantaggi dell'analisi degli avvisi tramite [Microsoft 365 Defender](/microsoft-365/security/defender/microsoft-365-defender) è che gli avvisi di Microsoft Defender for Identity sono ulteriormente correlati alle informazioni ottenute da ognuno degli altri prodotti della famiglia di prodotti. Questi avvisi migliorati sono coerenti con gli altri formati Microsoft 365 Defender di avviso provenienti da [Microsoft Defender per Office 365](/microsoft-365/security/office-365-security) e Microsoft Defender for [Endpoint.](/microsoft-365/security/defender-endpoint) La nuova pagina elimina in modo efficace la necessità di passare a un altro portale del prodotto per analizzare gli avvisi associati all'identità. Gli avvisi provenienti da Defender for Identity possono ora attivare le funzionalità di analisi e risposta automatizzate [(AIR)](/microsoft-365/security/defender/m365d-autoir) di Microsoft 365 Defender, tra cui la correzione automatica degli avvisi e la mitigazione di strumenti e processi che possono contribuire all'attività sospetta. > [!IMPORTANT] > Come parte della convergenza con Microsoft 365 Defender, alcune opzioni e dettagli sono cambiati rispetto alla posizione nel portale defender per l'identità. Leggi i dettagli seguenti per scoprire dove trovare sia le funzionalità familiari che le nuove funzionalità. ## <a name="review-security-alerts"></a>Esaminare gli avvisi di sicurezza È possibile accedere agli avvisi da più posizioni, tra cui la pagina Avvisi, la pagina Eventi imprevisti, le pagine dei singoli dispositivi e dalla **pagina Ricerca** avanzata. In questo esempio verrà esaminata la **pagina Avvisi**. In [Microsoft 365 Defender](https://security.microsoft.com/), passare a **Eventi imprevisti & avvisi** e quindi a **Avvisi**. ![Vai a Eventi imprevisti e avvisi, quindi Avvisi.](../../media/defender-identity/incidents-alerts.png) Per visualizzare gli avvisi di Defender for Identity, in alto a destra seleziona **Filtro** e quindi **in** Origini dei servizi seleziona Microsoft Defender for **Identity** e seleziona **Applica:** ![Filtra defender per gli eventi Identity.](../../media/defender-identity/filter-defender-for-identity.png) Gli avvisi vengono visualizzati con informazioni nelle colonne seguenti: Nome avviso, Tag, Gravità, Stato indagine, **Stato,** **Categoria,** Origine **rilevamento,** Asset con **impatto,** Prima attività e **Ultima attività.** ![Eventi Defender for Identity.](../../media/defender-identity/filtered-alerts.png) ## <a name="manage-alerts"></a>Gestire gli avvisi Se si fa clic **sul nome** dell'avviso per uno degli avvisi, si passa alla pagina con i dettagli relativi all'avviso. Nel riquadro sinistro verrà visualizzato un riepilogo di **Cosa è successo:** ![Cosa è successo nell'avviso.](../../media/defender-identity/what-happened.png) Sopra la **casella Cosa è** successo sono presenti i pulsanti **Account,** **Host di destinazione** e Host **di** origine dell'avviso. Per altri avvisi, potresti visualizzare pulsanti per informazioni dettagliate su host, account, indirizzi IP, domini e gruppi di sicurezza aggiuntivi. Seleziona una di queste per ottenere ulteriori dettagli sulle entità coinvolte. Nel riquadro destro, vedrai i dettagli **dell'avviso.** Qui è possibile visualizzare ulteriori dettagli ed eseguire diverse attività: - **Classificare questo avviso-** Qui è possibile designare questo avviso come **avviso True o** **False** ![Classificare l'avviso.](../../media/defender-identity/classify-alert.png) - **Stato avviso:** in **Imposta classificazione** è possibile classificare l'avviso **come True** o **False.** In **Assegnato a** è possibile assegnare l'avviso a se stessi o annullarne l'assegnazione. ![Stato dell'avviso.](../../media/defender-identity/alert-state.png) - Dettagli **avviso-** In Dettagli avviso **è** possibile trovare ulteriori informazioni sull'avviso specifico, seguire un collegamento alla documentazione relativa al tipo di avviso, vedere l'evento imprevisto a cui è associato l'avviso, esaminare eventuali indagini automatizzate collegate a questo tipo di avviso e vedere i dispositivi e gli utenti a cui è stato generato l'impatto. ![Dettagli dell'avviso.](../../media/defender-identity/alert-details.png) - **Commenti &** cronologia: qui è possibile aggiungere i commenti all'avviso e visualizzare la cronologia di tutte le azioni associate all'avviso. ![Commenti e cronologia.](../../media/defender-identity/comments-history.png) - **Gestisci avviso-** Se si seleziona **Gestisci** avviso, si passa a un riquadro che consente di modificare: - **Stato:** è possibile scegliere **Nuovo,** **Risolto** **o In corso.** - **Classificazione:** è possibile scegliere **True alert o** False **alert.** - **Commento:** è possibile aggiungere un commento sull'avviso. Se si selezionano i tre punti accanto a Gestisci **avviso,** è possibile consultare un esperto di **minacce,** Esportare l'avviso in un file di Excel o Collega **a** un altro **evento imprevisto.** ![Gestisci avviso.](../../media/defender-identity/manage-alert.png) > [!NOTE] > Nel file Excel ora sono disponibili **due collegamenti:** Visualizza **in Microsoft Defender per** l'identità e Visualizza in Microsoft 365 Defender . Ogni collegamento consente di accedere al portale pertinente e di fornire informazioni sull'avviso. ## <a name="see-also"></a>Vedere anche - [Analizzare gli avvisi in Microsoft 365 Defender](../defender/investigate-alerts.md)
77.382979
716
0.780726
ita_Latn
0.997919
87b44ad4f5178a3a2c180c17b7dfbd1623c120e3
729
md
Markdown
_posts/2014-07-04-play-with-locate.md
rohitishere1/rohitishere1.github.io
7e6692703333c90a2721cb0022d584641815c5d4
[ "MIT" ]
null
null
null
_posts/2014-07-04-play-with-locate.md
rohitishere1/rohitishere1.github.io
7e6692703333c90a2721cb0022d584641815c5d4
[ "MIT" ]
null
null
null
_posts/2014-07-04-play-with-locate.md
rohitishere1/rohitishere1.github.io
7e6692703333c90a2721cb0022d584641815c5d4
[ "MIT" ]
null
null
null
--- layout: post title: "Play with locate" description: "A minor tweak with locate which one might require" category: "OS" tags: ["OS","Linux","MacOS"] --- locate command ======================================== We are going to talk about locate utility commonly available on Linux and MacOS. I happen to be a frequent user of locate command and have figured that its data does not get updated that frequently so, if you are using it for any critical purpose, you might have erronous responses. So the way out is manually updating its database, how to do it is below. ***MacOS :*** {% highlight bash %} sudo /usr/libexec/locate.updatedb {% endhighlight %} ***Linux :*** {% highlight bash %} sudo updatedb {% endhighlight %}
28.038462
201
0.691358
eng_Latn
0.997795
87b512848866e788725af498c08dca11ce770ca0
33,867
md
Markdown
docs/authentication.md
dunglumy123/nakama-docs
dee620d2c95a8003db57e5e5bef760a88612131a
[ "Apache-2.0" ]
null
null
null
docs/authentication.md
dunglumy123/nakama-docs
dee620d2c95a8003db57e5e5bef760a88612131a
[ "Apache-2.0" ]
null
null
null
docs/authentication.md
dunglumy123/nakama-docs
dee620d2c95a8003db57e5e5bef760a88612131a
[ "Apache-2.0" ]
1
2020-12-10T02:35:43.000Z
2020-12-10T02:35:43.000Z
# Authentication Server đã có built-in authentication nên client chỉ cần gửi request và kết nối nếu đã có [server key](install-configuration.md#socket) để đăng nhập. Sau khi đăng nhập thành công, client nhận được 1 session tương ứng với [tài khoản người dùng](user-accounts.md). !!! Warning "Quan trọng" Server key mặc định là `defaultkey`, bạn cần đặt 1 giá trị bí mật và [duy nhất](install-configuration.md#socket). Giá trị này sẽ được đưa vào trong client code. ```js tab="JavaScript" var client = new nakamajs.Client("defaultkey", "127.0.0.1", 7350); client.ssl = false; ``` ```csharp tab=".NET" // Use "https" scheme if you've setup SSL. var client = new Client("http", "127.0.0.1", 7350, "defaultkey"); ``` ```csharp tab="Unity" // Use "https" scheme if you've setup SSL. var client = new Client("http", "127.0.0.1", 7350, "defaultkey"); ``` ```cpp tab="Cocos2d-x C++" NClientParameters parameters; parameters.serverKey = "defaultkey"; parameters.host = "127.0.0.1"; parameters.port = DEFAULT_PORT; NClientPtr client = NCocosHelper::createDefaultClient(parameters); ``` ```js tab="Cocos2d-x JS" var serverkey = "defaultkey"; var host = "127.0.0.1"; var port = 7350; var useSSL = false; var timeout = 7000; // ms var client = new nakamajs.Client(serverkey, host, port, useSSL, timeout); ``` ```cpp tab="C++" NClientParameters parameters; parameters.serverKey = "defaultkey"; parameters.host = "127.0.0.1"; parameters.port = DEFAULT_PORT; NClientPtr client = createDefaultClient(parameters); ``` ```java tab="Java" Client client = new DefaultClient("defaultkey", "127.0.0.1", 7349, false) // or same as above. Client client = DefaultClient.defaults("defaultkey"); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let client : Client = Builder("defaultkey") .host("127.0.0.1") .port(7350) .ssl(false) .build() // or same as above. let client : Client = Builder.defaults(serverKey: "defaultkey") ``` Mỗi tài khoản người dùng được tạo từ [các cách đăng nhập](#authenticate). Chúng tôi gọi mỗi cách là "link" bởi vì đó là 1 cách để truy cập vào thông tin của người dùng. Bạn có thể thêm một hoặc nhiều link để cho phép người dùng có thể login theo nhiều cách khác nhau trên nhiều thiết bị. ## Authenticate Trước khi tương tác với server, bạn cần phải lấy được session token bằng cách đăng nhập vào hệ thống. Server hỗ trợ các phương thức authentication rất mềm dẻo, bạn có thể đăng ký user với email, facebook hoặc device. !!! Note "Chú ý" Mặc định hệ thống sẽ tự động tạo tài khoản người dùng nếu thông tin này không có trước đây. Để biết cách đăng ký và đăng nhập, hãy quan sát các ví dụ cho các ngôn ngữ lập trình phổ biến dưới đây. ### Device Định danh thiết bị có thể được sử dụng như một cách để đăng ký người dùng với server. Phương thức này có thể làm người dùng thấy thoải mái vì không phải nhớ mật khẩu nhưng có thể không đáng tin cậy vì số định danh thiết bị đôi khi có thể thay đổi trong bản cập nhật thiết bị. Bạn có thể chọn tên người dùng khi tạo tài khoản. Để làm điều này, hãy đặt giá trị `username`. Nếu bạn chỉ muốn xác thực mà không tạo tài khoản người dùng, hãy đặt `create` là false. Một định danh thiết bị phải chứa các ký tự chữ và số có dấu gạch ngang và nằm trong khoảng từ 10 đến 128 byte. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/authenticate/custom?create=true&username=mycustomusername" \ --user 'defaultkey:' \ --data '{"id":"uniqueidentifier"}' ``` ```js tab="JavaScript" // This import is only required with React Native var deviceInfo = require('react-native-device-info'); var deviceId = null; try { const value = await AsyncStorage.getItem('@MyApp:deviceKey'); if (value !== null){ deviceId = value } else { deviceId = deviceInfo.getUniqueID(); AsyncStorage.setItem('@MyApp:deviceKey', deviceId).catch(function(error){ console.log("An error occured: %o", error); }); } } catch (error) { console.log("An error occured: %o", error); } const session = await client.authenticateDevice({ id: deviceId, create: true, username: "mycustomusername" }); console.info("Successfully authenticated:", session); ``` ```csharp tab=".NET" // Should use a platform API to obtain a device identifier. var deviceId = System.Guid.NewGuid().ToString(); var session = await client.AuthenticateDeviceAsync(deviceId); System.Console.WriteLine("New user: {0}, {1}", session.Created, session); ``` ```csharp tab="Unity" var deviceId = PlayerPrefs.GetString("nakama.deviceid"); if (string.IsNullOrEmpty(deviceId)) { deviceId = SystemInfo.deviceUniqueIdentifier; PlayerPrefs.SetString("nakama.deviceid", deviceId); // cache device id. } var session = await client.AuthenticateDeviceAsync(deviceId); Debug.LogFormat("New user: {0}, {1}", session.Created, session); ``` ```cpp tab="Cocos2d-x C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { CCLOG("Successfully authenticated"); }; std::string deviceId = "unique device id"; client->authenticateDevice( deviceId, opt::nullopt, opt::nullopt, {}, loginSucceededCallback, loginFailedCallback); ``` ```js tab="Cocos2d-x JS" var deviceId = "unique device id"; client.authenticateDevice({ id: deviceId, create: true, username: "mycustomusername" }) .then(function(session) { cc.log("Authenticated successfully. User id:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { cout << "Successfully authenticated" << endl; }; std::string deviceId = "unique device id"; client->authenticateDevice( deviceId, opt::nullopt, opt::nullopt, {}, loginSucceededCallback, loginFailedCallback); ``` ```java tab="Java" String id = UUID.randomUUID().toString(); Session session = client.authenticateDevice(id).get(); System.out.format("Session: %s ", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let defaults = UserDefaults.standard let deviceKey = "device_id" var deviceId : String? = defaults.string(forKey: deviceKey) if deviceId == nil { deviceId = UIDevice.current.identifierForVendor!.uuidString defaults.set(deviceId!, forKey: deviceKey) } let message = AuthenticateMessage(device: deviceId!) client.login(with: message).then { session in print("Login successful") }.catch{ err in if (err is NakamaError) { switch err as! NakamaError { case .userNotFound(_): let _ = self.client.register(with: message) return default: break } } print("Could not login: %@", err) } ``` ```tab="REST" POST /v2/account/authenticate/device?create=true&username=mycustomusername Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Basic base64(ServerKey:) { "id": "uniqueidentifier" } ``` Trong các trò chơi, nên sử dụng [Google](#google) hoặc [Game Center](#game-center) để đăng ký người dùng. ### Email Người dùng có thể đăng ký bằng email và mật khẩu. Mật khẩu được hash trước khi lưu trữ trong database và không thể được đọc hoặc "recovered" bởi các quản trị viên. Bạn có thể yên tâm về bảo vệ sự riêng tư của người dùng. Bạn có thể chọn tên người dùng khi tạo tài khoản. Để làm điều này, đặt giá trị `username`. Nếu bạn muốn chỉ xác thực mà không tạo tài khoản người dùng, hãy đặt `create` là false. Địa chỉ email phải hợp lệ theo quy định của RFC-5322 và mật khẩu phải có ít nhất 8 ký tự. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/authenticate/email?create=true&username=mycustomusername" \ --user 'defaultkey:' \ --data '{"email":"email@example.com", "password": "3bc8f72e95a9"}' ``` ```js tab="JavaScript" const email = "email@example.com"; const password = "3bc8f72e95a9"; const session = await client.authenticateEmail({ email: email, password: password, create: true, username: "mycustomusername" }) console.info("Successfully authenticated:", session); ``` ```csharp tab=".NET" const string email = "email@example.com"; const string password = "3bc8f72e95a9"; var session = await client.AuthenticateEmailAsync(email, password); System.Console.WriteLine("New user: {0}, {1}", session.Created, session); ``` ```csharp tab="Unity" const string email = "email@example.com"; const string password = "3bc8f72e95a9"; var session = await client.AuthenticateEmailAsync(email, password); Debug.LogFormat("New user: {0}, {1}", session.Created, session); ``` ```cpp tab="Cocos2d-x C++" auto successCallback = [](NSessionPtr session) { CCLOG("Authenticated successfully. User ID: %s", session->getUserId().c_str()); }; auto errorCallback = [](const NError& error) { }; string email = "email@example.com"; string password = "3bc8f72e95a9"; string username = "mycustomusername"; bool create = true; client->authenticateEmail(email, password, username, create, {}, successCallback, errorCallback); ``` ```js tab="Cocos2d-x JS" const email = "email@example.com"; const password = "3bc8f72e95a9"; client.authenticateEmail({ email: email, password: password, create: true, username: "mycustomusername" }) .then(function(session) { cc.log("Authenticated successfully. User ID:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto successCallback = [](NSessionPtr session) { std::cout << "Authenticated successfully. User ID: " << session->getUserId() << std::endl; }; auto errorCallback = [](const NError& error) { }; string email = "email@example.com"; string password = "3bc8f72e95a9"; string username = "mycustomusername"; bool create = true; client->authenticateEmail(email, password, username, create, {}, successCallback, errorCallback); ``` ```java tab="Java" String email = "email@example.com"; String password = "3bc8f72e95a9"; Session session = client.authenticateEmail(email, password).get(); System.out.format("Session: %s ", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let email = "email@example.com" let password = "3bc8f72e95a9" let message = AuthenticateMessage(email: email, password: password) client.register(with: message).then { session in NSLog("Session: %@", session.token) }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } // Use client.login(...) after register. ``` ```tab="REST" POST /v2/account/authenticate/email?create=true&username=mycustomusername Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Basic base64(ServerKey:) { "email": "email@example.com", "password": "3bc8f72e95a9" } ``` ### Đăng nhập bằng tài khoản mạng xã hội Server hỗ trợ rất nhiều mạng xã hội khác nhau. Với mỗi nhà cung cấp, tài khoản người dùng của mạng sẽ được sử dụng để thiết lập người dùng trên Game Platform. Trong một số trường hợp, [friends](social-friends.md) của người dùng cũng sẽ được tải về và thêm vào danh sách friend của họ. #### Facebook Để làm việc với Facebook, bạn cần thêm Facebook SDK vào project. Tham khảo <a href="https://developers.facebook.com/docs/" target="\_blank">tại đây</a>. Thực hiện theo các hướng dẫn về cách tích hợp code. Nếu là project mobile, bạn cần hoàn thành các hướng dẫn riêng cho cả iOS và Android. Bạn có thể chọn tên người dùng khi tạo tài khoản. Để làm điều này, đặt giá trị `username`. Nếu bạn muốn chỉ xác thực mà không tạo tài khoản người dùng, hãy đặt `create` là false. Bạn có thể tùy ý import [bạn bè](social-friends.md) trên Facebook khi xác thực. Để làm điều này, đặt `import` là true. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/authenticate/facebook?create=true&username=mycustomusername&import=true" \ --user 'defaultkey:' \ --data '{"token":"valid-oauth-token"}' ``` ```js tab="JavaScript" const oauthToken = "..."; const session = await client.authenticateFacebook({ token: oauthToken, create: true, username: "mycustomusername", import: true }); console.log("Successfully authenticated:", session); ``` ```csharp tab=".NET" const string oauthToken = "..."; var session = await client.AuthenticateFacebookAsync(oauthToken); System.Console.WriteLine("New user: {0}, {1}", session.Created, session); ``` ```csharp tab="Unity" // using Facebook.Unity; // https://developers.facebook.com/docs/unity/examples#init var perms = new List<string>(){"public_profile", "email"}; FB.LogInWithReadPermissions(perms, async (ILoginResult result) => { if (FB.IsLoggedIn) { var accessToken = Facebook.Unity.AccessToken.CurrentAccessToken; var session = await client.LinkFacebookAsync(session, accessToken); Debug.LogFormat("New user: {0}, {1}", session.Created, session); } }); ``` ```cpp tab="Cocos2d-x C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { CCLOG("Authenticated successfully. User ID: %s", session->getUserId().c_str()); }; std::string oauthToken = "..."; bool importFriends = true; client->authenticateFacebook( oauthToken, "mycustomusername", true, importFriends, {}, loginSucceededCallback, loginFailedCallback); ``` ```js tab="Cocos2d-x JS" const oauthToken = "..."; client.authenticateFacebook({ token: oauthToken, create: true, username: "mycustomusername", import: true }) .then(function(session) { cc.log("Authenticated successfully. User ID:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { cout << "Authenticated successfully. User ID: " << session->getUserId() << endl; }; std::string oauthToken = "..."; bool importFriends = true; client->authenticateFacebook( oauthToken, "mycustomusername", true, importFriends, {}, loginSucceededCallback, loginFailedCallback); ``` ```java tab="Java" String oauthToken = "..."; Session session = client.authenticateFacebook(oauthToken).get(); System.out.format("Session %s", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let oauthToken = "..." let message = AuthenticateMessage(facebook: oauthToken) client.register(with: message).then { session in NSLog("Session: %@", session.token) }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } ``` ```tab="REST" POST /v2/account/authenticate/facebook?create=true&username=mycustomusername&import=true Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Basic base64(ServerKey:) { "token": "...", } ``` Bạn có thể thêm một button vào giao diện của mình để đăng nhập bằng Facebook. ```csharp tab="Unity" FB.Login("email", (ILoginResult result) => { if (FB.IsLoggedIn) { var oauthToken = Facebook.Unity.AccessToken.CurrentAccessToken.TokenString; var session = await client.LinkFacebookAsync(session, accesstoken); Debug.LogFormat("Session: '{0}'.", session.AuthToken); }); ``` #### Google Tương tự như Facebook để đăng ký và đăng nhập, bạn cần sử dụng một trong các client SDK của Google. Bạn có thể chọn tên người dùng khi tạo tài khoản. Để làm điều này, đặt giá trị `username`. Nếu bạn muốn chỉ xác thực mà không tạo tài khoản người dùng, hãy đặt `create` là false. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/authenticate/google?create=true&username=mycustomusername" \ --user 'defaultkey:' \ --data '{"token":"valid-oauth-token"}' ``` ```js tab="JavaScript" const playerIdToken = "..."; const session = await client.authenticateGoogle({ token: oauthToken, create: true, username: "mycustomusername" }); console.info("Successfully authenticated: %o", session); ``` ```csharp tab=".NET" const string playerIdToken = "..."; var session = await client.AuthenticateGoogleAsync(playerIdToken); System.Console.WriteLine("New user: {0}, {1}", session.Created, session); ``` ```csharp tab="Unity" const string playerIdToken = "..."; var session = await client.AuthenticateGoogleAsync(playerIdToken); Debug.LogFormat("New user: {0}, {1}", session.Created, session); ``` ```cpp tab="Cocos2d-x C++" auto successCallback = [](NSessionPtr session) { CCLOG("Authenticated successfully. User ID: %s", session->getUserId().c_str()); }; auto errorCallback = [](const NError& error) { }; string oauthToken = "..."; client->authenticateGoogle(oauthToken, "mycustomusername", true, {}, successCallback, errorCallback); ``` ```js tab="Cocos2d-x JS" const oauthToken = "..."; client.authenticateGoogle({ token: oauthToken, create: true, username: "mycustomusername" }) .then(function(session) { cc.log("Authenticated successfully. User ID:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto successCallback = [](NSessionPtr session) { std::cout << "Authenticated successfully. User ID: " << session->getUserId() << std::endl; }; auto errorCallback = [](const NError& error) { }; string oauthToken = "..."; client->authenticateGoogle(oauthToken, "mycustomusername", true, {}, successCallback, errorCallback); ``` ```java tab="Java" String playerIdToken = "..."; Session session = client.authenticateGoogle(oauthToken).get(); System.out.format("Session %s", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let playerIdToken = "..." let message = AuthenticateMessage(google: oauthToken) client.register(with: message).then { session in NSLog("Session: %@", session.token) }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } ``` ```tab="REST" POST /v2/account/authenticate/google?create=true&username=mycustomusername Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Basic base64(ServerKey:) { "token": "...", } ``` ### Custom id Một __custom id__ có thể được sử dụng theo cách tương tự như __device id__ để đăng nhập hoặc đăng ký. Bạn chỉ cần cung cấp 1 custom id duy nhất cho 1 người dùng.. Một __custom id__ phải chứa các ký tự chữ và số có dấu gạch ngang và nằm trong khoảng từ 6 đến 128 byte. Bạn có thể chọn tên người dùng khi tạo tài khoản. Để làm điều này, hãy đặt giá trị `username`. Nếu bạn muốn chỉ xác thực mà không tạo tài khoản người dùng, hãy đặt `create` là false. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/authenticate/custom?create=true&username=mycustomusername" \ --user 'defaultkey:' \ --data '{"id":"some-custom-id"}' ``` ```js tab="JavaScript" const customId = "some-custom-id"; const session = await client.authenticateCustom({ id: customId, create: true, username: "mycustomusername" }); console.info("Successfully authenticated:", session); ``` ```csharp tab=".NET" const string customId = "some-custom-id"; var session = await client.AuthenticateCustomAsync(customId); System.Console.WriteLine("New user: {0}, {1}", session.Created, session); ``` ```csharp tab="Unity" const string customId = "some-custom-id"; var session = await client.AuthenticateCustomAsync(customId); Debug.LogFormat("New user: {0}, {1}", session.Created, session); ``` ```cpp tab="Cocos2d-x C++" auto successCallback = [](NSessionPtr session) { CCLOG("Authenticated successfully. User ID: %s", session->getUserId().c_str()); }; auto errorCallback = [](const NError& error) { }; string id = "some-custom-id"; string username = "mycustomusername"; bool create = true; client->authenticateCustom(id, username, create, {}, successCallback, errorCallback); ``` ```js tab="Cocos2d-x JS" const customId = "some-custom-id"; client.authenticateCustom({ id: customId, create: true, username: "mycustomusername" }) .then(function(session) { cc.log("Authenticated successfully. User ID:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto successCallback = [](NSessionPtr session) { std::cout << "Authenticated successfully. User ID: " << session->getUserId() << std::endl; }; auto errorCallback = [](const NError& error) { }; string id = "some-custom-id"; string username = "mycustomusername"; bool create = true; client->authenticateCustom(id, username, create, {}, successCallback, errorCallback); ``` ```java tab="Java" String customId = "some-custom-id"; Session session = client.authenticateCustom(customId).get(); System.out.format("Session %s", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let customID = "some-custom-id" let message = AuthenticateMessage(custom: customID) client.register(with: message).then { session in NSLog("Session: %@", session.token) }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } // Use client.login(...) after register. ``` ```tab="REST" POST /v2/account/authenticate/custom?create=true&username=mycustomusername Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Basic base64(ServerKey:) { "id": "some-custom-id", } ``` ## Sessions Api đăng nhập thành công sẽ trả lại thông tin về session. Session chứa các thông tin về userId, userName cho đến khi hết hạn. !!! Tip "Mẹo" Bạn có thể thay đổi thời gian hết hạn (expire) của sessiong ở mục [configuration](install-configuration.md) của Console server. Thời gian hết hạn mặc định là 60 giây. ```js tab="JavaScript" const id = "3e70fd52-7192-11e7-9766-cb3ce5609916"; const session = await client.authenticateDevice({ id: id }) console.info("id:", session.user_id, "username:", session.username); console.info("Session expired?", session.isexpired(Date.now() / 1000)); ``` ```csharp tab=".NET" const string id = "3e70fd52-7192-11e7-9766-cb3ce5609916"; var session = await client.AuthenticateDeviceAsync(id); System.Console.WriteLine("Id '{0}' Username '{1}'", session.UserId, session.Username); System.Console.WriteLine("Session expired? {0}", session.IsExpired); ``` ```csharp tab="Unity" var deviceId = SystemInfo.deviceUniqueIdentifier; var session = await client.AuthenticateDeviceAsync(deviceId); Debug.LogFormat("Id '{0}' Username '{1}'", session.UserId, session.Username); Debug.LogFormat("Session expired? {0}", session.IsExpired); ``` ```cpp tab="Cocos2d-x C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { CCLOG("id %s username %s", session->getUserId().c_str(), session->getUsername().c_str()); CCLOG("Session expired? %s", session->isExpired() ? "yes" : "no"); }; std::string deviceId = "3e70fd52-7192-11e7-9766-cb3ce5609916"; client->authenticateDevice( deviceId, opt::nullopt, opt::nullopt, {}, loginSucceededCallback, loginFailedCallback); ``` ```js tab="Cocos2d-x JS" var deviceId = "3e70fd52-7192-11e7-9766-cb3ce5609916"; client.authenticateDevice({ id: deviceId }) .then(function(session) { cc.log("Authenticated successfully. User id:", session.user_id); }, function(error) { cc.error("authenticate failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto loginFailedCallback = [](const NError& error) { }; auto loginSucceededCallback = [](NSessionPtr session) { cout << "id " << session->getUserId() << " username " << session->getUsername() << endl; cout << "Session expired? " << (session->isExpired() ? "yes" : "no") << endl; }; std::string deviceId = "3e70fd52-7192-11e7-9766-cb3ce5609916"; client->authenticateDevice( deviceId, opt::nullopt, opt::nullopt, {}, loginSucceededCallback, loginFailedCallback); ``` ```java tab="Java" var deviceid = SystemInfo.deviceUniqueIdentifier; Session session = client.authenticateDevice(deviceid).get(); System.out.format("Session %s", session.getAuthToken()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let id = "3e70fd52-7192-11e7-9766-cb3ce5609916" let message = AuthenticateMessage(device: id) client.login(with: message).then { session in let expired = session.isExpired(currentTimeSince1970: Date().timeIntervalSince1970) NSLog("Session id '%@' handle '%@'.", session.userID, session.handle) NSLog("Session expired: '%@'", expired) }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } ``` ### Connect Với session, bạn có thể gọi API trên server, trao đổi message với server. Tuy nhiên, phần lớn các client đều không tự động kết nối lại (auto-reconnect). Do đó, bạn cần phải tự thực hiện auto-reconnect. Hãy tham khảo các đoạn code dưới đây. ```js tab="JavaScript" var socket = client.createSocket(); session = await socket.connect(session); console.info("Socket connected."); ``` ```csharp tab=".NET" var socket = Socket.From(client); await socket.ConnectAsync(session); System.Console.WriteLine("Socket connected."); ``` ```csharp tab="Unity" var socket = client.NewSocket(); await socket.ConnectAsync(session); Debug.Log("Socket connected."); ``` ```cpp tab="Cocos2d-x C++" #include "NakamaCocos2d/NWebSocket.h" int port = 7350; // different port to the main API port bool createStatus = true; // if the server should show the user as online to others. // define realtime client in your class as NRtClientPtr rtClient; rtClient = client->createRtClient(port, NRtTransportPtr(new NWebSocket())); // define listener in your class as NRtDefaultClientListener listener; listener.setConnectCallback([]() { CCLOG("Socket connected."); }); rtClient->setListener(&listener); rtClient->connect(session, createStatus); ``` ```js tab="Cocos2d-x JS" const socket = client.createSocket(); socket.connect(session) .then( function() { cc.log("Socket connected."); }, function(error) { cc.error("connect failed:", JSON.stringify(error)); } ); ``` ```cpp tab="C++" int port = 7350; // different port to the main API port bool createStatus = true; // if the server should show the user as online to others. // define realtime client in your class as NRtClientPtr rtClient; rtClient = client->createRtClient(port); // define listener in your class as NRtDefaultClientListener listener; listener.setConnectCallback([]() { cout << "Socket connected." << endl; }); rtClient->setListener(&listener); rtClient->connect(session, createStatus); ``` ```java tab="Java" SocketClient socket = client.createSocket(); socket.connect(session, new AbstractSocketListener() {}).get(); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let session : Session = someSession // obtained from register or login. client.connect(with: session).then { _ in NSLog("Socket connected.") }); ``` ### Expiry Session có thể hết hạn và trở thành không hợp lệ. Nếu điều này xảy ra, bạn cần phải xác nhận lại với server và nhận session mới. Bạn có thể điều chỉnh thời gian tồn tại của session trên server bằng cmdflag "--session.token_exiry_sec 604800". Xem trang [Cấu hình session] (#install-configuration.md#session). Bạn có thể kiểm tra thời hạn của phiên bằng code sau: ```js tab="JavaScript" const nowUnixTime = Math.floor(Date.now() / 1000); if (session.isexpired(nowUnixTime)) { console.log("Session has expired. Must reauthenticate!"); } ``` ```csharp tab=".NET" var nowUnixTime = DateTime.UtcNow; if (session.HasExpired(nowUnixTime)) { System.Console.WriteLine("Session has expired. Must reauthenticate!"); } ``` ```csharp tab="Unity" var nowUnixTime = DateTime.UtcNow; if (session.HasExpired(nowUnixTime)) { Debug.Log("Session has expired. Must reauthenticate!"); } ``` ```cpp tab="Cocos2d-x C++" if (session->isExpired()) { CCLOG("Session has expired. Must reauthenticate!"); } ``` ```js tab="Cocos2d-x JS" const nowUnixTime = Math.floor(Date.now() / 1000); if (session.isexpired(nowUnixTime)) { cc.log("Session has expired. Must reauthenticate!"); } ``` ```cpp tab="C++" if (session->isExpired()) { cout << "Session has expired. Must reauthenticate!"; } ``` ```java tab="Java" if (session.isExpired(new Date())) { System.out.println("Session has expired. Must reauthenticate!"); } ``` ## Link, unlink Bạn có thể liên kết một hoặc nhiều tùy chọn đăng nhập khác với người dùng hiện tại. Điều này giúp dễ dàng hỗ trợ nhiều lần đăng nhập với mỗi người dùng và dễ dàng xác định người dùng trên các thiết bị. Bạn chỉ có thể liên kết các device Id, custom Ids, social provider Ids... chưa được sử dụng với tài khoản người dùng khác. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/link/custom" \ --header 'Authorization: Bearer $session' \ --data '{"id":"some-custom-id"}' ``` ```js tab="JavaScript" const customId = "some-custom-id"; const success = await client.linkCustom(session, { id: customId }); console.log("Successfully linked custom ID to current user."); ``` ```csharp tab=".NET" const string customId = "some-custom-id"; await client.LinkCustomAsync(session, customId); System.Console.WriteLine("Id '{0}' linked for user '{1}'", customId, session.UserId); ``` ```csharp tab="Unity" const string customid = "some-custom-id"; await client.LinkCustomAsync(session, customId); Debug.LogFormat("Id '{0}' linked for user '{1}'", customId, session.UserId); ``` ```cpp tab="Cocos2d-x C++" auto linkFailedCallback = [](const NError& error) { }; auto linkSucceededCallback = []() { CCLOG("Linked successfully"); }; std::string customid = "some-custom-id"; client->linkCustom(customid, linkSucceededCallback, linkFailedCallback); ``` ```js tab="Cocos2d-x JS" const customId = "some-custom-id"; client.linkCustom(session, { id: customId }) .then(function() { cc.log("Linked successfully"); }, function(error) { cc.error("link failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto linkFailedCallback = [](const NError& error) { }; auto linkSucceededCallback = []() { cout << "Linked successfully" << endl; }; std::string customid = "some-custom-id"; client->linkCustom(customid, linkSucceededCallback, linkFailedCallback); ``` ```java tab="Java" String customId = "some-custom-id"; client.linkCustom(session, customId).get(); System.out.format("Id %s linked for user %s", customId, session.getUserId()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let id = "some-custom-id" var message = SelfLinkMessage(device: id); client.send(with: message).then { NSLog("Successfully linked device ID to current user.") }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } ``` ```tab="REST" POST /v2/account/link/custom Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Bearer <session token> { "id":"some-custom-id" } ``` You can unlink any linked login options for the current user. ```sh tab="cURL" curl "http://127.0.0.1:7350/v2/account/unlink/custom" \ --header 'Authorization: Bearer $session' \ --data '{"id":"some-custom-id"}' ``` ```js tab="JavaScript" const customId = "some-custom-id"; const success = await client.unlinkCustom(session, { id: customId }); console.info("Successfully unlinked custom ID from the current user."); ``` ```csharp tab=".NET" const string customId = "some-custom-id"; await client.UnlinkCustomAsync(session, customId); System.Console.WriteLine("Id '{0}' unlinked for user '{1}'", customId, session.UserId); ``` ```csharp tab="Unity" const string customId = "some-custom-id"; await client.UnlinkCustomAsync(session, customId); Debug.LogFormat("Id '{0}' unlinked for user '{1}'", customId, session.UserId); ``` ```cpp tab="Cocos2d-x C++" auto unlinkFailedCallback = [](const NError& error) { }; auto unlinkSucceededCallback = []() { CCLOG("Successfully unlinked custom ID from the current user."); }; std::string customid = "some-custom-id"; client->unlinkCustom(customid, unlinkSucceededCallback, unlinkFailedCallback); ``` ```js tab="Cocos2d-x JS" const customId = "some-custom-id"; client.unlinkCustom(session, { id: customId }) .then(function() { cc.log("Successfully unlinked custom ID from the current user."); }, function(error) { cc.error("unlink failed:", JSON.stringify(error)); }); ``` ```cpp tab="C++" auto unlinkFailedCallback = [](const NError& error) { }; auto unlinkSucceededCallback = []() { cout << "Successfully unlinked custom ID from the current user." << endl; }; std::string customid = "some-custom-id"; client->unlinkCustom(customid, unlinkSucceededCallback, unlinkFailedCallback); ``` ```java tab="Java" String customId = "some-custom-id"; client.unlinkCustom(session, customId).get(); System.out.format("Id %s unlinked for user %s", customId, session.getUserId()); ``` ```swift tab="Swift" // Requires Itme-platform 1.x let id = "some-custom-id" var message = SelfUnlinkMessage(device: id); client.send(with: message).then { NSLog("Successfully unlinked device ID from current user.") }.catch { err in NSLog("Error %@ : %@", err, (err as! NakamaError).message) } ``` ```tab="REST" POST /v2/account/unlink/custom Host: 127.0.0.1:7350 Accept: application/json Content-Type: application/json Authorization: Bearer <session token> { "id":"some-custom-id" } ``` Bạn có thể link hoặc unlink nhiều tuỳ chọn tài khoản khác nhau. | Link | Mô tả | | ---- | ----------- | | Custom | Một định danh tùy chỉnh từ một dịch vụ nhận dạng khác. | | Device | Một định danh duy nhất cho một thiết bị thuộc về người dùng. | | Email | Một email và mật khẩu được đặt bởi người dùng. | | Facebook | Một tài khoản xã hội Facebook. Bạn có thể tùy chọn nhập Bạn bè trên Facebook khi liên kết. | | Google | Một tài khoản Google. |
29.838767
307
0.699088
vie_Latn
0.685161
87b5408fe06ab78c905acbab7846dd14cb358112
34,183
md
Markdown
docs/sql-server/sql-server-2016-release-notes.md
gordonmeurer/sql-docs.de-de
61bb002cf1c70b9d1969c506ed3b7ca8a9c08d01
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/sql-server/sql-server-2016-release-notes.md
gordonmeurer/sql-docs.de-de
61bb002cf1c70b9d1969c506ed3b7ca8a9c08d01
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/sql-server/sql-server-2016-release-notes.md
gordonmeurer/sql-docs.de-de
61bb002cf1c70b9d1969c506ed3b7ca8a9c08d01
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Versionsanmerkungen zu SQL Server 2016|Microsoft-Dokumente ms.date: 04/25/2018 ms.prod: sql ms.reviewer: '' ms.custom: '' ms.technology: install ms.topic: conceptual helpviewer_keywords: - build notes - release issues ms.assetid: c64077a2-bec8-4c87-9def-3dbfb1ea1fb6 author: craigg-msft ms.author: craigg manager: jhubbard monikerRange: = sql-server-2016 || = sqlallproducts-allversions ms.openlocfilehash: fc441d3247d5320e0a9913c0df48cc2573f22858 ms.sourcegitcommit: 50b60ea99551b688caf0aa2d897029b95e5c01f3 ms.translationtype: HT ms.contentlocale: de-DE ms.lasthandoff: 11/15/2018 ms.locfileid: "51700472" --- # <a name="sql-server-2016-release-notes"></a>Versionsanmerkungen zu SQL Server 2016 [!INCLUDE[tsql-appliesto-ss2016-xxxx-xxxx-xxx-md](../includes/tsql-appliesto-ss2016-xxxx-xxxx-xxx-md.md)] Im folgenden Artikel werden Einschränkungen und Probleme mit Releases von SQL Server 2016, Service Packs inbegriffen, beschrieben. Informationen zu Neuerungen finden Sie unter [Neues im Berichts-Generator für SQL Server 2016](https://docs.microsoft.com/sql/sql-server/what-s-new-in-sql-server-2016). - [![Download aus dem Evaluation Center](../includes/media/download2.png)](https://www.microsoft.com/evalcenter/evaluate-sql-server-2016) Laden Sie SQL Server 2016 aus dem **[Evaluation Center](https://www.microsoft.com/evalcenter/evaluate-sql-server-2016)** herunter. - [![Azure Virtual Machine (klein)](../includes/media/azure-vm.png)](https://azure.microsoft.com/marketplace/partners/microsoft/sqlserver2016sp1standardwindowsserver2016/) Haben Sie ein Azure-Konto? Wechseln Sie anschließend **[hierhin](https://azure.microsoft.com/marketplace/partners/microsoft/sqlserver2016sp1standardwindowsserver2016/)** , um einen virtuellen Computer zu starten, auf dem SQL Server 2016 SP1 bereits installiert ist. - [![SSMS herunterladen](../includes/media/download2.png)](../ssms/download-sql-server-management-studio-ssms.md) Wechseln Sie zu **[Herunterladen von SQL Server Management Studio](../ssms/download-sql-server-management-studio-ssms.md)**, um die neueste Version von SQL Server Management Studio (SSMS) abzurufen. ## <a name="bkmk_2016sp2"></a>SQL Server 2016 Service Pack 2 (SP2) ![info_tip](../sql-server/media/info-tip.png) SQL Server 2016 SP2 enthält alle kumulativen Updates bis einschließlich CU8, die nach SQL Server 2016 SP1 veröffentlicht wurden. - [![Microsoft Download Center](../includes/media/download2.png)](https://go.microsoft.com/fwlink/?linkid=869608) [Herunterladen von SQL Server 2016 Service Pack 2 (SP2)](https://go.microsoft.com/fwlink/?linkid=869608) - Eine vollständige Liste der Updates finden Sie unter [SQL Server 2016 Service Pack 2 release information (Releaseinformationen zu SQL Server 2016 Service Pack 2)](https://support.microsoft.com/help/4052908/sql-server-2016-service-pack-2-release-information). Sie müssen Ihren Computer möglicherweise nach der Installation von SQL Server 2016 SP2 neu starten. Als bewährte Methode wird empfohlen, nach der Installation von SQL Server 2016 SP2 einen Neustart zu planen und durchzuführen. Leistungs- und skalierungsbasierte Verbesserungen, die in SQL Server 2016 SP2 enthalten sind |Funktion|und Beschreibung|Weitere Informationen| | --- | --- | --- | |Verbesserter Cleanupvorgang für die Verteilungsdatenbank | Übergroße Verteilungsdatenbanktabellen führten zu Blockierungs- und Deadlocksituationen. Im Rahmen einer verbesserten Cleanup-Prozedur sollen einige davon beseitigt werden. | [KB4040276](https://support.microsoft.com/help/4040276/fix-indirect-checkpoints-on-the-tempdb-database-cause-non-yielding) | |Bereinigen der Änderungsnachverfolgung | Verbesserte Leistung und Effizienz des Cleanups für Nebentabellen bei der Änderungsnachverfolgung. | [KB4052129](https://support.microsoft.com//help/4052129/update-for-manual-change-tracking-cleanup-procedure-in-sql-server-2016) | |Verwenden von CPU-Timeout zum Abbrechen von Resource Governor-Anforderungen | Verbessert die Handhabung von Abfrageanforderungen, indem die Anforderung tatsächlich abgebrochen wird, falls CPU-Schwellenwerte für eine Anforderung erreicht werden. Dieses Verhalten wird unter dem Ablaufverfolgungsflag 2422 aktiviert. | [KB4038419](https://support.microsoft.com/help/4038419/add-cpu-timeout-to-resource-governor-request-max-cpu-time-sec) | |Verwenden von SELECT INTO zum Erstellen einer Zieltabelle in der Dateigruppe | Ab SQL Server 2016 SP2 unterstützt die T-SQL-Syntax SELECT INTO das Laden einer Tabelle in eine Dateigruppe, die nicht die Standarddateigruppe des Benutzers ist, die das Schlüsselwort ON <Filegroup name> in der T-SQL-Syntax verwendet. | | |Verbesserter indirekter Prüfpunkt für tempdb | Das Verwenden eines indirekten Prüfpunkts für tempdb wurde verbessert, um die Spinlock-Konflikte für DPLists zu reduzieren. Diese Verbesserung erlaubt der tempdb-Workload in SQL Server 2016 eine sofort einsatzbereite zentrale Hochskalierung, falls die Verwendung des indirekten Prüfpunkts für tempdb ON (Ein) ist. | [KB4040276](https://support.microsoft.com/help/4040276) | |Verbesserte Leistung bei der Datenbanksicherung auf Computern mit großem Speicher | SQL Server 2016 SP2 optimiert die Art und Weise, wie wir die fortlaufende E/A während der Sicherung ausgleichen, wodurch wir eine deutliche Leistungssteigerung bei der Sicherung für kleine bis mittlere Datenbanken erreichen. Wenn Sicherungen der Systemdatenbank auf einem Computer mit 2 TB durchgeführt werden, ist die Leistung stark verbessert. Die Leistungssteigerung sinkt mit steigender Datenbankgröße, da die zu sichernden Seiten und die Sicherungs-E/A im Vergleich zur Iteration des Pufferpools mehr Zeit in Anspruch nehmen. Durch diese Änderung können Sie die Sicherungsleistung für Kunden optimieren, die mehrere kleine Datenbanken auf hochwertigen Servern mit hoher Speicherkapazität hosten. | | |Unterstützung für die VDI-Sicherungskomprimierung für mit TDE aktivierte Datenbanken | SQL Server 2016 SP2 fügt VDI-Support hinzu, damit VDI-Sicherungslösungen die Komprimierung für mit TDE aktivierte Datenbanken nutzen können. Durch diese Verbesserung wurde ein neues Sicherungsformat eingeführt, um die Sicherungskomprimierung für mit TDE-aktivierte Datenbanken zu unterstützen. Die SQL Server-Engine verarbeitet transparent neue und alte Sicherungsformate, um die Sicherungen wiederherzustellen. | | |Dynamisches Laden von Profilparametern des Replikations-Agents | Mithilfe dieser neuen Erweiterungen können Replikations-Agent-Parameter dynamisch geladen werden, ohne dass der Agent neu gestartet werden muss. Diese Änderung gilt nur für die am häufigsten verwendeten Agent-Profilparameter. | | |Support der MAXDOP-Option für die CREATE STATISTICS-Anweisung und UPDATE STATISTICS-Anweisung | Mit dieser Erweiterung kann die MAXDOP-Option für eine CREATE STATISTICS-Anweisung und UPDATE STATISTICS-Anweisung angegeben werden. Zudem kann sichergestellt werden, dass die richtige MAXDOP-Einstellung verwendet wird, wenn Statistiken als Teil eines Erstellungs- oder Neuerstellungsvorgangs für alle Indextypen aktualisiert werden (falls die MAXDOP-Option nicht vorhanden ist). | [KB4041809](https://support.microsoft.com/help/4041809) | |Verbessertes Update für automatische Statistiken für inkrementelle Statistiken | Wenn z.B. mehrere Datenänderungen für mehrere Partitionen in einer Tabelle vorgenommen wurden und die Anzahl aller Änderung für inkrementelle Statistiken – jedoch keine einzelnen Partitionen – den Grenzwert für automatische Updates überschreitet, kann sich das Statistikupdate so lange verzögern, bis weitere Änderungen in der Tabelle vorgenommen werden. Dieses Verhalten wird unter dem Ablaufverfolgungsflag 11024 korrigiert. | | Unterstützbarkeit und Diagnose betreffende Verbesserungen sind in SQL Server 2016 SP2 beinhaltet. |Funktion |und Beschreibung |Weitere Informationen | | --- | --- | --- | |Vollständige DTC-Unterstützung für Datenbanken in einer Verfügbarkeitsgruppe | Datenbankübergreifende Transaktionen für Datenbanken, die Teil einer Verfügbarkeitsgruppe sind, werden für SQL Server 2016 nicht unterstützt. Mit SQL Server 2016 SP2 führen wir die vollständige Unterstützung für verteilte Transaktionen mit Datenbanken der Verfügbarkeitsgruppe ein. | | |Update für die Spalte „is_encrypted“ in „sys.database“ zur genauen Wiedergabe des Verschlüsselungsstatus für TempDB | Der Wert der Spalte „is_encrypted“ in „sys.databases“ beträgt 1 für tempdb, auch wenn Sie die Verschlüsselung für alle Benutzerdatenbanken deaktivieren und SQL Server neu starten. Das erwartete Verhalten ist, dass der Wert 0 (null) ist, da tempdb in dieser Situation nicht länger verschlüsselt ist. Beginnend mit SQL Server 2016 SP2 gibt „is_encrypted“ in „sys.databases“ nun den Verschlüsselungsstatus für tempdb genau wieder. | | |Neue DBCC CLONEDATABASE-Optionen zum Generieren verifizierter Klone und Sicherungsklone | Mit SQL Server 2016 SP2 lässt DBCC CLONEDATABASE jetzt zwei neue Optionen zu: das Erstellen eines verifizierten Klons und das Erstellen eines Sicherungsklons. Wenn eine Klondatenbank mithilfe der Option WITH VERIFY_CLONEDB erstellt wird, wird auch ein konsistenter Datenbankklon erstellt und geprüft, der von Microsoft in der Produktion unterstützt wird. Eine neue Eigenschaft wird eingeführt, um zu überprüfen, ober der Klon die verifizierte Eigenschaft SELECT DATABASEPROPERTYEX(‘clone_database_name’, ‘IsVerifiedClone’) ist. Wenn ein Klon mit der Option BACKUP_CLONEDB erstellt wird, wird eine Sicherung im selben Ordner, in dem sich auch die Datendatei befindet, erstellt. Dadurch können Kunden den Klon leichter auf unterschiedliche Server verschieben oder an den Microsoft-Kundenservice (CSS) zur Problembehandlung senden. | | |Unterstützung für Service Broker (SSB) für DBCC CLONEDATABASE | Der erweiterte DBCC CLONEDATABASE-Befehl zum Zulassen der Skripterstellung von SSB-Objekten | [KB4092075](https://support.microsoft.com/help/4092075) | |Neue dynamische Verwaltungssicht (DMV) zur Überwachung des tempdb-Speicherplatzverbrauchs | Die neue DMV „sys.dm_tran_version_store_space_usage“ wird in SQL Server 2016 SP2 eingeführt, um die Überwachung von tempdb für die Versionsspeichernutzung zu aktivieren. Datenbankadministratoren können tempdb-Größen basierend auf der Anforderung an die Versionsspeichernutzung pro Datenbank proaktiv planen. Dies geschieht ohne Mehraufwand an Leistung, wenn die Ausführung auf Produktionsservern erfolgt. | | |Unterstützung von vollständigen Speicherabbildern für Replikations-Agents | Wenn Replikations-Agents auf Ausnahmefehler stoßen, erstellen sie aktuell standardmäßig ein Miniabbild der Ausnahmesymptome. Dadurch wird die Problembehandlung von Ausnahmefehlern sehr schwierig. Mit dieser Änderung führen wir einen neuen Registrierungsschlüssel ein, der die Erstellung eines vollständigen Speicherabbilds für Replikations-Agents ermöglicht. | | |Verbesserung für erweiterte Ereignisse zum Lesen von Routingfehlern für eine Verfügbarkeitsgruppe | Derzeit wird „read_only_rout_fail xEvent“ nur ausgelöst, wenn zwar eine Routingliste besteht, jedoch zu keinem der Server auf der Liste eine Verbindung hergestellt werden kann. SQL Server 2016 SP2 beinhaltet zusätzliche Informationen zur Unterstützung bei der Problembehandlung und Erweiterungen zu den Codepunkten, bei denen xEvent ausgelöst wird. | | |Neue DMV, um das Transaktionsprotokoll zu überwachen | Die neue DMV „sys.dm_db_log_stats“ wurde hinzugefügt, die zusammenfassende Ebenenattribute und Informationen zu den Transaktionsprotokolldateien von Datenbanken zurückgibt. | | |Neue DMV zum Überwachen von VLF-Informationen | Die neue DMV „sys.dm_db_log_info“ wird in SQL Server 2016 SP2 eingeführt, mit der Sie VLF-Informationen ähnlich wie DBCC LOGINFO verfügbar machen können, um potentielle Transaktionsprotokollprobleme, die von Kunden gemeldet wurden, zu überwachen, davor zu warnen und diese zu vermeiden. | | |Prozessorinformationen in „sys.dm_os_sys_info“| Neue Spalten, die der DMV „sys.dm_os_sys_info“ hinzugefügt wurden, um die sich auf den Prozessor beziehenden Informationen verfügbar zu machen, wie „socket_count und cores_per_numa“ | | |Erweiterung modifizierter Informationen in „sys.dm_db_file_space_usage“| Eine neue Spalte, die „sys.dm_db_file_space_usage“ hinzugefügt wurde, um die Anzahl der geänderten Erweiterungen seit der letzten vollständigen Sicherung nachzuverfolgen | | |Segmentinformation in „sys.dm_exec_query_stats“ | Neue Spalten wurden zu „sys.dm_exec_query_stats“ hinzugefügt, um die Anzahl der übersprungenen und gelesenen Columnstore-Segmente nachzuverfolgen, zum Beispiel „total_columnstore_segment_reads“ und „total_columnstore_segment_skips“. | [KB4051358](https://support.microsoft.com/help/4051358) | |Festlegen des korrekten Kompatibilitätsgrads für die Verteilungsdatenbank | Nach der Installation des Service Packs ändert sich der Kompatibilitätsgrad der Verteilungsdatenbank in 90. Der Grund dafür war ein Codepfad in der gespeicherten Prozedur „sp_vupgrade_replication“. Das Service Pack wurde nun dahingehend geändert, dass es nun einen angemessenen Kompatibilitätsgrad für die Verteilungsdatenbank bestimmt. | | |Verfügbarmachen der letzten bekannten funktionierenden DBCC CHECKDB-Information | Es wurde eine neue Datenbankoption hinzugefügt, damit die Datumsangabe der letzten erfolgreichen Ausführung von DBCC CHECKDB programmgesteuert zurückgegeben werden kann. Benutzer können nun DATABASEPROPERTYEX([database], ‘lastgoodcheckdbtime’) abfragen, um einen einzelnen Wert abzurufen, der das Datum und die Uhrzeit der letzten erfolgreichen DBCC CHECKDB-Ausführung auf der angegebenen Datenbank darstellt. | | |Showplan XML-Erweiterungen| [Informationen darüber, welche Statistiken zum Kompilieren des Abfrageplans verwendet wurden](https://blogs.msdn.microsoft.com/sql_server_team/sql-server-2017-showplan-enhancements/), einschließlich der Name der Statistik, des Änderungszählers und des Prozentsatzes der Stichproben sowie wann das letzte Update für die Statistik stattfand. Gilt nur für CE-Modelle 120 und höher. Dies wird beispielsweise nicht für CE 70 unterstützt.| | | |Showplan XML wird ein neues Attribut, [EstimateRowsWithoutRowgoal](https://blogs.msdn.microsoft.com/sql_server_team/more-showplan-enhancements-row-goal/) hinzugefügt, wenn der Abfrageoptimierer die „row goal“-Logik verwendet.| | | |Die neuen Runtime-Attribute [UdfCpuTime und UdfElapsedTime](https://blogs.msdn.microsoft.com/sql_server_team/more-showplan-enhancements-udfs/) in der tatsächlichen Showplan-XML zum Nachverfolgen der Zeit, die für skalare benutzerdefinierte Funktionen gebraucht wird| | | |Das Hinzufügen des CXPACKET-Wartetyps zur [Liste der möglichen Top 10 der Wartevorgänge](https://blogs.msdn.microsoft.com/sql_server_team/new-showplan-enhancements/) im tatsächlichen Showplan XML-Ereignis: Eine parallele Abfrageausführung erfordert häufig CXPACKET-Wartevorgänge. Dieser Wartevorgangstyp wurde jedoch nicht im aktuellen Showplan XML-Ereignis gemeldet. | | | |Die Warnung vor dem Runtime-Überlauf wird erweitert, um die Anzahl der Seiten, die in tempdb geschrieben wurden, während des Überlaufs eines Parallelitätsoperators zu melden.| | |Replikationsunterstützung für Datenbanken mit ergänzenden Zeichensortierungen | Die Replikation kann nun auf Datenbanken unterstützt werden, die die ergänzende Zeichensortierung verwenden. | | |Verbesserte Handhabung von Service Broker mit Failover einer Verfügbarkeitsgruppe | Wenn Service Broker für Datenbanken der Verfügbarkeitsgruppe in der aktuellen Implementierung aktiviert ist, bleiben bei einem Failover der Verfügbarkeitsgruppe alle Service Broker-Verbindungen, die dem primären Replikat entstammen, geöffnet. Mit dieser Verbesserung sollen all diese offenen Verbindungen während eines Failover der Verfügbarkeitsgruppe geschlossen werden. | | |Verbesserte Problembehandlung für parallele Wartevorgänge | durch Hinzufügen eines neuen [CXCONSUMER](https://blogs.msdn.microsoft.com/sql_server_team/making-parallelism-waits-actionable/)-Wartevorgangs. | | |Verbesserte Konsistenz zwischen DMVs für dieselbe Information | Die DMV „sys.dm_exec_session_wait_stats“ verfolgt nun CXPACKET- und CXCONSUMER-Wartevorgänge konsistent mit der DMV„sys.dm_os_wait_stats“. | | |Verbesserte Problembehandlung von abfrageinternen Parallelitätsdeadlocks | Ein neues erweitertes exchange_spill-Ereignis, um die Anzahl der in tempdb geschriebenen Ereignisse während des Überlaufs eines Parallelitätsoperators im xEvent-Feldnamen „worktable_physical_writes“ zu melden.| | | |Die Überlaufspalten in den DMVs „sys.dm_exec_query_stats“, „sys.dm_exec_procedure_stats“ und „sys.dm_exec_trigger_stats“ (wie z.B. „total_spills“) enthalten jetzt auch die Daten, die von Parallelitätsoperatoren zum Überlauf gebracht werden.| | | |Das XML-Deadlockdiagramm wurde für Szenarios für Parallelitätsdeadlocks verbessert. Der exchangeEvent-Ressource wurden mehr Attribute hinzugefügt.| | | |Das XML-Deadlockdiagramm wurde für Deadlocks verbessert, die Operatoren im Batchmodus einbeziehen, und der SyncPoint-Ressource wurden mehr Attribute hinzugefügt.| | |Dynamisches erneutes Laden einiger Profilparametern des Replikations-Agents | In der aktuellen Implementierung von Replikations-Agents erfordert jede Änderung im Agent-Profilparameter, dass der Agent angehalten und neu gestartet wird. Durch diese Verbesserungen können Parameter dynamisch neu geladen werden, ohne dass der Replikations-Agent neu geladen werden muss. | | ![horizontal-bar.png](media/horizontal-bar.png) ## <a name="bkmk_2016sp1"></a>SQL Server 2016 Service Pack 1 (SP1) ![info_tip](../sql-server/media/info-tip.png) SQL Server 2016 SP1 umfasst alle kumulativen Updates bis SQL Server 2016 RTM CU3 Security einschließlich des Sicherheitsupdates MS16-136. Es enthält ein Rollup der in den kumulativen Updates von SQL Server 2016 bereitgestellten Lösungen und umfasst das aktuelle kumulative Update CU3 sowie das Sicherheitsupdate MS16-136 mit Veröffentlichungsdatum am 8. November 2016. Die folgenden Features sind in der Standard, Web, Express und Local DB Edition von SQL Server SP1 verfügbar (sofern nicht anders angegeben): - Always Encrypted - Geänderte Datenerfassung (in Express nicht verfügbar) - columnstore - Komprimierung - Dynamische Datenmaskierung - Differenzierte Überwachung - In-Memory OLTP (in Local DB nicht verfügbar) - Mehrere FileStream-Container (in Local DB nicht verfügbar) - Partitionierung - PolyBase - Sicherheit auf Zeilenebene In der folgenden Tabelle werden wichtige Verbesserungen in SQL Server 2016 SP1 zusammengefasst. |Funktion|und Beschreibung|Weitere Informationen finden Sie unter| |---|---|---| |Masseneinfügung in Heaps mit automatischem TABLOCK unter TF 715| Das Ablaufverfolgungsflag 715 aktiviert Tabellensperren für Massenladevorgänge in einen Heap ohne nicht gruppierte Indizes.|[Migrating SAP workloads to SQL Server just got 2.5x faster (Beschleunigung der Migration von SAP-Workloads zu SQL Server um das 2,5-fache)](https://blogs.msdn.microsoft.com/sql_server_team/migrating-sap-workloads-to-sql-server-just-got-2-5x-faster/)| |CREATE OR ALTER|Bereitstellen von Objekten wie gespeicherten Prozeduren, Triggern, benutzerdefinierten Funktionen und Ansichten.|[Blog der SQL Server-Datenbank-Engine](https://blogs.msdn.microsoft.com/sqlserverstorageengine/2016/11/17/create-or-alter-another-great-language-enhancement-in-sql-server-2016-sp1/)| |DROP TABLE-Unterstützung für die Replikation|DROP TABLE DDL-Unterstützung für die Replikation zum Löschen von Replikationsartikeln.|[KB 3170123](https://support.microsoft.com/help/3170123/supports-drop-table-ddl-for-articles-that-are-included-in-transactiona)| |Signieren des FileStream RsFx-Treibers|Der Filestream RsFx-Treiber wird im Windows Hardware Developer Center Dashboard-Portal (Dev-Portal) signiert und zertifiziert. Dadurch kann der Filestream RsFx-Treiber von SQL Server 2016 SP1 problemlos unter Windows Server 2016/Windows 10 installiert werden.|[Migrating SAP workloads to SQL Server just got 2.5x faster (Beschleunigung der Migration von SAP-Workloads zu SQL Server um das 2,5-fache)](https://blogs.msdn.microsoft.com/sql_server_team/migrating-sap-workloads-to-sql-server-just-got-2-5x-faster/)| |LPIM für SQL-Dienstkonto – programmgesteuerte Identifikation|Ermöglicht DBAs die programmgesteuerte Identifikation, wenn die Berechtigung zum Sperren von Seiten im Speicher (Lock Pages in Memory, LPIM) zur Startzeit des Service besteht.|[Developers Choice: Programmatically identify LPIM and IFI privileges in SQL Server (Wahl für Entwickler: LPIM- und IFI-Berechtigungen zur programmgesteuerten Identifikation in SQL Server)](https://blogs.msdn.microsoft.com/sql_server_team/developers-choice-programmatically-identify-lpim-and-ifi-privileges-in-sql-server)| |Manuelles Bereinigen der Änderungsnachverfolgung|Die neue gespeicherte Prozedur löscht die interne Tabelle der Änderungsnachverfolgung bedarfsgesteuert.| [KB 3173157](https://support.microsoft.com/help/3173157/adds-a-stored-procedure-for-the-manual-cleanup-of-the-change-tracking)| |Parallele INSERT..SELECT-Änderungen in lokalen temporären Tabellen|Neue parallele INSERT in INSERT..SELECT-Vorgänge.|[SQL Server-Kundenberatungsteam](https://blogs.msdn.microsoft.com/sqlcat/2016/07/21/real-world-parallel-insert-what-else-you-need-to-know/)| |Showplan XML|Erweiterte Diagnose einschließlich einer aktivierten Zuweisungswarnung und eines aktivierten maximalen Arbeitsspeichers für eine Abfrage, aktivierter Ablaufverfolgungsflags und Oberflächen für andere Diagnoseinformationen. | [KB 3190761](https://support.microsoft.com/help/3190761/update-to-improve-diagnostics-by-expose-data-type-of-the-parameters-fo)| |Speicherklassenspeicher|Steigern Sie mit dem Speicherklassenspeicher in Windows Server 2016 die Transaktionsverarbeitung, wodurch die Commitzeiten für Transaktionen nach Größenordnungen beschleunigt werden können.|[Blog der SQL Server-Datenbank-Engine](https://blogs.msdn.microsoft.com/sqlserverstorageengine/2016/12/02/transaction-commit-latency-acceleration-using-storage-class-memory-in-windows-server-2016sql-server-2016-sp1/)| |USE HINT|Verwenden Sie die Abfrageoption `OPTION(USE HINT('<option>'))`, um das Verhalten des Abfrageoptimierers mit unterstützten Hinweisen auf Abfrageebene zu ändern. Anders als bei der Option QUERYTRACEON sind bei der Option USE HINT keine Systemadministratorberechtigungen erforderlich.|[Developers Choice: USE HINT query hints (Wahl der Entwickler: USE HINT-Abfragehinweise)](https://blogs.msdn.microsoft.com/sql_server_team/developers-choice-use-hint-query-hints/)| |XEvent-Ergänzungen|Neue XEvents- und Perfmon-Diagnosefunktionen bewirken eine Optimierung der Wartezeit bei der Problembehandlung.|[Erweiterte Ereignisse](https://docs.microsoft.com/sql/relational-databases/extended-events/extended-events)| Zudem sollten Sie folgende Problembehandlungen beachten: - Basierend auf Feedback von DBAs und der SQL-Community werden die Hekaton-Protokollierungsnachrichten ab SQL 2016 SP1 auf ein Minimum reduziert. - Überprüfen neuer [Ablaufverfolgungsflags](https://docs.microsoft.com/sql/t-sql/database-console-commands/dbcc-traceon-trace-flags-transact-sql). - Die vollständigen Versionen der WideWorldImporters-Beispieldatenbanken können jetzt beginnend mit SQL Server 2016 SP1 mit der Standard Edition und der Express Edition ausgeführt werden und sind auf [Github]( https://github.com/Microsoft/sql-server-samples/releases/tag/wide-world-importers-v1.0) verfügbar. Im Beispiel müssen keine Änderungen vorgenommen werden. Die Datenbanksicherungen, die zum RTM der Enterprise Edition erstellt wurden, werden in SP1 von der Standard und der Express Edition unterstützt. Für die Installation von SQL Server 2016 SP1 ist nach der Installation möglicherweise ein Neustart erforderlich. Als bewährte Methode wird empfohlen, nach der Installation von SQL Server 2016 SP1 einen Neustart zu planen und durchzuführen. ### <a name="download-pages-and-more-information"></a>Downloadseiten und weitere Informationen - [Service Pack 1 für Microsoft SQL Server 2016 herunterladen](https://www.microsoft.com/download/details.aspx?id=54276) - [SQL Server 2016 Service Pack 1 (SP1) freigegeben](https://blogs.msdn.microsoft.com/sqlreleaseservices/sql-server-2016-service-pack-1-sp1-released/) - [Releaseinformationen zu SQL Server 2016 Service Pack 1](https://support.microsoft.com/kb/3182545) - ![info_tip](../sql-server/media/info-tip.png) Links und Informationen zu allen unterstützten Versionen finden Sie unter [SQL Server-Update Center](https://msdn.microsoft.com/library/ff803383.aspx). Dort finden Sie auch Informationen zu Service Packs von [!INCLUDE[ssNoVersion_md](../includes/ssnoversion-md.md)] ![horizontal-bar.png](media/horizontal-bar.png) ## <a name="bkmk_2016_ga"></a> SQL Server 2016 Release - General Availability (GA) - [Datenbank-Engine (GA)](#bkmk_ga_instalpatch) - [Stretch Database (GA)](#bkmk_ga_stretch) - [Abfragespeicher (GA)](#bkmk_ga_query_store) - [Produktdokumentation (GA)](#bkmk_ga_docs) ### ![repl_icon_warn](../database-engine/availability-groups/windows/media/repl-icon-warn.gif) <a name="bkmk_ga_instalpatch"></a> Install Patch Requirement (GA) **Problem und Kundenbeeinträchtigung:** Microsoft hat ein Problem erkannt, das die Microsoft VC++ 2013 Runtime-Binarys betrifft, die von SQL Server 2016 als erforderliche Komponente installiert werden. Ein Update ist verfügbar, um dieses Problem zu beheben. Wenn dieses Update an den VC++ Runtime-Binarys nicht installiert wird, können bei SQL Server 2016 in bestimmten Szenarien Stabilitätsprobleme auftreten. Bevor Sie SQL Server 2016 installieren, überprüfen Sie, ob der Computer das in [KB 3164398](https://support.microsoft.com/kb/3164398)beschriebenen Patch benötigt. Der Patch ist auch im [Kumulativen Updatepaket 1 (CU1) für SQL Server 2016 RTM](https://www.microsoft.com/download/details.aspx?id=53338) enthalten. **Lösung:** Greifen Sie auf eine der folgenden Lösungen zurück: - Installieren Sie [KB 3138367 – Update für Visual C++ 2013 und Visual C++ Redistributable Package](https://support.microsoft.com/kb/3138367). Der KB-Hotfix ist die bevorzugte Lösung. Sie können dies vor oder nach der Installation von SQL Server 2016 installieren. Wenn SQL Server 2016 bereits installiert ist, führen Sie nacheinander die folgenden Schritte aus: 1. Laden Sie die passende *vcredist_\*exe* herunter. 1. Beenden Sie den SQL Server-Dienst für alle Instanzen der Datenbank-Engine. 1. Installieren Sie **KB 3138367**. 1. Starten Sie den Computer neu. - Installieren Sie [KB 3164398 – kritisches Update für erforderliche MSVCRT-Komponenten für SQL Server 2016](https://support.microsoft.com/kb/3164398). Wenn Sie **KB 3164398**verwenden, können Sie während der Installation von SQL Server über Microsoft Update oder Microsoft Download Center installieren. - **Während der Installation von SQL Server 2016:** Wenn der Computer, auf dem Sie SQL Server einrichten, Zugang zum Internet hat, überprüft das SQL Server-Setup im Rahmen der gesamten SQL Server-Installation, ob das Update vorhanden ist. Wenn Sie das Update akzeptieren, lädt das Setupprogramm die Binärdateien im Rahmen der Installation herunter und aktualisiert sie. - **Microsoft Update:** Das Update ist bei Microsoft Update als wichtiges, nicht sicherheitsrelevantes Update für SQL Server 2016 verfügbar. Installation über Microsoft Update, nachdem SQL Server 2016 nach dem Update den Neustart des Servers fordert. - **Download Center:** Das Update steht im Microsoft Download Center zur Verfügung. Sie können die Software für das Update herunterladen und nach der Installation von SQL Server 2016 auf Servern installieren. ### <a name="bkmk_ga_stretch"></a>Stretch Database #### <a name="problem-with-a-specific-character-in-a-database-or-table-name"></a>Problem mit einem bestimmten Zeichen in einem Datenbank- oder Tabellennamen **Problem und Beeinträchtigung des Kunden:** Beim Versuch, Stretch Database auf einer Datenbank oder in einer Tabelle zu aktivieren, ist ein Fehler aufgetreten. Das Problem tritt auf, wenn der Name des Objekts ein Zeichen enthält, das beim Konvertieren von Klein- und Großschreibung als ein anderes Zeichen behandelt wird. Ein Beispiel für ein Zeichen, das dieses Problem auslöst, ist das Zeichen „ƒ“ (durch Drücken von ALT+159 erstellt). **Lösung:** Wenn Sie Stretch Database in einer Datenbank oder Tabelle aktivieren möchten, ist die einzige Option, das Objekt umzubenennen und das Problemzeichen zu entfernen. #### <a name="problem-with-an-index-that-uses-the-include-keyword"></a>Problem mit einem Index, der das Schlüsselwort INCLUDE verwendet **Problem und Kundenbeeinträchtigung:** Der Versuch, Stretch Database in einer Tabelle zu aktivieren, die über einen Index verfügt, der das INCLUDE-Schlüsselwort verwendet, um zusätzliche Spalten in den Index einzubeziehen, resultiert in einem Fehler. **Lösung:** Löschen Sie den Index, der das INCLUDE-Schlüsselwort verwendet, aktivieren Sie Stretch Database in der Tabelle, und erstellen Sie den Index neu. Wenn Sie dies tun, achten Sie darauf, die Wartungsmethoden und Richtlinien Ihrer Organisation einzuhalten, um sicherzustellen, dass für Benutzer der betreffenden Tabelle höchstens minimale Auswirkungen entstehen. ### <a name="bkmk_ga_query_store"></a>Query Store #### <a name="problem-with-automatic-data-cleanup-on-editions-other-than-enterprise-and-developer"></a>Problem mit automatischer Datenbereinigung bei anderen Editionen als Enterprise und Developer **Problem und Kundenbeeinträchtigung:** Fehler bei automatischer Datenbereinigung bei anderen Editionen als Enterprise und Developer. In Folge dessen wächst der vom Abfragespeicher verwendete Speicherplatz mit der Zeit, bis das konfigurierte Limit erreicht ist, wenn Daten nicht manuell gelöscht werden. Wenn dieses Problem nicht beseitigt wird, wird auch der Speicherplatz aufgefüllt, der den Fehlerprotokollen zugeordnet ist, da jeder Bereinigungsversuch eine Dumpdatei erzeugt. Der Aktivierungszeitraum für die Bereinigung hängt von der Häufigkeit der Arbeitsauslastung ab, ist aber nicht länger als 15 Minuten. **Lösung:** Wenn Sie den Abfragespeicher bei anderen Editionen als Enterprise und Developer verwenden möchten, müssen Sie explizit die Bereinigungsrichtlinien deaktivieren. Dies kann über SQL Server Management Studio (Seite „Datenbankeigenschaften“) oder über das Transact-SQL-Skript erfolgen: ```ALTER DATABASE <database name> SET QUERY_STORE (OPERATION_MODE = READ_WRITE, CLEANUP_POLICY = (STALE_QUERY_THRESHOLD_DAYS = 0), SIZE_BASED_CLEANUP_MODE = OFF)``` Darüber hinaus sollten Sie manuelle Bereinigungsoptionen erwägen, um zu verhindern, dass der Abfragespeicher in den schreibgeschützten Modus übergeht. Führen Sie z.B. die folgende Abfrage zum Bereinigen des gesamten Datenbereichs in regelmäßigen Abständen aus: ```ALTER DATABASE <database name> SET QUERY_STORE CLEAR``` Führen Sie außerdem die folgenden gespeicherten Prozeduren des Abfragespeichers regelmäßig aus, um Laufzeitstatistik, bestimmte Abfragen oder Plänen zu bereinigen: - `sp_query_store_reset_exec_stats` - `sp_query_store_remove_plan` - `sp_query_store_remove_query` ### <a name="bkmk_ga_docs"></a> Produktdokumentation (GA) **Problem und Kundenbeeinträchtigung:** Eine herunterladbare Version der Dokumentation zu SQL Server 2016 ist noch nicht verfügbar. Wenn Sie den Hilfebibliotheks-Manager verwenden, um zu versuchen, **Inhalt vom Onlinespeicherort zu installieren**, wird Ihnen die Dokumentationen zu SQL Server 2012 und SQL Server 2014 angezeigt. Allerdings gibt es keine Optionen zum Anzeigen der SQL Server 2016-Dokumentation. **Problemumgehung:** Sie können eine der folgenden Problemumgehungen verwenden: ![Verwalten der Hilfeeinstellungen für SQL Server](../sql-server/media/docs-sql2016-managehelpsettings.png "Verwalten der Hilfeeinstellungen für SQL Server") - Verwenden Sie die Option **Onlinehilfe oder lokale Hilfe auswählen** , und konfigurieren Sie die Hilfe für „Ich möchte Onlinehilfe verwenden“. - Verwenden Sie die Option **Inhalt von Onlinespeicherort installieren** , und laden Sie den Inhalt von SQL Server 2014 herunter. **F1-Hilfe:** Beim Drücken von F1 in [!INCLUDE[ssManStudioFull](../includes/ssmanstudiofull-md.md)] wird standardmäßig die Onlineversion des F1-Hilfeartikels im Browser angezeigt. Für die Probleme wird browserbasierte Hilfe angeboten, auch wenn Sie die lokale Hilfe konfiguriert und installiert haben. **Aktualisieren von Inhalt:** In SQL Server Management Studio und Visual Studio kann die Anwendung Help Viewer während des Hinzufügens der Dokumentation einfrieren (hängen bleiben). Führen Sie zur Lösung des Problems folgende Schritte aus. Weitere Informationen zu diesem Problem finden Sie unter [Visual Studio Help Viewer friert beim Begrüßungsbildschirm ein](https://msdn.microsoft.com/library/mt654096.aspx). * Öffnen Sie die Datei „%LOCALAPPDATA%\Microsoft\HelpViewer2.2\HlpViewer_SSMS16_en-US.settings“ | „HlpViewer_VisualStudio14_en-US.settings“ im Editor, und ändern Sie das Datum im folgenden Code in einen Zeitpunkt in der Zukunft. ``` Cache LastRefreshed="12/31/2017 00:00:00" ``` ## <a name="additional-information"></a>Zusätzliche Informationen + [SQL Server 2016-Installation](../database-engine/install-windows/installation-for-sql-server-2016.md) + [SQL Server-Update Center – Links und Informationen zu allen unterstützten Versionen](https://msdn.microsoft.com/library/ff803383.aspx) [!INCLUDE[get-help-options](../includes/paragraph-content/get-help-options.md)] [!INCLUDE[contribute-to-content](../includes/paragraph-content/contribute-to-content.md)] ![MS_Logo_X-Small](../sql-server/media/ms-logo-x-small.png "MS_Logo_X-Small")
144.84322
934
0.811163
deu_Latn
0.98642
87b6e49746ad3a5770000221986a27f1bbf7eac9
218
md
Markdown
README.md
PixelTheDeveloper/esx_rpchat
19614c62fe1e629472d2b3d54a7c93ab877f8886
[ "MIT" ]
null
null
null
README.md
PixelTheDeveloper/esx_rpchat
19614c62fe1e629472d2b3d54a7c93ab877f8886
[ "MIT" ]
null
null
null
README.md
PixelTheDeveloper/esx_rpchat
19614c62fe1e629472d2b3d54a7c93ab877f8886
[ "MIT" ]
null
null
null
### esx_rpchat ESX RPChat Most Clean Version! ### Original Version https://github.com/PapajONse/esx_rpchat ### Support Me https://discord.gg/fJekXq5jUP ### Commands - ool - ooc - tweet - atweet - pa - dc - ad - oos
11.473684
39
0.688073
kor_Hang
0.511222
87b7340d8c60f16d9df63266579375754fef9a92
1,509
md
Markdown
articles/data-factory/concepts-data-flow-manage-graph.md
GordenW/azure-docs.zh-cn
2b69134b6401663a0fe76e07cd81d97da080bda1
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/data-factory/concepts-data-flow-manage-graph.md
GordenW/azure-docs.zh-cn
2b69134b6401663a0fe76e07cd81d97da080bda1
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/data-factory/concepts-data-flow-manage-graph.md
GordenW/azure-docs.zh-cn
2b69134b6401663a0fe76e07cd81d97da080bda1
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: 数据流图表 description: 如何使用数据工厂数据流图表 author: kromerm ms.author: makromer ms.service: data-factory ms.topic: conceptual ms.custom: seo-lt-2019 ms.date: 11/04/2019 ms.openlocfilehash: 0d357c4c671070a5c5e9d4587e2f90b6628996f4 ms.sourcegitcommit: 877491bd46921c11dd478bd25fc718ceee2dcc08 ms.translationtype: MT ms.contentlocale: zh-CN ms.lasthandoff: 07/02/2020 ms.locfileid: "81605361" --- # <a name="mapping-data-flow-graphs"></a>映射数据流图表 [!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)] 映射数据流设计图面是一个 "构造" 图面,可在其中自上而下、从左向右构建数据流。 具有加号 (+) 符号的每个转换都附加有一个工具箱。 请专注于你的业务逻辑,而非在自由形式的 DAG 环境中通过边缘来连接节点。 下面是用于管理数据流图形的内置机制。 ## <a name="move-nodes"></a>移动节点 ![聚合转换选项](media/data-flow/agghead.png "聚合器标头") 如果不使用拖放模式,"移动" 转换节点的方式是更改传入流。 你将通过更改“传入流”来四处移动转换。 ## <a name="streams-of-data-inside-of-data-flow"></a>数据流中的数据流 在 Azure 数据工厂数据流中,流表示数据的流动。 在 "转换设置" 窗格中,您将看到 "传入流" 字段。 它告诉你哪个传入数据流在为该转换提供数据。 可以通过单击传入流名称并选择另一个数据流在图形上更改转换节点的物理位置。 然后,该流上的当前转换以及所有后续转换都将移动到新位置。 如果要移动的转换后面有一个或多个转换,则会通过一个新分支来联接数据流中的新位置。 如果在你选择的节点后面没有后续转换,则只有该转换会移动到新位置。 ## <a name="hide-graph-and-show-graph"></a>隐藏关系图并显示关系图 底部的 "配置" 窗格的右下角有一个按钮,可以在使用转换配置时将底部窗格扩展到全屏。 这样,便可以使用 "上一步" 和 "下一步" 按钮在图形配置中导航。 若要移回图形视图,请单击 "下一步" 按钮并返回到 "拆分屏幕"。 ## <a name="search-graph"></a>搜索关系图 您可以在设计图面上搜索包含 "搜索" 按钮的关系图。 ![搜索](media/data-flow/search001.png "搜索关系图") ## <a name="next-steps"></a>后续步骤 完成数据流设计后,打开 "调试" 按钮,并在调试模式下直接在 "数据流[设计器](concepts-data-flow-debug-mode.md)" 或 "[管道调试](control-flow-execute-data-flow-activity.md)" 中对其进行测试。
29.019231
142
0.76607
yue_Hant
0.629791
87b740b6175b73a970d4c57cd9e019f83b8105ee
227
md
Markdown
CHANGELOG.md
betagouv/ansible-role-webhook
6c5d8a65776c3f70953c550331ffee169f83791f
[ "MIT" ]
1
2022-02-25T01:04:11.000Z
2022-02-25T01:04:11.000Z
CHANGELOG.md
betagouv/ansible-role-webhook
6c5d8a65776c3f70953c550331ffee169f83791f
[ "MIT" ]
null
null
null
CHANGELOG.md
betagouv/ansible-role-webhook
6c5d8a65776c3f70953c550331ffee169f83791f
[ "MIT" ]
2
2020-12-24T05:36:53.000Z
2021-11-29T12:47:42.000Z
# Changelog ## v1.0.0 - Initial stable release - Added Github Actions ## v0.1.0 - Documented Variables - Added variables for username - Added variables for user group - Added variable for Port - Added variable for Versions
15.133333
32
0.744493
eng_Latn
0.985515
87b7cc5521ac75776883e7a0c6cfb630044a9bbb
4,083
md
Markdown
openwhisk/nl/pt/BR/openwhisk_watson_translator.md
Ranjanasingh9/docs
45286f485cd18ff628e697ddc5091f396bfdc46b
[ "Apache-2.0" ]
null
null
null
openwhisk/nl/pt/BR/openwhisk_watson_translator.md
Ranjanasingh9/docs
45286f485cd18ff628e697ddc5091f396bfdc46b
[ "Apache-2.0" ]
null
null
null
openwhisk/nl/pt/BR/openwhisk_watson_translator.md
Ranjanasingh9/docs
45286f485cd18ff628e697ddc5091f396bfdc46b
[ "Apache-2.0" ]
1
2020-09-24T04:33:34.000Z
2020-09-24T04:33:34.000Z
--- copyright: years: 2016, 2017 lastupdated: "2017-02-21" --- {:shortdesc: .shortdesc} {:new_window: target="_blank"} {:codeblock: .codeblock} {:screen: .screen} {:pre: .pre} # Usando o pacote do Tradutor do Watson {: #openwhisk_catalog_watson_translator} O pacote `/whisk.system/watson-translator` oferece uma maneira conveniente de chamar APIs do Watson para traduzir. O pacote inclui as ações a seguir. | Entidade | Tipo | Parâmetros | Descrição | | --- | --- | --- | --- | | `/whisk.system/watson-translator` | pacote | username, password | Pacote para tradução de texto e identificação de idioma | | `/whisk.system/watson-translator/translator` | ação | payload, translateFrom, translateTo, translateParam, username, password | Traduzir texto | | `/whisk.system/watson-translator/languageId` | ação | payload, username, password | Identificar idioma | **Nota**: o pacote `/whisk.system/watson` está descontinuado, incluindo as ações `/whisk.system/watson/translate` e `/whisk.system/watson/languageId`. ## Configurando o pacote do Tradutor do Watson no Bluemix Se você estiver usando o OpenWhisk a partir do Bluemix, o OpenWhisk criará automaticamente as ligações de pacote para as suas instâncias de serviço do Watson do Bluemix. 1. Crie uma instância de serviço do Tradutor do Watson em seu [painel](http://console.ng.Bluemix.net) do Bluemix. Certifique-se de lembrar do nome da instância de serviço e da organização e do espaço do Bluemix nos quais você está. 2. Atualize os pacotes em seu namespace. A atualização cria automaticamente uma ligação de pacote para a instância de serviço do Watson que você criou. ``` wsk package refresh ``` {: pre} ``` created bindings: Bluemix_Watson_Translator_Credentials-1 ``` ``` wsk package list ``` {: pre} ``` packages /myBluemixOrg_myBluemixSpace/Bluemix_Watson_Translator_Credentials-1 private ``` ## Configurando um pacote do Tradutor do Watson fora do Bluemix Se você não estiver usando o OpenWhisk no Bluemix ou se desejar configurar o seu Tradutor do Watson fora do Bluemix, deverá criar manualmente uma ligação de pacote para o seu serviço de Tradutor do Watson. Você precisa do nome do usuário e da senha do serviço de Tradutor do Watson. - Crie uma ligação de pacote que esteja configurada para o seu serviço de Tradutor do Watson. ``` wsk package bind /whisk.system/watson-translator myWatsonTranslator -p username MYUSERNAME -p password MYPASSWORD ``` {: pre} ## Traduzindo texto A ação `/whisk.system/watson-translator/translator` traduz o texto de um idioma para outro. Os parâmetros são como segue: - `username`: o nome do usuário da API do Watson. - `password`: a senha da API do Watson. - `payload`: o texto a ser traduzido. - `translateParam`: o parâmetro de entrada indicando o texto a ser traduzido. Por exemplo, se `translateParam=payload`, o valor do parâmetro `payload` que é passado à ação será traduzido. - `translateFrom`: um código de dois dígitos do idioma de origem. - `translateTo`: um código de dois dígitos do idioma de destino. - Chame a ação `translator` em sua ligação de pacote para traduzir algum texto do inglês para o francês. ``` wsk action invoke myWatsonTranslator/translator \ --blocking --result \ --param payload "Blue skies ahead" --param translateFrom "en" \ --param translateTo "fr" ``` {: pre} ```json { "payload": "Ciel bleu a venir" } ``` ## Identificando o idioma de algum texto A ação `/whisk.system/watson-translator/languageId` identifica o idioma de algum texto. Os parâmetros são como segue: - `username`: o nome do usuário da API do Watson. - `password`: a senha da API do Watson. - `payload`: o texto para identificar. - Chame a ação `languageId` em sua ligação do pacote para identificar o idioma. ``` wsk action invoke myWatsonTranslator/languageId \ --blocking --result \ --param payload "Ciel bleu a venir" ``` {: pre} ```json { "payload": "Ciel bleu a venir", "language": "fr", "confidence": 0.710906 } ```
32.664
157
0.723243
por_Latn
0.997262
87b8216fca1e1ee065ad635686476d5ace9af792
252
md
Markdown
translations/ja-JP/data/reusables/actions/enterprise-http-proxy.md
nyanthanya/Cuma_Info
d519c49504fc3818c1294f14e63ee944d2f4bd89
[ "CC-BY-4.0", "MIT" ]
17
2021-01-05T16:29:05.000Z
2022-02-26T09:08:44.000Z
translations/ja-JP/data/reusables/actions/enterprise-http-proxy.md
nyanthanya/Cuma_Info
d519c49504fc3818c1294f14e63ee944d2f4bd89
[ "CC-BY-4.0", "MIT" ]
222
2021-04-08T20:13:34.000Z
2022-03-18T22:37:27.000Z
translations/ja-JP/data/reusables/actions/enterprise-http-proxy.md
nyanthanya/Cuma_Info
d519c49504fc3818c1294f14e63ee944d2f4bd89
[ "CC-BY-4.0", "MIT" ]
3
2021-08-31T03:18:06.000Z
2021-10-30T17:49:09.000Z
{% data variables.product.product_location %}で**HTTPプロキシサーバー**を設定してイルなら、**HTTPプロキシの除外**リストに`localhost`と`127.0.0.1`を追加しなければなりません。 プロキシの設定変更に関する詳しい情報については「[アウトバウンドのWebプロキシサーバーの設定](/admin/configuration/configuring-an-outbound-web-proxy-server)」を参照してください。
126
251
0.829365
yue_Hant
0.241696
87b8432302b436d23eb36f42211640e6685d2ea9
334
md
Markdown
src/v6/transitions/warp.md
deulos/vue-flux-docs
52a18b2caf82080dc6950cd44052b9d594abe321
[ "MIT" ]
1
2020-03-27T18:41:53.000Z
2020-03-27T18:41:53.000Z
src/v6/transitions/warp.md
ragnarlotus/vue-flux-docs
52a18b2caf82080dc6950cd44052b9d594abe321
[ "MIT" ]
null
null
null
src/v6/transitions/warp.md
ragnarlotus/vue-flux-docs
52a18b2caf82080dc6950cd44052b9d594abe321
[ "MIT" ]
null
null
null
--- sidebarDepth: 0 --- # Warp ## Description A concentric effect is performed by rotating the image converted into circles in alternate direction. ## Options | Name | Type | Default | |------|------|---------| | circles | Numeric | 7 | | tileDuration | Numeric | 800 | | tileDelay | Numeric | 150 | | easing | String | linear |
17.578947
101
0.628743
eng_Latn
0.95829
87b84c452b8c22265ea96694ae4f8815305d608e
854
md
Markdown
README.md
thebevrishot/satang-pro-signer
e0d6affdaf3b3bf5a670bda160f8a7d341b41707
[ "MIT" ]
null
null
null
README.md
thebevrishot/satang-pro-signer
e0d6affdaf3b3bf5a670bda160f8a7d341b41707
[ "MIT" ]
null
null
null
README.md
thebevrishot/satang-pro-signer
e0d6affdaf3b3bf5a670bda160f8a7d341b41707
[ "MIT" ]
null
null
null
# satang-pro-signer [![Build Status](https://travis-ci.org/thebevrishot/satang-pro-signer.svg?branch=master)](https://travis-ci.org/thebevrishot/satang-pro-signer) [![PyPI version](https://badge.fury.io/py/satang-pro-signer-x.svg)](https://badge.fury.io/py/satang-pro-signer-x) An implementation of Satang Pro request signing scheme. https://docs.satang.pro/authentication ## installation ``` pip install satang-pro-signer-x ``` ## usage ```python import json import satang_pro_signer # import signer # prepare secret secret = bytes.fromhex('8781e58f94f8b2a58b6aa30649fd6a46') # create signer signer = satang_pro_signer.Signer(secret) # prepare payload to be sign payload = json.loads('{"type":"limit","pair":"btc_thb", "side":"sell", "price":"100000", "amount":"100", "none":"1570763737"}') # sign signature = signer.sign(payload) # bytes ```
25.878788
143
0.734192
yue_Hant
0.154351
87b9f98ef84d68106dfdba88d382372f3429cfb6
3,140
md
Markdown
README.md
obiknows/ipfs-nodes
b1ff59af377cb7263412b5807f625dda068152b6
[ "MIT" ]
null
null
null
README.md
obiknows/ipfs-nodes
b1ff59af377cb7263412b5807f625dda068152b6
[ "MIT" ]
null
null
null
README.md
obiknows/ipfs-nodes
b1ff59af377cb7263412b5807f625dda068152b6
[ "MIT" ]
null
null
null
# go-ipfs ![banner](https://ipfs.io/ipfs/QmVk7srrwahXLNmcDYvyUEJptyoxpndnRa57YJ11L4jV26/ipfs.go.png) ## Usage ``` ipfs - Global p2p merkle-dag filesystem. ipfs [<flags>] <command> [<arg>] ... SUBCOMMANDS BASIC COMMANDS init Initialize ipfs local configuration add <path> Add a file to ipfs cat <ref> Show ipfs object data get <ref> Download ipfs objects ls <ref> List links from an object refs <ref> List hashes of links from an object DATA STRUCTURE COMMANDS block Interact with raw blocks in the datastore object Interact with raw dag nodes files Interact with objects as if they were a unix filesystem ADVANCED COMMANDS daemon Start a long-running daemon process mount Mount an ipfs read-only mountpoint resolve Resolve any type of name name Publish or resolve IPNS names dns Resolve DNS links pin Pin objects to local storage repo Manipulate an IPFS repository NETWORK COMMANDS id Show info about ipfs peers bootstrap Add or remove bootstrap peers swarm Manage connections to the p2p network dht Query the DHT for values or peers ping Measure the latency of a connection diag Print diagnostics TOOL COMMANDS config Manage configuration version Show ipfs version information update Download and apply go-ipfs updates commands List all available commands Use 'ipfs <command> --help' to learn more about each command. ipfs uses a repository in the local file system. By default, the repo is located at ~/.ipfs. To change the repo location, set the $IPFS_PATH environment variable: export IPFS_PATH=/path/to/ipfsrepo ``` ### Docker usage An IPFS docker image is hosted at [hub.docker.com/r/ipfs/go-ipfs](https://hub.docker.com/r/ipfs/go-ipfs/). To make files visible inside the container you need to mount a host directory with the `-v` option to docker. Choose a directory that you want to use to import/export files from IPFS. You should also choose a directory to store IPFS files that will persist when you restart the container. export ipfs_staging=</absolute/path/to/somewhere/> export ipfs_data=</absolute/path/to/somewhere_else/> Start a container running ipfs and expose ports 4001, 5001 and 8080: docker run -d --name ipfs_host -v $ipfs_staging:/export -v $ipfs_data:/data/ipfs -p 4001:4001 -p 127.0.0.1:8080:8080 -p 127.0.0.1:5001:5001 ipfs/go-ipfs:latest Watch the ipfs log: docker logs -f ipfs_host Wait for ipfs to start. ipfs is running when you see: Gateway (readonly) server listening on /ip4/0.0.0.0/tcp/8080 You can now stop watching the log. Run ipfs commands: docker exec ipfs_host ipfs <args...> For example: connect to peers docker exec ipfs_host ipfs swarm peers Add files: cp -r <something> $ipfs_staging docker exec ipfs_host ipfs add -r /export/<something> Stop the running container: docker stop ipfs_host ## License MIT
30.192308
163
0.690446
eng_Latn
0.948228
87ba30c3ff6494c519bd6fb81392f8a9771d96b0
7,886
md
Markdown
articles/dataverse/create-update-delete-trigger.md
modery/power-automate-docs
3ae7440345ffb33536f43157cf0ff8076ba47f54
[ "CC-BY-4.0", "MIT" ]
98
2019-11-12T21:50:27.000Z
2022-03-24T10:47:49.000Z
articles/dataverse/create-update-delete-trigger.md
modery/power-automate-docs
3ae7440345ffb33536f43157cf0ff8076ba47f54
[ "CC-BY-4.0", "MIT" ]
566
2019-11-13T01:05:02.000Z
2022-03-31T21:56:41.000Z
articles/dataverse/create-update-delete-trigger.md
modery/power-automate-docs
3ae7440345ffb33536f43157cf0ff8076ba47f54
[ "CC-BY-4.0", "MIT" ]
126
2019-11-13T04:41:24.000Z
2022-03-20T12:22:53.000Z
--- title: Trigger flows when a row is added, modified, or deleted | Microsoft Docs description: "Learn how to trigger flows by using the When a row is added, modified, or deleted trigger." services: '' suite: flow documentationcenter: na author: MSFTMAN manager: KVIVEK ms.author: Deonhe editor: '' tags: '' ms.service: flow ms.devlang: na ms.topic: article ms.tgt_pltfrm: na ms.workload: na ms.date: 09/13/2021 search.app: - Flow - Powerplatform search.audienceType: - maker --- # Trigger flows when a row is added, modified, or deleted The **When a row is added, modified or deleted** trigger runs a flow whenever a row of a selected table and scope changes or is created. ## Prerequisites - To create a flow that triggers when you create, modify, or delete a row, you must have user-level permissions for create, read, write, and delete on the **Callback Registration** table. - Additionally, depending on the scopes defined in the flow, you might need at least that level of read on the same table. You can get more information about [Environment security](/power-platform/admin/database-security). ![Dataverse triggers.](../media/create-update-delete-trigger/triggers.png "Dataverse triggers") The following information is required to use the **When a row is added, modified or deleted** trigger. - Trigger condition - Table name - Scope ### Trigger condition The trigger condition, **Change type**, precisely defines which combination of changes to a row would run the flow. ![Trigger conditions.](../media/create-update-delete-trigger/2.png "Trigger conditions") When the flow is triggered by the creation, update, or deletion of a row, the value of `triggerOutputs()['body/SdkMessage']` will be `Create`, `Update`, or `Delete`, respectively. ### Table name The **Table name** list filters the rows to indicate precisely which kind of rows should change before the flow triggers. See [Tables in Dataverse](/powerapps/maker/common-data-service/entity-overview). ![Select a table name.](../media/create-update-delete-trigger/created-modified-deleted.png "Select a table name") ### Scope The **Scope** list indicates those rows should be monitored to determine if the flow should be run. ![Select scope for triggering flow.](../media/create-update-delete-trigger/scope.png "Select scope for triggering flow") Here’s what each scope means: |**Scope**| **Row ownership level** | |---------| ----------------------- | |Business Unit | Actions are taken on rows owned by anyone in your [business unit](/power-platform/admin/wp-security-cds#business-units). | | Organization | Actions are taken by anyone within the [environment](/power-platform/admin/environments-overview). | | Parent: Child business unit | Actions are taken on rows that are owned by anyone in your [business unit or a child business unit](/power-platform/admin/wp-security-cds#business-units). | | User | Actions are taken on rows owned by you. | ### Advanced options You can set additional properties to define more granularly when the flow runs and the user profile under which it runs. ## Filter conditions Use filter conditions to set conditions for when to trigger flows. ![Filter condition.](../media/create-update-delete-trigger/filter-conditions.png "Filter condition") ## Filtering columns Use the **Column filter** box to define the specific columns of the row that should cause the flow to run when changed, as a comma-separated list of unique column names. ![Filter columns by firstname.lastname.](../media/create-update-delete-trigger/filter-columns.png "Filter columns by firstname.lastname") >[!NOTE] >This property applies to the **Update** condition only. **Create** and **Delete** apply to all columns of a row. ### Filter expression The filter expression provides a way for you to define an OData style filter expression to help you to define the trigger conditions even more precisely. The flow runs only when the expression evaluates to *true* after the change is saved in Dataverse. In the following example, the flow triggers when `firstname` is updated to "John". See the following examples, [standard filter operators](/powerapps/developer/common-data-service/webapi/query-data-web-api#standard-filter-operators), and [query functions](/powerapps/developer/common-data-service/webapi/query-data-web-api#standard-query-functions) to learn how to construct these filter expressions. >[!NOTE] >Unlike the examples in the reference links, your expression must not contain the string **$filter=**. This string applies only when you use the APIs directly. ![Row filter equal.](../media/create-update-delete-trigger/row-filter.png) ![Row filter contains.](../media/create-update-delete-trigger/row-filter-contains.png) ### Wait condition using delay until Use an OData-style time stamp in the **Delay until** property to delay the flow trigger until a specific UTC time. The key benefit of using the Dataverse **Delay until** property instead of the standard **Delay until** *action* is the Dataverse **Delay until** property never expires, allowing the flow run to wait for long periods of time. ![Delay until.](../media/create-update-delete-trigger/delay-until.png "Delay until") ### User impersonation using Run As >[!IMPORTANT] >The flow owner must have the Microsoft Dataverse privilege **Act on Behalf of Another User** (prvActOnBehalfOfAnotherUser). The **Delegate** security role includes this privilege by default. You can enable it on any security role. For more details, go to [Impersonate another user](/powerapps/developer/common-data-service/impersonate-another-user). When you create flows with the **When a row is added, modified or deleted** trigger, you can set each Microsoft Dataverse action in the flow to be performed using the context of a user, other than the flow owner. Follow these steps to impersonate a user: 1. In the Power Automate flow definition, select **Show advanced options** in the **When a row is added, modified or deleted** trigger. 1. Select a value for **Run as** to tell Microsoft Dataverse which user’s context you intend to use for subsequent Dataverse actions. 1. For each Dataverse action that you want to run as a different user, select the menu in the upper-right corner (...), as shown in the following image, and select the **Use invoker’s connection** setting. For the steps in which it is not selected, the default user is assumed. This would call the underlying APIs as per the selected user, and not as the flow owner. ![Run as the modifying user.](../media/create-update-delete-trigger/run-as.png "Run as the modifying user") If nothing is specified, it defaults to the flow owner who created the flow&mdash;essentially, the author. Here are the other options: - **Flow owner**: The user who created the flow. - **Row owner**: The user who owns the Microsoft Dataverse row that underwent a change, causing the flow to be triggered. If a row is owned by a team, then this option falls back to run as the flow owner. - **Modifying user**: The user that took the action on the Microsoft Dataverse row, causing the flow to get triggered or modified. ![Run as options.](../media/create-update-delete-trigger/11.png "Run as options") Additionally, instant flows allow running the steps of any other [connector](/connectors/) (such as [Microsoft Teams](/connectors/teams/), [Microsoft 365 Outlook](/connectors/office365/), or [SharePoint](/connectors/sharepointonline/) in the same flow using the invoker’s connection. To do so, follow these steps: 1. Go to the flow overview page. 1. Select **Edit** on the **Run only users** settings. 1. In the **Manage run-only permissions** pane, go to the **User and groups** tab, and then select **Provided by run-only user** under the **Connections Used** list.
52.573333
366
0.751458
eng_Latn
0.995684
87baa1a96e5eeb4eabba21fa5943cbdc0625ff37
2,220
md
Markdown
docs/framework/winforms/controls/printpreviewcontrol-control-windows-forms.md
AlejandraHM/docs.es-es
5f5b056e12f9a0bcccbbbef5e183657d898b9324
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/controls/printpreviewcontrol-control-windows-forms.md
AlejandraHM/docs.es-es
5f5b056e12f9a0bcccbbbef5e183657d898b9324
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/controls/printpreviewcontrol-control-windows-forms.md
AlejandraHM/docs.es-es
5f5b056e12f9a0bcccbbbef5e183657d898b9324
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: PrintPreviewControl (Control, formularios Windows Forms) ms.date: 03/30/2017 helpviewer_keywords: - printing [Windows Forms], print preview - PrintPreviewControl control (using designer) - PrintPreview control (using designer) - print preview [Windows Forms], custom interface (using designer) ms.assetid: 3fdb2e46-92a3-4e26-bb8d-63a89087b337 ms.openlocfilehash: 7f4b6f71427a750799d102a8602d3a2c1d43d034 ms.sourcegitcommit: 3d5d33f384eeba41b2dff79d096f47ccc8d8f03d ms.translationtype: MT ms.contentlocale: es-ES ms.lasthandoff: 05/04/2018 ms.locfileid: "33534318" --- # <a name="printpreviewcontrol-control-windows-forms"></a>PrintPreviewControl (Control, formularios Windows Forms) Los Windows Forms `PrintPreviewControl` se utiliza para mostrar un documento tal como aparecerá cuando se imprima. Este control no tiene botones ni otros elementos de la interfaz de usuario, por lo que normalmente utiliza `PrintPreviewControl` solo si desea volver a escribir su propia interfaz de usuario de vista previa de impresión. Si quiere la interfaz de usuario estándar, utilice un control <xref:System.Windows.Forms.PrintPreviewDialog>. ## <a name="in-this-section"></a>En esta sección [Información general sobre el control PrintPreviewControl](../../../../docs/framework/winforms/controls/printpreviewcontrol-control-overview-windows-forms.md) Presenta los conceptos generales de `PrintPreviewControl`, que puede utilizar para diseñar su propio componente o cuadro de diálogo de vista previa de impresión. ## <a name="reference"></a>Referencia <xref:System.Windows.Forms.PrintPreviewControl> Contiene información de referencia sobre la clase y sus miembros. ## <a name="related-sections"></a>Secciones relacionadas [PrintPreviewDialog (control)](../../../../docs/framework/winforms/controls/printpreviewdialog-control-windows-forms.md) Describe una alternativa para crear funcionalidades de vista previa de impresión. [Controles que se utilizan en formularios Windows Forms](../../../../docs/framework/winforms/controls/controls-to-use-on-windows-forms.md) Proporciona una lista completa de controles de Windows Forms, con vínculos a información sobre su uso.
65.294118
447
0.787387
spa_Latn
0.719744
87bac6540dce7862586a83161607822de4b19f06
463
md
Markdown
README.md
karun012/scotty-todo-sample
7933750fd23b0064f69295de4be0c74aa1fd509e
[ "Unlicense" ]
2
2017-01-07T06:23:53.000Z
2019-05-14T11:47:48.000Z
README.md
karun012/scotty-todo-sample
7933750fd23b0064f69295de4be0c74aa1fd509e
[ "Unlicense" ]
null
null
null
README.md
karun012/scotty-todo-sample
7933750fd23b0064f69295de4be0c74aa1fd509e
[ "Unlicense" ]
null
null
null
scotty-todo-sample ================== Sample Todo App Using Scotty and ReactJS #How to run it I recommend using a sandbox so that the dependencies do not mess up your environment. ```cabal cabal update cabal sandbox init cabal install .cabal-sandbox/bin/scotty-todo ``` If the install fails with a "Backjump limit reached" error, use ```cabal cabal install --max-backjumps=9999 ``` Access [http://localhost:3000](http://localhost:3000) to get to the index page
25.722222
86
0.736501
eng_Latn
0.772815
87bb9aed05839f971fa2fd3f415dec37c4b3818c
3,010
md
Markdown
drafts/web/2017-12-17-book-JS-Object-Oriented.md
ngminhtrung/travis-blog
968b013f5ea28d0b74223d4acd12667b4750b933
[ "MIT" ]
1
2022-01-20T11:49:38.000Z
2022-01-20T11:49:38.000Z
drafts/web/2017-12-17-book-JS-Object-Oriented.md
ngminhtrung/travis-blog
968b013f5ea28d0b74223d4acd12667b4750b933
[ "MIT" ]
15
2018-04-27T23:48:18.000Z
2022-02-26T02:25:42.000Z
drafts/web/2017-12-17-book-JS-Object-Oriented.md
ngminhtrung/travis-blog
968b013f5ea28d0b74223d4acd12667b4750b933
[ "MIT" ]
1
2020-11-05T04:58:47.000Z
2020-11-05T04:58:47.000Z
## Về tác giả Nicholas C. Zakas: - Từng là software engineer tại Yahoo! (vị trí principal front-end engineer), Box - Tác giả viết sách và diễn giả về chủ đề JavaScript - Sách đã viết: - Maintainable JavaScript (O’Reilly Media, 2012) - Professional JavaScript for Web Developers (Wrox, 2012) - Tác giả của: ESlint ## Dịch tóm tắt sách ### Chương 1 - Primitive và Reference Types - Mặc dù JavaScript (tính đến thời điểm xuất bản sách năm 2014) không có khái niệm về "classes", nó vẫn có hai loại types là **primitive** và **reference**. - Mỗi variable hoặc data được gắn với 1 trong 2 loại type nói trên. - *Primitive types* dùng để lưu các dữ liệu đơn giản - *Reference types* dùng để lưu các objects, vốn đơn thuần chỉ là những tham chiếu đến các vị trí khác nhau trong bộ nhớ. - Điều hay ho chính là JavaScript cho phép người dùng làm việc với primitive type giống như với reference type để giúp việc viết code trở nên nhất quán. - Với các ngôn ngữ lập trình khác, primitive type được lưu ở trên *stack*, còn refernce type được lưu ở trong *heap*, thì JavaScript lại xử lý vấn đề theo một cách khác: Nó lưu thông tin về *scope* của variables với 1 object đặc biệt gọi là `variable object`. - Primitive values được lưu trực tiếp trên `variable object`. - Reference values được đặt như một con trở (pointer) trong `variable object`, đóng vai trò là tham chiếu đến vị trí trong bộ nhớ nơi mà object được lưu. - Có 05 primitive types, bao gồm: `strings`, `numbers`, `Booleans`, `null`, và `undefined`. - Người dùng có thể dùng `typeof` để kiểm tra primitive type của data, ngoại trừ `null` - Riêng với `null`, để kiểm tra, cần so sánh trực tiếp variable/ data với `null`. - Reference type là thứ gần nhất với "class" trong JavaScript, mỗi object là 1 instance của reference type. - Mỗi object có thể được tạo thông qua toán tử `new` hoặc sử dụng `reference literal` (chính là dấu [], {}, hoặc ()) - Để truy cập properties hoặc methods, lập trình viên có thể dùng *dot notation" (chính là dấu "."), hoặc *bracket notation* (chính là [ ]) - Function là objects trong JavaScript, có thể được xác định thông qua toán tử `typeof`. ### Chương 2 - Functions - Function trong JavaScript độc đáo ở chỗ nó cũng là những objects, nghĩa là có thể truy cập, copy, ghi đè, nói chung là có thể xử lý như xử lý value của bất kỳ object nào. - Điểm khác biệt lớn nhất giữa *function* và các *objects* khác ở chỗ *function* có một property đặc biệt, [[Call]], có chứa mô tả để thực thi function. - Toán tử `typeof` tìm kiểm property đặc biệt trên để quyết định xem đây có phải là function hay không. - Về mặt khai báo, có hai cách tạo function: *declaration* và *expression*. - *Function declaration* có chứa tên function ở bên phải của từ khóa `function`, và được *hoisted* lên trên cùng của context mà function đó được định nghĩa. - *Function expression* được dùng để gán giá trị vào expression, function parameters, hoặc để trả về giá trị của một function khác.
75.25
260
0.72392
vie_Latn
1.00001
87bba025b0693789ac581db0e63cc1dc7f411b61
19,429
md
Markdown
articles/key-vault/key-vault-get-started.md
SunnyDeng/azure-content-dede
edb0ac8eec176b64971ec219274a4a922dd00fec
[ "CC-BY-3.0" ]
2
2020-08-29T21:10:59.000Z
2021-07-25T10:13:02.000Z
articles/key-vault/key-vault-get-started.md
SunnyDeng/azure-content-dede
edb0ac8eec176b64971ec219274a4a922dd00fec
[ "CC-BY-3.0" ]
null
null
null
articles/key-vault/key-vault-get-started.md
SunnyDeng/azure-content-dede
edb0ac8eec176b64971ec219274a4a922dd00fec
[ "CC-BY-3.0" ]
null
null
null
<properties pageTitle="Erste Schritte mit dem Azure-Schlüsseltresor | Microsoft Azure" description="Verwenden Sie dieses Tutorials für den Einstieg in den Azure-Schlüsseltresor, um einen geschützten Container in Azure zu erstellen, in dem Sie kryptografischen Schlüssel und geheime Schlüssel in Azure speichern und verwalten." services="key-vault" documentationCenter="" authors="cabailey" manager="mbaldwin" tags="azure-resource-manager"/> <tags ms.service="key-vault" ms.workload="identity" ms.tgt_pltfrm="na" ms.devlang="na" ms.topic="hero-article" ms.date="11/10/2015" ms.author="cabailey"/> # Erste Schritte mit dem Azure-Schlüsseltresor # Azure-Tresorschlüssel ist in den meisten Regionen verfügbar. Weitere Informationen finden Sie auf der Seite [Preisübersicht für Schlüsseltresor](../../../../pricing/details/key-vault/). ## Einführung Verwenden Sie dieses Tutorial für den Einstieg in Azure-Schlüsseltresor, um einen geschützten Container (einen Tresor) in Azure zu erstellen, in dem Sie kryptografische und geheime Schlüssel in Azure speichern und verwalten. Sie werden durch den Vorgang der Verwendung von Azure PowerShell zum Erstellen eines Tresors geleitet, der einen Schlüssel oder ein Kennwort enthält, den/das Sie anschließend mit einer Azure-Anwendung verwenden können. Anschließend wird gezeigt, wie eine Anwendung diesen Schlüssel bzw. das Kennwort verwenden kann. *Geschätzter Zeitaufwand*:* 20 Minuten. >[AZURE.NOTE]Dieses Tutorial enthält keine Anweisungen zum Schreiben der in einem der Schritte verwendeten Azure-Anwendung, die dazu dient, eine Anwendung zum Verwenden eines Schlüssels oder geheimen Schlüssels im Schlüsseltresor zu autorisieren. > >Derzeit können Sie den Azure-Schlüsseltresor nicht im Azure-Portal konfigurieren. Sie müssen stattdessen die Anweisungen für Azure PowerShell verwenden. Anleitungen für die plattformübergreifende Befehlszeilenschnittstelle finden Sie in [diesem entsprechenden Tutorial](key-vault-manage-with-cli.md). Eine Übersicht über den Azure-Schlüsseltresor finden Sie unter [Was ist der Azure-Schlüsseltresor?](key-vault-whatis.md) ## Voraussetzungen Für dieses Tutorial benötigen Sie Folgendes: - Ein Abonnement für Microsoft Azure. Wenn Sie kein Abonnement haben, können Sie sich für eine [kostenlose Testversion](../../../../pricing/free-trial) registrieren. - Azure PowerShell, **mindestens Version 1.0**. Um Azure PowerShell zu installieren und Ihrem Azure-Abonnement zuzuordnen, lesen Sie [Installieren und Konfigurieren von Azure PowerShell](../powershell-install-configure.md). Wenn Sie Azure PowerShell bereits installiert haben und die Version nicht kennen, geben Sie über die Azure PowerShell-Konsole `(Get-Module azure -ListAvailable).Version` ein. Wenn Sie Azure PowerShell in den Versionen 0.9.1 bis 0.9.8 installiert haben, können Sie diese Tutorial mit einigen kleineren Änderungen verwenden. Sie müssen z. B. den Befehl `Switch-AzureMode AzureResourceManager` verwenden, und einige der Befehle des Azure-Schlüsseltresors wurden geändert. Eine Liste der Schlüsseltresor-Cmdlets für die Versionen 0.9.1 bis 0.9.8 finden Sie unter [Cmdlets für den Azure-Schlüsseltresor](https://msdn.microsoft.com/library/azure/dn868052(v=azure.98).aspx). - Eine Anwendung, die zur Verwendung des Schlüssels oder Kennworts konfiguriert wird, den bzw. das Sie in diesem Lernprogramm erstellen. Eine Beispielanwendung erhalten Sie im [Microsoft Download Center](http://www.microsoft.com/de-DE/download/details.aspx?id=45343). Anweisungen finden Sie in der zugehörigen Readme-Datei. Dieses Tutorial richtet sich an Azure PowerShell-Anfänger. Es wird aber vorausgesetzt, dass Sie die grundlegenden Konzepte kennen, z. B. Module, Cmdlets und Sitzungen. Weitere Informationen finden Sie unter [Erste Schritte mit Windows PowerShell](https://technet.microsoft.com/library/hh857337.aspx). Um detaillierte Hilfe zu einem Cmdlet aus dem Tutorial zu erhalten, verwenden Sie das **Get-Help**-Cmdlet. Get-Help <cmdlet-name> -Detailed Geben Sie beispielsweise Folgendes ein, um Hilfe zum **Add-AzureAccount**-Cmdlet zu erhalten: Get-Help Add-AzureAccount -Detailed Lesen Sie bitte auch die folgenden Tutorials, um sich mit dem Azure-Ressourcen-Manager in Azure PowerShell vertraut zu machen: - [Installieren und Konfigurieren von Azure PowerShell](../powershell-install-configure.md) - [Verwenden von Azure PowerShell mit dem Ressourcen-Manager](../powershell-azure-resource-manager.md) ## <a id="connect"></a>Verbindungsherstellung mit Ihren Abonnements ## Starten Sie eine Azure PowerShell-Sitzung, und melden Sie sich mit dem folgenden Befehl bei Ihrem Azure-Konto an: Login-AzureRmAccount Geben Sie im Popup-Browserfenster den Benutzernamen und das Kennwort Ihres Azure-Kontos ein. Azure PowerShell ruft alle Abonnements ab, die diesem Konto zugeordnet sind, und verwendet standardmäßig das erste Abonnement. Wenn Sie über mehrere Abonnements verfügen und zur Verwendung für den Azure-Schlüsseltresor ein spezifisches Abonnement verwenden möchten, geben Sie den folgenden Befehl ein, um die Abonnements für Ihr Konto anzuzeigen: Get-AzureRmSubscription Geben Sie Folgendes ein, um das zu verwendende Abonnement anzugeben: Set-AzureRmContext -SubscriptionId <subscription ID> Weitere Informationen zum Konfigurieren von Azure PowerShell finden Sie unter [Installieren und Konfigurieren von Azure PowerShell](../powershell-install-configure.md). ## <a id="resource"></a>Erstellen einer neuen Ressourcengruppe ## Wenn Sie den Azure-Ressourcen-Manager verwenden, werden alle zugehörigen Ressourcen in einer Ressourcengruppe erstellt. Im Rahmen dieses Tutorials erstellen wir eine neue Ressourcengruppe mit dem Namen **ContosoResourceGroup**: New-AzureRmResourceGroup –Name 'ContosoResourceGroup' –Location 'East Asia' ## <a id="vault"></a>Erstellen eines Schlüsseltresors ## Verwenden Sie das Cmdlet [New-AzureRmKeyVault](https://msdn.microsoft.com/library/azure/mt603736.aspx), um einen Schlüsseltresor zu erstellen. Dieses Cmdlet verfügt über drei erforderliche Parameter: einen für den **Ressourcengruppennamen**, einen für den **Schlüsseltresornamen** und einen für den **geografischen Standort**. Wenn Sie beispielsweise den Tresornamen **ContosoKeyVault**, den Ressourcengruppennamen **ContosoResourceGroup** und den Speicherort **East Asia** verwenden möchten, geben Sie Folgendes ein: New-AzureRmKeyVault -VaultName 'ContosoKeyVault' -ResourceGroupName 'ContosoResourceGroup' -Location 'East Asia' Die Ausgabe dieses Cmdlets zeigt die Eigenschaften des Schlüsseltresors, den Sie soeben erstellt haben. Die zwei wichtigsten Eigenschaften sind diese: - **Tresorname**: In diesem Beispiel ist dies **ContosoKeyVault**. Sie verwenden diesen Namen für andere Schlüsseltresor-Cmdlets. - **Tresor-URI**: In diesem Beispiel ist dies https://contosokeyvault.vault.azure.net/. Anwendungen, die Ihren Tresor über die zugehörige REST-API nutzen, müssen diesen URI verwenden. Ihr Azure-Konto ist jetzt autorisiert, Vorgänge in diesem Schlüsseltresor durchzuführen. Bisher sind Sie der einzige Benutzer mit dieser Berechtigung. ## <a id="add"></a>Hinzufügen eines Schlüssels oder geheimen Schlüssels zum Schlüsseltresor ## Wenn Sie mit dem Azure-Schlüsseltresor einen softwaregeschützten Schlüssel erstellen möchten, verwenden Sie hierzu das Cmdlet [Add-AzureKeyVaultKey](https://msdn.microsoft.com/library/azure/dn868048.aspx). Geben Sie Folgendes ein: $key = Add-AzureKeyVaultKey -VaultName 'ContosoKeyVault' -Name 'ContosoFirstKey' -Destination 'Software' Wenn Sie über einen vorhandenen softwaregeschützten Schlüssel in einer PFX-Datei auf Ihrem Laufwerk "C:\" in einer Datei namens "softkey.pfx" verfügen, die Sie in den Azure-Schlüsseltresor hochladen möchten, geben Sie den folgenden Befehl ein. Mit diesem Befehl wird die Variable **securepfxpwd** gesetzt, um das Kennwort **123** für die PFX-Datei festzulegen: $securepfxpwd = ConvertTo-SecureString –String '123' –AsPlainText –Force Geben Sie anschließend den folgenden Code ein, um den Schlüssel aus der PFX-Datei zu importieren, die den Schlüssel mithilfe von Software im Schlüsseltresordienst schützt: $key = Add-AzureKeyVaultKey -VaultName 'ContosoKeyVault' -Name 'ContosoFirstKey' -KeyFilePath 'c:\softkey.pfx' -KeyFilePassword $securepfxpwd Jetzt können Sie mit dem zugehörigen URI auf den erstellten oder in den Azure-Schlüsseltresor hochgeladenen Schlüssel verweisen. Verwenden Sie ****https://ContosoKeyVault.vault.azure.net/keys/ContosoFirstKey**, um immer die aktuelle Version zu erhalten, und verwenden Sie ****https://ContosoKeyVault.vault.azure.net/keys/ContosoFirstKey/cgacf4f763ar42ffb0a1gca546aygd87**, um diese bestimmte Version abzurufen. Geben Sie zur Anzeige des URI für diesen Schlüssel Folgendes ein: $Key.key.kid Um einen geheimen Schlüssel zum Schlüsseltresor hinzuzufügen – in diesem Fall das Kennwort "SQLPassword" mit dem Wert "Pa$$w0rd" für den Azure-Schlüsseltresor – konvertieren Sie zunächst den Wert "Pa$$w0rd" in eine sichere Zeichenfolge, indem Sie Folgendes eingeben: $secretvalue = ConvertTo-SecureString 'Pa$$w0rd' -AsPlainText -Force Geben Sie anschließend Folgendes ein: $secret = Set-AzureKeyVaultSecret -VaultName 'ContosoKeyVault' -Name 'SQLPassword' -SecretValue $secretvalue Jetzt können Sie mit dem zugehörigen URI auf das Kennwort verweisen, das Sie dem Azure-Schlüsseltresor hinzugefügt haben. Verwenden Sie ****https://ContosoVault.vault.azure.net/secrets/SQLPassword**, um immer die aktuelle Version zu erhalten, und verwenden Sie ****https://ContosoVault.vault.azure.net/secrets/SQLPassword/90018dbb96a84117a0d2847ef8e7189d**, um diese bestimmte Version abzurufen. Geben Sie zur Anzeige des URI für diesen geheimen Schlüssel Folgendes ein: $secret.Id Zeigen wir den Schlüssel oder geheimen Schlüssel an, den Sie soeben erstellt haben: - Geben Sie Folgendes ein, um Ihren Schlüssel anzuzeigen: `Get-AzureKeyVaultKey –VaultName 'ContosoKeyVault'` - Geben Sie Folgendes ein, um Ihren geheimen Schlüssel anzuzeigen: `Get-AzureKeyVaultSecret –VaultName 'ContosoKeyVault'` Jetzt sind Schlüsseltresor und Schlüssel oder geheimer Schlüssel bereit für die Verwendung durch die Anwendung. Sie müssen die Anwendungen zur Verwendung der Schlüssel bzw. geheimen Schlüssel autorisieren. ## <a id="register"></a>Registrieren einer Anwendung mit Azure Active Directory ## Dieser Schritt wird üblicherweise durch einen Entwickler auf einem separaten Computer durchgeführt. Dieser Schritt ist nicht spezifisch für den Azure-Schlüsseltresor, wird der Vollständigkeit halber jedoch hier aufgeführt. >[AZURE.IMPORTANT]Zum Abschließen dieses Tutorials müssen sich Ihr Konto, der Schlüsseltresor und die in diesem Schritt registrierte Anwendung im selben Azure-Verzeichnis befinden. Anwendungen, die einen Schlüsseltresor verwenden, müssen sich mithilfe eines Azure Active Directory-Tokens authentifizieren. Hierzu muss der Besitzer der Anwendung die Anwendung zunächst in Azure Active Directory registrieren. Zum Abschluss der Registrierung erhält der Anwendungsbesitzer die folgenden Werte: - Eine **Anwendungs-ID** (auch Client-ID genannt) und einen **Authentifizierungsschlüssel** (auch als gemeinsamer geheimer Schlüssel bezeichnet). Die Anwendung muss beide dieser Werte in Azure Active Directory vorlegen, um ein Token zu erhalten. Wie die Anwendung konfiguriert wird, um dies zu erreichen, richtet sich nach der Anwendung. Bei der Beispielanwendung für den Schlüsseltresor legt der Anwendungsbesitzer diese Werte in der Datei "app.config" fest. So registrieren Sie die Anwendungen in Azure Active Directory 1. Melden Sie sich beim Azure-Portal an. 2. Klicken Sie auf der linken Seite auf **Active Directory**, und wählen Sie das Verzeichnis, in dem Sie Ihre Anwendung registrieren möchten. <br> <br> **Hinweis**: Sie müssen dasselbe Verzeichnis auswählen, das auch das Azure-Abonnement enthält, mit dem Sie Ihren Schlüsseltresor erstellt haben. Wenn Sie nicht wissen, welches Verzeichnis dies ist, klicken Sie auf **Einstellungen**, ermitteln Sie das Abonnement, mit dem Sie Ihren Schlüsseltresor erstellt haben, und notieren Sie sich den Namen des Verzeichnisses, der in der letzten Spalte angezeigt wird. 3. Klicken Sie auf **ANWENDUNGEN**. Wenn Ihrem Verzeichnis keine Apps hinzugefügt wurden, wird auf dieser Seite der Link **App hinzufügen** angezeigt. Klicken Sie auf den Link, oder klicken Sie alternativ auf der Befehlsleiste auf **HINZUFÜGEN**. 4. Klicken Sie im Assistenten **ANWENDUNG HINZUFÜGEN** auf der Seite **Was möchten Sie tun?** auf **Eine von meinem Unternehmen entwickelte Anwendung hinzufügen**. 5. Geben Sie auf der Seite **Erzählen Sie uns von Ihrer App** einen Namen für Ihre Anwendung ein, und wählen Sie dann **WEBANWENDUNG UND/ODER WEB-API** (die Standardeinstellung) aus. Klicken Sie auf das Symbol **Weiter**. 6. Geben Sie auf der Seite **App-Eigenschaften** die **ANMELDE-URL** und den **APP-ID-URI** für Ihre Webanwendung an. Wenn Ihre Anwendung über keine solchen Werte verfügt, können Sie in diesem Schritt Pseudowerte verwenden (Sie können beispielsweise in beide Felder den Wert http://test1.contoso.com eingeben). Es spielt keine Rolle, ob diese Websites vorhanden sind. Es ist nur wichtig, dass der App-ID-URI für jede Anwendung in Ihrem Verzeichnis eindeutig ist. Das Verzeichnis verwendet diese Zeichenfolge zur Identifizierung Ihrer App. 7. Klicken Sie auf das Symbol **Abschließen**, um Ihre Änderungen im Assistenten zu speichern. 8. Klicken Sie auf der Seite **Schnellstart** auf **KONFIGURIEREN**. 9. Führen Sie einen Bildlauf zum Abschnitt **Schlüssel** durch, wählen Sie die Dauer aus, und klicken Sie dann auf **SPEICHERN**. Die Seite wird aktualisiert und zeigt jetzt einen Schlüsselwert. Sie müssen Ihre Anwendung mit diesem Schlüsselwert und der **CLIENT-ID** konfigurieren. (Die Anweisungen für diese Konfigurationen sind anwendungsspezifisch.) 10. Kopieren Sie den Client-ID-Wert von dieser Seite. Sie verwenden ihn im nächsten Schritt, um Berechtigungen für Ihren Tresor festzulegen. ## <a id="authorize"></a>Autorisieren der Anwendung zum Verwenden des Schlüssels oder geheimen Schlüssels ## Verwenden Sie das Cmdlet [Set-AzureRmKeyVaultAccessPolicy](https://msdn.microsoft.com/library/azure/mt603625.aspx), um die Anwendung zum Zugreifen auf den Schlüssel oder geheimen Schlüssel im Tresor zu autorisieren. Wenn Ihr Tresorname beispielsweise **ContosoKeyVault** lautet, die Anwendung, die Sie autorisieren möchten, über die Client-ID 8f8c4bbd-485b-45fd-98f7-ec6300b7b4ed verfügt und Sie die Anwendung zum Entschlüsseln und Anmelden mit Schlüsseln in Ihrem Tresor autorisieren möchten, führen Sie Folgendes aus: Set-AzureRmKeyVaultAccessPolicy -VaultName 'ContosoKeyVault' -ServicePrincipalName 8f8c4bbd-485b-45fd-98f7-ec6300b7b4ed -PermissionsToKeys decrypt,sign Wenn Sie dieselbe Anwendung so autorisieren möchten, dass sie geheime Schlüssel im Tresor liest, führen Sie Folgendes aus: Set-AzureRmKeyVaultAccessPolicy -VaultName 'ContosoKeyVault' -ServicePrincipalName 8f8c4bbd-485b-45fd-98f7-ec6300b7b4ed -PermissionsToSecrets Get ## <a id="HSM"></a>Verwenden eines Hardwaresicherheitsmoduls (HSM) ## Zur Steigerung der Sicherheit können Sie Schlüssel in HSMs importieren oder in diesen generieren. Diese Schlüssel verbleiben immer innerhalb der HSM-Grenzen. Die HSMs sind FIPS 140-2 Ebene 2 überprüft. Wenn diese Anforderung auf Sie nicht zutrifft, überspringen Sie diesen Abschnitt und wechseln Sie zu [Löschen des Schlüsseltresors und zugeordneter Schlüssel und geheimer Schlüssel](#delete). Um diese HSM-geschützten Schlüssel zu erstellen, müssen Sie über ein [Tresorabonnement mit Unterstützung von HSM-geschützten Schlüsseln](../../../pricing/free-trial) verfügen. Wenn Sie den Tresor erstellen, fügen Sie den **-SKU**-Parameter hinzu. New-AzureRmKeyVault -VaultName 'ContosoKeyVaultHSM' -ResourceGroupName 'ContosoResourceGroup' -Location 'East Asia' -SKU 'Premium' Sie können diesem Tresor softwaregeschützte Schlüssel (wie weiter oben gezeigt) und HSM-geschützte Schlüssel hinzufügen. Legen Sie den Parameter **-Destination** auf "HSM" fest, um einen HSM-geschützten Schlüssel zu erstellen: $key = Add-AzureKeyVaultKey -VaultName 'ContosoKeyVaultHSM' -Name 'ContosoFirstHSMKey' -Destination 'HSM' Sie können mit dem folgenden Befehl einen Schlüssel aus einer PFX-Datei auf Ihrem Computer importieren. Dieser Befehl importiert den Schlüssel in HSMs im Schlüsseltresordienst: $key = Add-AzureKeyVaultKey -VaultName 'ContosoKeyVaultHSM' -Name 'ContosoFirstHSMKey' -KeyFilePath 'c:\softkey.pfx' -KeyFilePassword $securepfxpwd -Destination 'HSM' Der nächste Befehl importiert ein BYOK-Paket (Bring Your Own Key). Auf diese Weise können Sie Ihren Schlüssel im lokalen HSM generieren und ihn in HSMs im Schlüsseltresordienst übertragen, ohne dass der Schlüssel die HSM-Grenzen verlässt: $key = Add-AzureKeyVaultKey -VaultName 'ContosoKeyVaultHSM' -Name 'ContosoFirstHSMKey' -KeyFilePath 'c:\ITByok.byok' -Destination 'HSM' Ausführlichere Informationen zum Generieren dieses BYOK-Pakets finden Sie unter [Generieren und Übertragen von HSM-geschützten Schlüsseln für den Azure-Schlüsseltresor](key-vault-hsm-protected-keys.md). ## <a id="delete"></a>Löschen des Schlüsseltresors und der zugeordneten Schlüssel und geheimen Schlüssel ## Wenn Sie den Schlüsseltresor und die darin enthaltenen Schlüssel und geheimen Schlüssel nicht länger benötigen, können Sie den Schlüsseltresor mit dem Cmdlet [Remove-AzureRmKeyVault](https://msdn.microsoft.com/library/azure/mt619485.aspx) löschen: Remove-AzureRmKeyVault -VaultName 'ContosoKeyVault' Alternativ können Sie eine gesamte Azure-Ressourcengruppe löschen, die den Schlüsseltresor und alle weiteren Ressourcen enthält, die Sie in diese Gruppe eingeschlossen haben: Remove-AzureRmResourceGroup -ResourceGroupName 'ContosoResourceGroup' ## <a id="other"></a>Weitere Azure PowerShell-Cmdlets ## Die folgenden weiteren Befehle sind möglicherweise ebenfalls für das Verwalten eines Azure-Schlüsseltresors von Nutzen: - `$Keys = Get-AzureKeyVaultKey -VaultName 'ContosoKeyVault'`: Dieser Befehl ruft eine tabellarische Anzeige aller Schlüssel und ausgewählten Eigenschaften ab. - `$Keys[0]`: Dieser Befehl zeigt eine vollständige Liste der Eigenschaften für den angegebenen Schlüssel an. - `Get-AzureKeyVaultSecret`: Dieser Befehl ruft eine tabellarische Anzeige der Namen aller geheimen Schlüssel und ausgewählten Eigenschaften ab. - `Remove-AzureKeyVaultKey -VaultName 'ContosoKeyVault' -Name 'ContosoFirstKey'`: Beispiel dazu, wie ein bestimmter Schlüssel entfernt wird. - `Remove-AzureKeyVaultSecret -VaultName 'ContosoKeyVault' -Name 'SQLPassword'`: Beispiel dazu, wie ein bestimmter geheimer Schlüssel entfernt wird. ## <a id="next"></a>Nächste Schritte ## Ein weiterführendes Tutorial zur Verwendung des Azure-Schlüsseltresors in einer Webanwendung finden Sie unter [Verwenden des Azure-Schlüsseltresors aus einer Webanwendung](key-vault-use-from-web-application.md). Eine Liste der Azure PowerShell 1.0-Cmdlets für den Azure-Schlüsseltresor finden Sie unter [Cmdlets für den Azure-Schlüsseltresor](https://msdn.microsoft.com/library/azure/dn868052.aspx). Eine Referenz zur Programmierung finden Sie im [Entwicklerhandbuch für den Azure-Schlüsseltresor](key-vault-developers-guide.md). <!---HONumber=Nov15_HO3-->
80.618257
892
0.808431
deu_Latn
0.994915
87bbd7dffd2a29e0e9c8b32af0c4b76ae12c0c14
7,798
md
Markdown
content/en/angular/Dealing-with-Angular-security-issues/index.md
manasa-arun/angularjswiki
d6aebc09e3132f66b2bbbe274ee288e248e0d780
[ "MIT" ]
null
null
null
content/en/angular/Dealing-with-Angular-security-issues/index.md
manasa-arun/angularjswiki
d6aebc09e3132f66b2bbbe274ee288e248e0d780
[ "MIT" ]
null
null
null
content/en/angular/Dealing-with-Angular-security-issues/index.md
manasa-arun/angularjswiki
d6aebc09e3132f66b2bbbe274ee288e248e0d780
[ "MIT" ]
null
null
null
+++ title = "Dealing With the Top 5 Most Known Security Risks in Angular" subtitle = "Security issues in Angular" summary ="As with any other modern software development instrument, security is the number one concern for early Angular application development." keywords=["Security issues, Angular"] date="2021-11-18T00:00:05+0000" lastmod="2021-11-18T00:01:00+0000" type="post" draft=false authors = ["admin"] [image] caption = "Security issues in Angular" # Focal point (optional) # Options: Smart, Center, TopLeft, Top, TopRight, Left, Right, BottomLeft, Bottom, BottomRight focal_point = "" # Show image only in page previews? preview_only = false +++ As with any other modern software development instrument, security is the number one concern for early Angular application development. Angular- a front-end component-based framework- undoubtedly helps mitigate a range of security threats without the need for additional lines of code. However, even with its magnificent protective shield coupled with Long-term Support (LTS) from [Google](https://www.angularjswiki.com/angular/angular-13-release/), applications built on Angular are still prone to attacks. Identifying vulnerabilities is the first step in threat actors’ playbooks. When an unpatched weakness is found, they exploit it to gain access to the application and launch an attack. Snyk has done a great job explaining [angular security best practices](https://snyk.io/blog/angular-security-best-practices/) that help prevent attacks on applications running on Angular. This article will examine five of the most known security risks that are a concern with Angular applications. ## Cross-Site Scripting (XSS) Attacks Cross-site scripting is a common application-layer web attack in which a hacker injects malicious scripts into a web application. The web page picks up the malicious data and interprets it as code. Upon execution, an XSS vulnerability gives the malicious actor complete control over the app. When a website with an XSS vulnerability executes a malicious code, it sends your cookie information to the attacker behind the scenes. This means that neither you nor the website will be aware that an XSS attack is going on. In most cases, attackers send the code as a combination of XSS and HTML. But XSS can also be in the form of executable plugins, downloads and media content. There are three types of XSS attacks: ### Persistent XSS Persistent XSS also known as stored XSS, is when the attacker injects a malicious code directly into a vulnerable web application. ### Non-persistent XSS (AKA Type II or reflected XSS) This attack happens when a malicious script is reflected off a web application onto your user’s browser. ### DOM-based XSS DOM-based XSS also called Type-0 XSS, is an attack wherein the attacker alters the Document Object Model (DOM) in the victim’s browser. Consequently, the attacker can run code in the target browser without the user’s knowledge. Although no change is done on the HTTP response itself, the damaging script can execute differently due to the change in the DOM environment. ## Clickjacking Attacks Clickjacking is when an attacker manipulates a user to click a button or link, thereby performing an action on another page when they intended to accomplish something different. Thus, the attacker is hijacking clicks by rerouting them to their pages. Typically, a clickjacking attack hides the target website’s UI several layers beneath the visible UI. As such, the victims aren’t aware that they are clicking a different website. That’s why these attacks are also commonly known as UI redress or UI redressing attacks. The consequences of these attacks can range from downloading malware, unwittingly giving likes on social networks, purchasing products on eCommerce stores and even transferring money. There are different types of clickjacking attacks, including; ### Password manager attacks This attack deceives password managers to utilize auto-fill functionality. ### Likejacking In Likejacking attack, the attacker hijacks the users’ clicks and converts them into likes on a social media network, such as Facebook. ### Cookiejacking This is where a user is made to interact with a UI element, such as a checkbox, consequently providing the attacker with cookies stored in their browser. This enables the attacker to perform actions on the target site on the user’s behalf. ## Client Side Template Injection (CSTI) Attacks This client-side vulnerability arises when an application running a client-side template framework places user input in web pages. Client-side template injection vulnerabilities are common in Angular because it’s a client-side template framework. In a Client-Side Template Injection Attack, the attacker can steal the victims’ data and use it to perform actions on their behalf. In most cases, CSTI vulnerabilities happen accidentally as a result of poor template design. However, it’s important to note that some CSTI vulnerabilities are sometimes implemented intentionally. For instance, if your website allows user inputs, for instance, allowing editors the privileges to submit templates by design, you’re at a high risk of a CSTI attack as soon as an attacker compromises a user account with such privileges. Angular version 1.2-1.6 has a sandbox that helps prevent CSTI attacks. But this was later found to be ineffective and removed in versions 1.6 and above. If you’re testing a web application running on Angular 1.2 to 1.5, the first step will be to search for the sandbox and bypass it for the payload to execute. ## SQL Injection Attacks Throughout the [history of Angular](https://www.angularjswiki.com/angular/history-of-angularjs/), SQL injection has always been at the top of the most common threats that developers and businesses wrestle with. An SQL Injection Attack is a common web hacking technique where the attacker attempts to insert a malicious SQL code into your database through vulnerable fields, for instance, textboxes. The fact that it’s executed on your database means that your system and sensitive data are at risk, making this the worst form of attack that you should beware of. Following a successful SQL Injection Attack, the malicious actor can: 1. Hack a user’s account 2. Steal and copy sensitive data from the website’s database 3. Delete sensitive data 4. Modify the structure of the database 5. Change or modify the system’s sensitive data 6. See other users’ private information 7. Take control of the database server and command actions at will ## JSONP Attacks JSONP (or JSON with padding) is a widespread attack among sites that rely on the Cross-Origin Resource Sharing (or CORS). The CORS is a concept in which one site provides explicit permission to another site to make specific requests. In addition to CORS, there is JSONP. This is another way of accessing data from another site. Using this method, only an authenticated user can retrieve personal data specific to the users. This means that other sites can’t have access to such data. However, JSONP allows this in the form of a CSRF-style attack. This attack takes advantage of the fact that JSONP relies on the Same-Origin Policy (SOP). The problem with SOP is that it does not prevent sites from executing external `<script>` tags. In that case, the attacker’s site adds a JSONP URL as a script. The browser executes the request and sends cookies for the authenticated user. The JSONP returns data for the authenticated user, which is read by the attacker’s site. The attacker uses the current session to perform additional requests. As you can imagine, this poses a big threat if the second site has sensitive and exploitable data.
58.19403
342
0.792896
eng_Latn
0.998711
87bc6217b658efaa682f4d04bceaebcb94ec152d
262
md
Markdown
matrix_equation/README.md
OyamaZemi/algorithms
a147079950cd432b3f5926b2bb7aac7a5fef66ff
[ "BSD-3-Clause" ]
null
null
null
matrix_equation/README.md
OyamaZemi/algorithms
a147079950cd432b3f5926b2bb7aac7a5fef66ff
[ "BSD-3-Clause" ]
null
null
null
matrix_equation/README.md
OyamaZemi/algorithms
a147079950cd432b3f5926b2bb7aac7a5fef66ff
[ "BSD-3-Clause" ]
null
null
null
行列方程式 ========== [matrix_eqn.py](https://github.com/QuantEcon/QuantEcon.py/blob/master/quantecon/matrix_eqn.py) を読んで,アルゴリズム・実装・パフォーマンスについて改善点を探る. * [Refactor matrix equations (Ricatti, Lyapunov, Sylvester)](https://github.com/QuantEcon/QuantEcon.py/issues/47)
32.75
113
0.763359
yue_Hant
0.452306