Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
5,184
2,910,456,310
IssuesEvent
2015-06-21 19:32:54
Benestar/asparagus
https://api.github.com/repos/Benestar/asparagus
closed
Add a more complex example to README
documentation
We need a more complex usage example in README.md to show how filters, optionals and unions work.
1.0
Add a more complex example to README - We need a more complex usage example in README.md to show how filters, optionals and unions work.
non_process
add a more complex example to readme we need a more complex usage example in readme md to show how filters optionals and unions work
0
201,194
15,802,021,808
IssuesEvent
2021-04-03 07:39:08
rachelljt/ped
https://api.github.com/repos/rachelljt/ped
opened
typos in UG
severity.VeryLow type.DocumentationBug
![Screenshot 2021-04-03 at 3.38.16 PM.png](https://raw.githubusercontent.com/rachelljt/ped/main/files/4bf8903b-d345-48b6-8167-a378f2057508.png) Small issue here but i realised there were many instances of address book being mentioned. I think it would be best to change to Module Planner which is the name of the app. <!--session: 1617435104577-1eaf30cc-d705-4441-bbe1-396c90e57a7c-->
1.0
typos in UG - ![Screenshot 2021-04-03 at 3.38.16 PM.png](https://raw.githubusercontent.com/rachelljt/ped/main/files/4bf8903b-d345-48b6-8167-a378f2057508.png) Small issue here but i realised there were many instances of address book being mentioned. I think it would be best to change to Module Planner which is the name of the app. <!--session: 1617435104577-1eaf30cc-d705-4441-bbe1-396c90e57a7c-->
non_process
typos in ug small issue here but i realised there were many instances of address book being mentioned i think it would be best to change to module planner which is the name of the app
0
601
3,074,419,123
IssuesEvent
2015-08-20 07:06:49
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Support multiple scope names in a single keyscope attribute
DITA 1.3 feature P2 preprocess
I publish the following DITA Map which has key scopes and references in it to XHTML: http://www.oxygenxml.com/forum/files/dot-bugs-testKSNPE.zip and I obtain at some point: BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:268: java.lang.NullPointerException at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:209) at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:84) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
1.0
Support multiple scope names in a single keyscope attribute - I publish the following DITA Map which has key scopes and references in it to XHTML: http://www.oxygenxml.com/forum/files/dot-bugs-testKSNPE.zip and I obtain at some point: BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:268: java.lang.NullPointerException at org.dita.dost.module.KeyrefModule.processFile(KeyrefModule.java:209) at org.dita.dost.module.KeyrefModule.execute(KeyrefModule.java:84) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
process
support multiple scope names in a single keyscope attribute i publish the following dita map which has key scopes and references in it to xhtml and i obtain at some point build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita base build preprocess xml java lang nullpointerexception at org dita dost module keyrefmodule processfile keyrefmodule java at org dita dost module keyrefmodule execute keyrefmodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java
1
286,182
31,335,075,886
IssuesEvent
2023-08-24 05:02:11
istio/istio
https://api.github.com/repos/istio/istio
closed
Re-enable dualuse workload cert generation (CommonName identical to the SAN)
kind/enhancement area/security lifecycle/stale
**Describe the feature request** ***Historical background*** There used to be a feature to instrument `node-agent` to generate workload certificates with a non-empty Subject. Today, Istio's generated workload certificates have an empty subject. ``` openssl x509 -in istio-sidecar-cert.pem -text -noout Certificate: Data: Version: 3 (0x2) Serial Number: fc:bd:66:97:96:e5:66:7a:bc:81:61:e5:18:d4:1b:94 Signature Algorithm: sha256WithRSAEncryption Issuer: O = Istio, CN = Intermediate CA, L = gcp-cluster Validity Not Before: Mar 10 14:13:26 2022 GMT Not After : Mar 11 14:15:26 2022 GMT Subject: Subject Public Key Info: Public Key Algorithm: rsaEncryption RSA Public-Key: (2048 bit) Modulus: ... Exponent: 65537 (0x10001) X509v3 extensions: X509v3 Key Usage: critical Digital Signature, Key Encipherment X509v3 Extended Key Usage: TLS Web Server Authentication, TLS Web Client Authentication X509v3 Basic Constraints: critical CA:FALSE X509v3 Authority Key Identifier: keyid:15:E1:F1:D3:12:D1:D4:C2:AA:26:64:D7:65:AA:6F:BC:0F:A3:01:5E X509v3 Subject Alternative Name: critical URI:spiffe://cluster.local/ns/default/sa/default Signature Algorithm: sha256WithRSAEncryption ... ``` The code to have this functionality is still available in the source tree, but the exposure of a flag or an ENV variable to activate it, is no longer available. Some docs from 1.1 and 1.5 that documented this feature (`--experimental-dual-use` to enable dual-use mode. Generates certificates with a CommonName identical to the SAN.): - https://istio.io/v1.1/docs/reference/commands/node_agent - https://istio.io/v1.5/pt-br/docs/reference/commands/node_agent ***Applicable use cases*** The use case and applicability for this feature is enormous. Every external service that Istio meshed workloads want to connect to, and have support for mTLS AuthN AND AuthZ based on Client certificates, can now leverage Istio workload ID based on SPIFFE to enforce AuthZ. > (IstioWorkLoad | IstioSideCar) ===[mTLS with Client AuthN/Z based on SpiffeID]===> ExternalService (no sidecar or proxy) Some examples include: - **MongoDB:** reference documentation [here](https://www.mongodb.com/docs/manual/tutorial/configure-x509-client-authentication/#add-x-509-certificate-subject-as-a-user) - **Kafka:** reference documentation [here](https://docs.confluent.io/platform/current/kafka/authorization.html#tls-ssl-principal-user-names) - **PostGres:** reference documentation [here](https://www.postgresql.org/docs/current/auth-cert.html) - ... Most services like DBs and Queues do support this **Client Cert Subject** based AuthNZ mechanism. All of them however, do NOT support **Client Cert X509 SAN Extension** based AuthNZ mechanism. > **NOTE:** the use case is to perform AuthN **AND** AuthZ with that external service, without the use of an intermediate egress GW or destination sidecar/proxy. We want to be independent from the wire protocol support of Envoy and rely on mTLS on top of any TCP connection. The external service could even be a SaaS service, when it allows its consumers to configure the mTLS configuration in terms of Certs/Subjects/etc and corresponding Authorization/ACL mappings. ***Proposal*** A solution would be to expose it as an ENV variable for pilot agent like this: - https://istio.io/latest/docs/reference/commands/pilot-agent/ Variable Name | Type | Default Value | Description -- | -- | -- | -- WORKLOAD_RSA_KEY_SIZE | Integer | 2048 | Specify the RSA key size to use for workload certificates. WORKLOAD_DUAL_USE_CERT | Bool | false | Enable dual-use mode. Generates certificates with a CommonName identical to the SAN. > **REMARK:** as a more flexible solution, end users would also have the option to determine which DN field the SPIFFE ID is placed in (CN, OU, O, ST or C) **Describe alternatives you've considered** The alternatives include - wait for every external DB or Queueing solution to adopt SPIFFE SAN based AuthN/Z. Istio provides a strong and flexible workload ID mechanism, but we are only able to leverage this within the Istio eco-system itself right now. - have a separate mechanism and/or set of credentials to AuthN/Z with the external service, potentially. This generally involves abandoning mTLS and/or setting up a secondary TLS connection, and/or maintaining and injecting traditional username/password credentials within workloads. **Affected product area (please put an X in all that apply)** [ ] Ambient [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Extensions and Telemetry [x ] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Affected features (please put an X in all that apply)** [ ] Multi Cluster [ ] Virtual Machine [ ] Multi Control Plane [ x] External Connectivity **Additional context** Within the existing code base, check for traces of the `options.IsDualUse` flag, spread across various source files.
True
Re-enable dualuse workload cert generation (CommonName identical to the SAN) - **Describe the feature request** ***Historical background*** There used to be a feature to instrument `node-agent` to generate workload certificates with a non-empty Subject. Today, Istio's generated workload certificates have an empty subject. ``` openssl x509 -in istio-sidecar-cert.pem -text -noout Certificate: Data: Version: 3 (0x2) Serial Number: fc:bd:66:97:96:e5:66:7a:bc:81:61:e5:18:d4:1b:94 Signature Algorithm: sha256WithRSAEncryption Issuer: O = Istio, CN = Intermediate CA, L = gcp-cluster Validity Not Before: Mar 10 14:13:26 2022 GMT Not After : Mar 11 14:15:26 2022 GMT Subject: Subject Public Key Info: Public Key Algorithm: rsaEncryption RSA Public-Key: (2048 bit) Modulus: ... Exponent: 65537 (0x10001) X509v3 extensions: X509v3 Key Usage: critical Digital Signature, Key Encipherment X509v3 Extended Key Usage: TLS Web Server Authentication, TLS Web Client Authentication X509v3 Basic Constraints: critical CA:FALSE X509v3 Authority Key Identifier: keyid:15:E1:F1:D3:12:D1:D4:C2:AA:26:64:D7:65:AA:6F:BC:0F:A3:01:5E X509v3 Subject Alternative Name: critical URI:spiffe://cluster.local/ns/default/sa/default Signature Algorithm: sha256WithRSAEncryption ... ``` The code to have this functionality is still available in the source tree, but the exposure of a flag or an ENV variable to activate it, is no longer available. Some docs from 1.1 and 1.5 that documented this feature (`--experimental-dual-use` to enable dual-use mode. Generates certificates with a CommonName identical to the SAN.): - https://istio.io/v1.1/docs/reference/commands/node_agent - https://istio.io/v1.5/pt-br/docs/reference/commands/node_agent ***Applicable use cases*** The use case and applicability for this feature is enormous. Every external service that Istio meshed workloads want to connect to, and have support for mTLS AuthN AND AuthZ based on Client certificates, can now leverage Istio workload ID based on SPIFFE to enforce AuthZ. > (IstioWorkLoad | IstioSideCar) ===[mTLS with Client AuthN/Z based on SpiffeID]===> ExternalService (no sidecar or proxy) Some examples include: - **MongoDB:** reference documentation [here](https://www.mongodb.com/docs/manual/tutorial/configure-x509-client-authentication/#add-x-509-certificate-subject-as-a-user) - **Kafka:** reference documentation [here](https://docs.confluent.io/platform/current/kafka/authorization.html#tls-ssl-principal-user-names) - **PostGres:** reference documentation [here](https://www.postgresql.org/docs/current/auth-cert.html) - ... Most services like DBs and Queues do support this **Client Cert Subject** based AuthNZ mechanism. All of them however, do NOT support **Client Cert X509 SAN Extension** based AuthNZ mechanism. > **NOTE:** the use case is to perform AuthN **AND** AuthZ with that external service, without the use of an intermediate egress GW or destination sidecar/proxy. We want to be independent from the wire protocol support of Envoy and rely on mTLS on top of any TCP connection. The external service could even be a SaaS service, when it allows its consumers to configure the mTLS configuration in terms of Certs/Subjects/etc and corresponding Authorization/ACL mappings. ***Proposal*** A solution would be to expose it as an ENV variable for pilot agent like this: - https://istio.io/latest/docs/reference/commands/pilot-agent/ Variable Name | Type | Default Value | Description -- | -- | -- | -- WORKLOAD_RSA_KEY_SIZE | Integer | 2048 | Specify the RSA key size to use for workload certificates. WORKLOAD_DUAL_USE_CERT | Bool | false | Enable dual-use mode. Generates certificates with a CommonName identical to the SAN. > **REMARK:** as a more flexible solution, end users would also have the option to determine which DN field the SPIFFE ID is placed in (CN, OU, O, ST or C) **Describe alternatives you've considered** The alternatives include - wait for every external DB or Queueing solution to adopt SPIFFE SAN based AuthN/Z. Istio provides a strong and flexible workload ID mechanism, but we are only able to leverage this within the Istio eco-system itself right now. - have a separate mechanism and/or set of credentials to AuthN/Z with the external service, potentially. This generally involves abandoning mTLS and/or setting up a secondary TLS connection, and/or maintaining and injecting traditional username/password credentials within workloads. **Affected product area (please put an X in all that apply)** [ ] Ambient [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Extensions and Telemetry [x ] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Affected features (please put an X in all that apply)** [ ] Multi Cluster [ ] Virtual Machine [ ] Multi Control Plane [ x] External Connectivity **Additional context** Within the existing code base, check for traces of the `options.IsDualUse` flag, spread across various source files.
non_process
re enable dualuse workload cert generation commonname identical to the san describe the feature request historical background there used to be a feature to instrument node agent to generate workload certificates with a non empty subject today istio s generated workload certificates have an empty subject openssl in istio sidecar cert pem text noout certificate data version serial number fc bd bc signature algorithm issuer o istio cn intermediate ca l gcp cluster validity not before mar gmt not after mar gmt subject subject public key info public key algorithm rsaencryption rsa public key bit modulus exponent extensions key usage critical digital signature key encipherment extended key usage tls web server authentication tls web client authentication basic constraints critical ca false authority key identifier keyid aa aa bc subject alternative name critical uri spiffe cluster local ns default sa default signature algorithm the code to have this functionality is still available in the source tree but the exposure of a flag or an env variable to activate it is no longer available some docs from and that documented this feature experimental dual use to enable dual use mode generates certificates with a commonname identical to the san applicable use cases the use case and applicability for this feature is enormous every external service that istio meshed workloads want to connect to and have support for mtls authn and authz based on client certificates can now leverage istio workload id based on spiffe to enforce authz istioworkload istiosidecar externalservice no sidecar or proxy some examples include mongodb reference documentation kafka reference documentation postgres reference documentation most services like dbs and queues do support this client cert subject based authnz mechanism all of them however do not support client cert san extension based authnz mechanism note the use case is to perform authn and authz with that external service without the use of an intermediate egress gw or destination sidecar proxy we want to be independent from the wire protocol support of envoy and rely on mtls on top of any tcp connection the external service could even be a saas service when it allows its consumers to configure the mtls configuration in terms of certs subjects etc and corresponding authorization acl mappings proposal a solution would be to expose it as an env variable for pilot agent like this variable name type default value description workload rsa key size integer specify the rsa key size to use for workload certificates workload dual use cert bool false enable dual use mode generates certificates with a commonname identical to the san remark as a more flexible solution end users would also have the option to determine which dn field the spiffe id is placed in cn ou o st or c describe alternatives you ve considered the alternatives include wait for every external db or queueing solution to adopt spiffe san based authn z istio provides a strong and flexible workload id mechanism but we are only able to leverage this within the istio eco system itself right now have a separate mechanism and or set of credentials to authn z with the external service potentially this generally involves abandoning mtls and or setting up a secondary tls connection and or maintaining and injecting traditional username password credentials within workloads affected product area please put an x in all that apply ambient docs installation networking performance and scalability extensions and telemetry security test and release user experience developer infrastructure affected features please put an x in all that apply multi cluster virtual machine multi control plane external connectivity additional context within the existing code base check for traces of the options isdualuse flag spread across various source files
0
84,626
15,724,730,485
IssuesEvent
2021-03-29 09:09:10
crouchr/learnage
https://api.github.com/repos/crouchr/learnage
opened
CVE-2013-2765 (Medium) detected in mod-securitymodsecurity-apache_2.5.11
security vulnerability
## CVE-2013-2765 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mod-securitymodsecurity-apache_2.5.11</b></p></summary> <p> <p>Library home page: <a href=https://sourceforge.net/projects/mod-security/>https://sourceforge.net/projects/mod-security/</a></p> <p>Found in HEAD commit: <a href="https://github.com/crouchr/learnage/commit/a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c">a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>learnage/blackrain2020/original-sources-3rd-party/modsecurity-apache_2.5.11/modsecurity-apache_2.5.11/apache2/msc_reqbody.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ModSecurity module before 2.7.4 for the Apache HTTP Server allows remote attackers to cause a denial of service (NULL pointer dereference, process crash, and disk consumption) via a POST request with a large body and a crafted Content-Type header. <p>Publish Date: 2013-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-2765>CVE-2013-2765</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=967615">https://bugzilla.redhat.com/show_bug.cgi?id=967615</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: Upgrade to version ModSecurity-2.7.4 or greater</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-2765 (Medium) detected in mod-securitymodsecurity-apache_2.5.11 - ## CVE-2013-2765 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mod-securitymodsecurity-apache_2.5.11</b></p></summary> <p> <p>Library home page: <a href=https://sourceforge.net/projects/mod-security/>https://sourceforge.net/projects/mod-security/</a></p> <p>Found in HEAD commit: <a href="https://github.com/crouchr/learnage/commit/a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c">a5f2b4a6eb346dbe0def97e83877b169dc4b8f8c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>learnage/blackrain2020/original-sources-3rd-party/modsecurity-apache_2.5.11/modsecurity-apache_2.5.11/apache2/msc_reqbody.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ModSecurity module before 2.7.4 for the Apache HTTP Server allows remote attackers to cause a denial of service (NULL pointer dereference, process crash, and disk consumption) via a POST request with a large body and a crafted Content-Type header. <p>Publish Date: 2013-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-2765>CVE-2013-2765</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=967615">https://bugzilla.redhat.com/show_bug.cgi?id=967615</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: Upgrade to version ModSecurity-2.7.4 or greater</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in mod securitymodsecurity apache cve medium severity vulnerability vulnerable library mod securitymodsecurity apache library home page a href found in head commit a href found in base branch master vulnerable source files learnage original sources party modsecurity apache modsecurity apache msc reqbody c vulnerability details the modsecurity module before for the apache http server allows remote attackers to cause a denial of service null pointer dereference process crash and disk consumption via a post request with a large body and a crafted content type header publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution upgrade to version modsecurity or greater step up your open source security game with whitesource
0
8,828
11,940,092,970
IssuesEvent
2020-04-02 16:09:42
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
opened
Spanner: TestDataTypes timing out on CI
api: spanner priority: p1 type: process
After 10 minutes. (Locally it passes after around 2 minutes.) I've skipped it in #1001 but it should be looked at. One thing would be to separate out in single tests. At least we could isolate the culprit easier. @skuruppu feel free to assign to someone from your team that can take care of it.
1.0
Spanner: TestDataTypes timing out on CI - After 10 minutes. (Locally it passes after around 2 minutes.) I've skipped it in #1001 but it should be looked at. One thing would be to separate out in single tests. At least we could isolate the culprit easier. @skuruppu feel free to assign to someone from your team that can take care of it.
process
spanner testdatatypes timing out on ci after minutes locally it passes after around minutes i ve skipped it in but it should be looked at one thing would be to separate out in single tests at least we could isolate the culprit easier skuruppu feel free to assign to someone from your team that can take care of it
1
22,546
11,743,727,138
IssuesEvent
2020-03-12 05:35:14
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Calls Between App Service Environments
Pri2 app-service/svc cxp product-question triaged
As I understand, the "Calls Between App Service Environments" section describes situation when "ASE Two" is external ASE. But what if it's an ILB ASE? In this case, it won't be an "Internet" call, right? The ILB address will be used as destination and "ASE One" subnet address - as a source, correct? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 27797a88-1843-5f70-b505-873200f3e0bc * Version Independent ID: 4cfadca3-e382-15c1-e9a1-a61f839e9153 * Content: [Network architecture v1 - Azure App Service Environment](https://docs.microsoft.com/en-us/azure/app-service/environment/app-service-app-service-environment-network-architecture-overview) * Content Source: [articles/app-service/environment/app-service-app-service-environment-network-architecture-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/environment/app-service-app-service-environment-network-architecture-overview.md) * Service: **app-service** * GitHub Login: @stefsch * Microsoft Alias: **stefsch**
1.0
Calls Between App Service Environments - As I understand, the "Calls Between App Service Environments" section describes situation when "ASE Two" is external ASE. But what if it's an ILB ASE? In this case, it won't be an "Internet" call, right? The ILB address will be used as destination and "ASE One" subnet address - as a source, correct? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 27797a88-1843-5f70-b505-873200f3e0bc * Version Independent ID: 4cfadca3-e382-15c1-e9a1-a61f839e9153 * Content: [Network architecture v1 - Azure App Service Environment](https://docs.microsoft.com/en-us/azure/app-service/environment/app-service-app-service-environment-network-architecture-overview) * Content Source: [articles/app-service/environment/app-service-app-service-environment-network-architecture-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/environment/app-service-app-service-environment-network-architecture-overview.md) * Service: **app-service** * GitHub Login: @stefsch * Microsoft Alias: **stefsch**
non_process
calls between app service environments as i understand the calls between app service environments section describes situation when ase two is external ase but what if it s an ilb ase in this case it won t be an internet call right the ilb address will be used as destination and ase one subnet address as a source correct document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login stefsch microsoft alias stefsch
0
9,887
12,889,437,717
IssuesEvent
2020-07-13 14:32:09
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Stages is no longer in-preview
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
The page has a note about stages being in-preview, they are GA so the note can be removed. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 4266f72c-c774-0046-4593-d01eb775d3c3 * Version Independent ID: f20827aa-a6c5-96a8-5969-e576ffbc2e38 * Content: [Stages in Azure Pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/stages?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/stages.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/stages.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Stages is no longer in-preview - The page has a note about stages being in-preview, they are GA so the note can be removed. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 4266f72c-c774-0046-4593-d01eb775d3c3 * Version Independent ID: f20827aa-a6c5-96a8-5969-e576ffbc2e38 * Content: [Stages in Azure Pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/stages?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/stages.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/stages.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
stages is no longer in preview the page has a note about stages being in preview they are ga so the note can be removed document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
7,375
10,513,808,742
IssuesEvent
2019-09-27 21:42:00
googlemaps/google-maps-services-java
https://api.github.com/repos/googlemaps/google-maps-services-java
closed
Include package-list file in Javadoc
help wanted priority: p3 type: process
Please include the `package-list` file when publishing Javadocs so that we can link to it when generating our own Javadocs. For example, I would expect to find it at the below URL, but a 404 is returned. https://googlemaps.github.io/google-maps-services-java/latest/javadoc/package-list
1.0
Include package-list file in Javadoc - Please include the `package-list` file when publishing Javadocs so that we can link to it when generating our own Javadocs. For example, I would expect to find it at the below URL, but a 404 is returned. https://googlemaps.github.io/google-maps-services-java/latest/javadoc/package-list
process
include package list file in javadoc please include the package list file when publishing javadocs so that we can link to it when generating our own javadocs for example i would expect to find it at the below url but a is returned
1
20,825
27,580,477,537
IssuesEvent
2023-03-08 15:53:39
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Process: support tiers for hardware for Zephyr v3.3 and later
Process
This issue covers a process discussion for introducing support tiers for boards with upstream Zephyr project support. The goal of this issue is to define a high level process for defining the support level for different Zephyr boards. This will aid users in knowing which platforms are well tested and debugged. For a related release criteria issue, see https://github.com/zephyrproject-rtos/zephyr/issues/46759. Proposal for board tier names and descriptions: - Tier 0: emulation boards used in CI; failures can block PRs. Supported by the Zephyr project itself, commitment to fix bugs in releases. One is required for each new architecture. - Tier 1: "real hardware" with commitment from a specific team to run tests using twister device testing for the "Zephyr compatibility test suite" (details TBD) mentioned above on a regular basis using open drivers. Commitment to fix bugs in time for releases. Not supported by "Zephyr Project" itself. Other quality of developer experience criteria TBD. General availability for purchase (modulo silicon supply chain crisis). - Tier 2: board implementation is available in upstream, no commitment to testing, may not be generally available. TL;DR no guarantees, if it breaks you get to keep both pieces. Has a dedicated maintainer who commits to respond to issues / review patches, however. - Tier 3: Deprecated board. Board implementation is available, but no owner or unresponsive owner. No commitment to support is available. May be removed from upstream if no one works to bring it up to tier 2 or better. If the last board for a particular SoC is removed, the SoC and drivers will be removed as well. This will also affect the release criteria for future versions of zephyr: - maintainers will guarantee testing for boards in higher-priority support tiers - bugs in boards in higher-priority support tiers are more likely to block the release Once the high level process is defined, the goal is for the testing working group to manage the details with input from interested stakeholders. Previous discussion: https://docs.google.com/document/d/1LOhgronKx3TxUypYLy9nex0_3UC58FMMjbXV4wbiuiE/edit#
1.0
Process: support tiers for hardware for Zephyr v3.3 and later - This issue covers a process discussion for introducing support tiers for boards with upstream Zephyr project support. The goal of this issue is to define a high level process for defining the support level for different Zephyr boards. This will aid users in knowing which platforms are well tested and debugged. For a related release criteria issue, see https://github.com/zephyrproject-rtos/zephyr/issues/46759. Proposal for board tier names and descriptions: - Tier 0: emulation boards used in CI; failures can block PRs. Supported by the Zephyr project itself, commitment to fix bugs in releases. One is required for each new architecture. - Tier 1: "real hardware" with commitment from a specific team to run tests using twister device testing for the "Zephyr compatibility test suite" (details TBD) mentioned above on a regular basis using open drivers. Commitment to fix bugs in time for releases. Not supported by "Zephyr Project" itself. Other quality of developer experience criteria TBD. General availability for purchase (modulo silicon supply chain crisis). - Tier 2: board implementation is available in upstream, no commitment to testing, may not be generally available. TL;DR no guarantees, if it breaks you get to keep both pieces. Has a dedicated maintainer who commits to respond to issues / review patches, however. - Tier 3: Deprecated board. Board implementation is available, but no owner or unresponsive owner. No commitment to support is available. May be removed from upstream if no one works to bring it up to tier 2 or better. If the last board for a particular SoC is removed, the SoC and drivers will be removed as well. This will also affect the release criteria for future versions of zephyr: - maintainers will guarantee testing for boards in higher-priority support tiers - bugs in boards in higher-priority support tiers are more likely to block the release Once the high level process is defined, the goal is for the testing working group to manage the details with input from interested stakeholders. Previous discussion: https://docs.google.com/document/d/1LOhgronKx3TxUypYLy9nex0_3UC58FMMjbXV4wbiuiE/edit#
process
process support tiers for hardware for zephyr and later this issue covers a process discussion for introducing support tiers for boards with upstream zephyr project support the goal of this issue is to define a high level process for defining the support level for different zephyr boards this will aid users in knowing which platforms are well tested and debugged for a related release criteria issue see proposal for board tier names and descriptions tier emulation boards used in ci failures can block prs supported by the zephyr project itself commitment to fix bugs in releases one is required for each new architecture tier real hardware with commitment from a specific team to run tests using twister device testing for the zephyr compatibility test suite details tbd mentioned above on a regular basis using open drivers commitment to fix bugs in time for releases not supported by zephyr project itself other quality of developer experience criteria tbd general availability for purchase modulo silicon supply chain crisis tier board implementation is available in upstream no commitment to testing may not be generally available tl dr no guarantees if it breaks you get to keep both pieces has a dedicated maintainer who commits to respond to issues review patches however tier deprecated board board implementation is available but no owner or unresponsive owner no commitment to support is available may be removed from upstream if no one works to bring it up to tier or better if the last board for a particular soc is removed the soc and drivers will be removed as well this will also affect the release criteria for future versions of zephyr maintainers will guarantee testing for boards in higher priority support tiers bugs in boards in higher priority support tiers are more likely to block the release once the high level process is defined the goal is for the testing working group to manage the details with input from interested stakeholders previous discussion
1
5,291
8,074,398,662
IssuesEvent
2018-08-06 23:05:15
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Enhance EKS driver
area/cluster process/needs-ui status/reopened status/resolved status/to-test version/2.0
Add these additional features if requested: > 1. Amazon suggests using a reusable service role for creating clusters. Right now we are creating a new service role for each EKS cluster we spin up. We could change Rancher to use an existing role that the user specifics, or have Rancher only create the service role if it doesn't exist (or both). Yes, we should be able to use an existing cluster service role. > 2. We are currently bypassing heptio authentication. Do we want to investigate either integrating AWS's EKS authentication system into Rancher? Not as of now > 3. We create a new VPC/Subnet for each EKS cluster. Can we allow the user to specify a VPC or Subnet? There are some specific requirements around what VPCs and Subnets should look like. Can we have Rancher automatically update a VPC or Subnet if it is not configured correctly? Yes, we should give flexibility to VPC/subnet. We should not automatically update a VPC/Subnet, but we should document what are the VPC/subnet requirements. > 4. Specifying security groups have the same problems as specifying VPCs and subnets (and may be required if specifying a VPC/subnet. We should give the ability to let them select their own, but document on what needs to happen for security group. > 5. Right now all VPCs are exposed to the public internet by default, maybe add an option to not do this. Yes, let's add this option when we create VPC, but not if they are using an existing one. > 6. We don't support modifying existing clusters at all right now. Looks like EKS clusters can't be modified (which means we can't change VPC, subnet, security groups, or even k8s version), but VPCs/subnets and worker nodes can. Hold off on this for now. > 7. Amazon says to not use root credentials when creating an EKS cluster, might be a good idea to remind the user of this or check and fail if root credentials are supplied. Add a note in the UI for this > 8. DNS propagation takes a few minutes when exposing a pod with a layer-4 LB. We should see if there is a way to show this to the user (healthchecks?) File a separate issue for this. We have this issue for all k8s managed providers. > 9. Our deletion process leaves a lot of resources dangling. When we delete VPCs it is often blocked because of network interfaces and LBs. Can we have rancher automatically find and delete these resources so that the VPC is deleted successfully? No, we should only delete the nodes themselves and put a note reminding the users that these network related resources are not deleted and need to be cleaned up themselves. ---------- Further details from talking with @vincent99 ``` 1. Vince will lookup the roles from the UI (no new api endpoint needed). If a role gets passed in, eks driver will use it, otherwise create a new one. We will need to modify EKS driver to support this. 3. For the UI piece Vince will make a API call to list Virtual Networks and Subnets. If no network is passed in we will create one. 4. For the UI piece Vince will make a API call to list security groups. If no security group is passed in we will create one. Document requirements for security groups. 5. Just checkbox/boolean option. Would not be applicable if using existing virtual network. 7. Vince will add note about root credentials or see if we can check. No backend work needed. 9. Vince says we already do this. ``` For hosting the templates we will vendor them inside kontainer-engine so we don't have to make an http call.
1.0
Enhance EKS driver - Add these additional features if requested: > 1. Amazon suggests using a reusable service role for creating clusters. Right now we are creating a new service role for each EKS cluster we spin up. We could change Rancher to use an existing role that the user specifics, or have Rancher only create the service role if it doesn't exist (or both). Yes, we should be able to use an existing cluster service role. > 2. We are currently bypassing heptio authentication. Do we want to investigate either integrating AWS's EKS authentication system into Rancher? Not as of now > 3. We create a new VPC/Subnet for each EKS cluster. Can we allow the user to specify a VPC or Subnet? There are some specific requirements around what VPCs and Subnets should look like. Can we have Rancher automatically update a VPC or Subnet if it is not configured correctly? Yes, we should give flexibility to VPC/subnet. We should not automatically update a VPC/Subnet, but we should document what are the VPC/subnet requirements. > 4. Specifying security groups have the same problems as specifying VPCs and subnets (and may be required if specifying a VPC/subnet. We should give the ability to let them select their own, but document on what needs to happen for security group. > 5. Right now all VPCs are exposed to the public internet by default, maybe add an option to not do this. Yes, let's add this option when we create VPC, but not if they are using an existing one. > 6. We don't support modifying existing clusters at all right now. Looks like EKS clusters can't be modified (which means we can't change VPC, subnet, security groups, or even k8s version), but VPCs/subnets and worker nodes can. Hold off on this for now. > 7. Amazon says to not use root credentials when creating an EKS cluster, might be a good idea to remind the user of this or check and fail if root credentials are supplied. Add a note in the UI for this > 8. DNS propagation takes a few minutes when exposing a pod with a layer-4 LB. We should see if there is a way to show this to the user (healthchecks?) File a separate issue for this. We have this issue for all k8s managed providers. > 9. Our deletion process leaves a lot of resources dangling. When we delete VPCs it is often blocked because of network interfaces and LBs. Can we have rancher automatically find and delete these resources so that the VPC is deleted successfully? No, we should only delete the nodes themselves and put a note reminding the users that these network related resources are not deleted and need to be cleaned up themselves. ---------- Further details from talking with @vincent99 ``` 1. Vince will lookup the roles from the UI (no new api endpoint needed). If a role gets passed in, eks driver will use it, otherwise create a new one. We will need to modify EKS driver to support this. 3. For the UI piece Vince will make a API call to list Virtual Networks and Subnets. If no network is passed in we will create one. 4. For the UI piece Vince will make a API call to list security groups. If no security group is passed in we will create one. Document requirements for security groups. 5. Just checkbox/boolean option. Would not be applicable if using existing virtual network. 7. Vince will add note about root credentials or see if we can check. No backend work needed. 9. Vince says we already do this. ``` For hosting the templates we will vendor them inside kontainer-engine so we don't have to make an http call.
process
enhance eks driver add these additional features if requested amazon suggests using a reusable service role for creating clusters right now we are creating a new service role for each eks cluster we spin up we could change rancher to use an existing role that the user specifics or have rancher only create the service role if it doesn t exist or both yes we should be able to use an existing cluster service role we are currently bypassing heptio authentication do we want to investigate either integrating aws s eks authentication system into rancher not as of now we create a new vpc subnet for each eks cluster can we allow the user to specify a vpc or subnet there are some specific requirements around what vpcs and subnets should look like can we have rancher automatically update a vpc or subnet if it is not configured correctly yes we should give flexibility to vpc subnet we should not automatically update a vpc subnet but we should document what are the vpc subnet requirements specifying security groups have the same problems as specifying vpcs and subnets and may be required if specifying a vpc subnet we should give the ability to let them select their own but document on what needs to happen for security group right now all vpcs are exposed to the public internet by default maybe add an option to not do this yes let s add this option when we create vpc but not if they are using an existing one we don t support modifying existing clusters at all right now looks like eks clusters can t be modified which means we can t change vpc subnet security groups or even version but vpcs subnets and worker nodes can hold off on this for now amazon says to not use root credentials when creating an eks cluster might be a good idea to remind the user of this or check and fail if root credentials are supplied add a note in the ui for this dns propagation takes a few minutes when exposing a pod with a layer lb we should see if there is a way to show this to the user healthchecks file a separate issue for this we have this issue for all managed providers our deletion process leaves a lot of resources dangling when we delete vpcs it is often blocked because of network interfaces and lbs can we have rancher automatically find and delete these resources so that the vpc is deleted successfully no we should only delete the nodes themselves and put a note reminding the users that these network related resources are not deleted and need to be cleaned up themselves further details from talking with vince will lookup the roles from the ui no new api endpoint needed if a role gets passed in eks driver will use it otherwise create a new one we will need to modify eks driver to support this for the ui piece vince will make a api call to list virtual networks and subnets if no network is passed in we will create one for the ui piece vince will make a api call to list security groups if no security group is passed in we will create one document requirements for security groups just checkbox boolean option would not be applicable if using existing virtual network vince will add note about root credentials or see if we can check no backend work needed vince says we already do this for hosting the templates we will vendor them inside kontainer engine so we don t have to make an http call
1
10,183
13,044,162,858
IssuesEvent
2020-07-29 03:47:37
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `TruncateUint` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `TruncateUint` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `TruncateUint` from TiDB - ## Description Port the scalar function `TruncateUint` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function truncateuint from tidb description port the scalar function truncateuint from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
13,512
8,247,838,271
IssuesEvent
2018-09-11 16:37:43
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
closed
Build module/nomodule scripts for v0.js
Category: Runtime Related to: Performance
AMP should be outputting v0-module.js and v0-nmodule.js in order to run these scripts inside `<script type=module` or `<script nomodule` for the same. These outputs will be identical in build except the perf.tick and perf.addEnabledExperiment parameters as shown here: https://github.com/ampproject/amphtml/blob/master/src/amp.js#L69-L74 - [x] Add experiment in v0.js - [x] Modify build script to yield both versions
True
Build module/nomodule scripts for v0.js - AMP should be outputting v0-module.js and v0-nmodule.js in order to run these scripts inside `<script type=module` or `<script nomodule` for the same. These outputs will be identical in build except the perf.tick and perf.addEnabledExperiment parameters as shown here: https://github.com/ampproject/amphtml/blob/master/src/amp.js#L69-L74 - [x] Add experiment in v0.js - [x] Modify build script to yield both versions
non_process
build module nomodule scripts for js amp should be outputting module js and nmodule js in order to run these scripts inside script type module or script nomodule for the same these outputs will be identical in build except the perf tick and perf addenabledexperiment parameters as shown here add experiment in js modify build script to yield both versions
0
16,529
21,557,297,025
IssuesEvent
2022-04-30 16:40:10
CynthiaChuang/CynthiaChuang.github.io
https://api.github.com/repos/CynthiaChuang/CynthiaChuang.github.io
closed
Typo in /HedgeDoc-a-New-Fork-of-CodiMD
Non Issue Processing Typo
https://cynthiachuang.github.io/HedgeDoc-a-New-Fork-of-CodiMD/ 因此基於種種原因,社群版的 CodiMD 後來更名成為 HedgeDoc,並將 `log` 換成一隻可愛小刺蝟。
1.0
Typo in /HedgeDoc-a-New-Fork-of-CodiMD - https://cynthiachuang.github.io/HedgeDoc-a-New-Fork-of-CodiMD/ 因此基於種種原因,社群版的 CodiMD 後來更名成為 HedgeDoc,並將 `log` 換成一隻可愛小刺蝟。
process
typo in hedgedoc a new fork of codimd 因此基於種種原因,社群版的 codimd 後來更名成為 hedgedoc,並將 log 換成一隻可愛小刺蝟。
1
55,120
30,593,549,104
IssuesEvent
2023-07-21 19:23:55
nexB/scancode-workbench
https://api.github.com/repos/nexB/scancode-workbench
closed
Blank space on BarChart view
bug D3 performance should have
<img width="1988" alt="screen shot 2018-08-24 at 6 30 18 pm" src="https://user-images.githubusercontent.com/7485204/44613682-de06fe00-a7cb-11e8-8eca-65b551e039a9.png"> As you can see from the above picture, there is a large blank space that shows up occasionally on the Bar Chart view. It seems the more items on the Bar Chart, the more blank space there is. Ultimately, we need to git rid of this white space. It is not useful at all, and makes scrolling around in this view much worse UX-wise.
True
Blank space on BarChart view - <img width="1988" alt="screen shot 2018-08-24 at 6 30 18 pm" src="https://user-images.githubusercontent.com/7485204/44613682-de06fe00-a7cb-11e8-8eca-65b551e039a9.png"> As you can see from the above picture, there is a large blank space that shows up occasionally on the Bar Chart view. It seems the more items on the Bar Chart, the more blank space there is. Ultimately, we need to git rid of this white space. It is not useful at all, and makes scrolling around in this view much worse UX-wise.
non_process
blank space on barchart view img width alt screen shot at pm src as you can see from the above picture there is a large blank space that shows up occasionally on the bar chart view it seems the more items on the bar chart the more blank space there is ultimately we need to git rid of this white space it is not useful at all and makes scrolling around in this view much worse ux wise
0
16,770
21,944,735,764
IssuesEvent
2022-05-23 22:26:25
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Batch Nominatim Geocoder output directly saved to file has no CRS
Processing Bug Projections/Transformations
### What is the bug or the crash? When batch-geocoding a table with the Batch Nominatim Geocoder and saving directly to file, the loaded layer does not have a CRS. It should have EPSG:4326. Both GPKG and SHP exports fail at storing the right CRS. However, _not_ saving as file but loading a temporary scratch layer does work (and one can then make the layer permanent to save the data to file). ### Steps to reproduce the issue 1. Create a CSV file which contains locations, for example: ``` city,responses "Brisbane, Australia",2 "Canberra, Australia",34 ``` 2. Load the CSV file as a layer 3. Open Processing Toolbox, open the Batch Nominatim geocoder 4. Use the table as the Input Layer, and the "city" field as the Address field 5. Use "Save to File..." or "Save to GeoPackage..." in the Geocoded field 6. Make sure "Open output file..." is ticked 7. Run the tool Result: no CRS for the layer. One can assign EPSG:4326 to fix it. ### Versions QGIS version | 3.22.6-Białowieża | QGIS code revision | b6f33b1a9f -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.0.4 PROJ version | 6.3.1 EPSG Registry database version | v9.8.6 (2020-01-22) Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1 SQLite version | 3.31.1 PDAL version | 2.0.1 PostgreSQL client version | 12.9 (Ubuntu 12.9-0ubuntu0.20.04.1) SpatiaLite version | 4.3.0a QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Ubuntu 20.04.4 LTS   |   |   |   Active Python plugins QuickOSM | 2.0.1 quick_map_services | 0.19.29 processing | 2.12.99 sagaprovider | 2.12.99 MetaSearch | 0.3.5 grassprovider | 2.12.99 db_manager | 0.1.20 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Batch Nominatim Geocoder output directly saved to file has no CRS - ### What is the bug or the crash? When batch-geocoding a table with the Batch Nominatim Geocoder and saving directly to file, the loaded layer does not have a CRS. It should have EPSG:4326. Both GPKG and SHP exports fail at storing the right CRS. However, _not_ saving as file but loading a temporary scratch layer does work (and one can then make the layer permanent to save the data to file). ### Steps to reproduce the issue 1. Create a CSV file which contains locations, for example: ``` city,responses "Brisbane, Australia",2 "Canberra, Australia",34 ``` 2. Load the CSV file as a layer 3. Open Processing Toolbox, open the Batch Nominatim geocoder 4. Use the table as the Input Layer, and the "city" field as the Address field 5. Use "Save to File..." or "Save to GeoPackage..." in the Geocoded field 6. Make sure "Open output file..." is ticked 7. Run the tool Result: no CRS for the layer. One can assign EPSG:4326 to fix it. ### Versions QGIS version | 3.22.6-Białowieża | QGIS code revision | b6f33b1a9f -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.0.4 PROJ version | 6.3.1 EPSG Registry database version | v9.8.6 (2020-01-22) Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1 SQLite version | 3.31.1 PDAL version | 2.0.1 PostgreSQL client version | 12.9 (Ubuntu 12.9-0ubuntu0.20.04.1) SpatiaLite version | 4.3.0a QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Ubuntu 20.04.4 LTS   |   |   |   Active Python plugins QuickOSM | 2.0.1 quick_map_services | 0.19.29 processing | 2.12.99 sagaprovider | 2.12.99 MetaSearch | 0.3.5 grassprovider | 2.12.99 db_manager | 0.1.20 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
batch nominatim geocoder output directly saved to file has no crs what is the bug or the crash when batch geocoding a table with the batch nominatim geocoder and saving directly to file the loaded layer does not have a crs it should have epsg both gpkg and shp exports fail at storing the right crs however not saving as file but loading a temporary scratch layer does work and one can then make the layer permanent to save the data to file steps to reproduce the issue create a csv file which contains locations for example city responses brisbane australia canberra australia load the csv file as a layer open processing toolbox open the batch nominatim geocoder use the table as the input layer and the city field as the address field use save to file or save to geopackage in the geocoded field make sure open output file is ticked run the tool result no crs for the layer one can assign epsg to fix it versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version compiled against geos capi running against geos capi sqlite version pdal version postgresql client version ubuntu spatialite version qwt version version os version ubuntu lts         active python plugins quickosm quick map services processing sagaprovider metasearch grassprovider db manager supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
98,804
4,031,464,799
IssuesEvent
2016-05-18 17:15:48
giantotter/giantotter_public
https://api.github.com/repos/giantotter/giantotter_public
opened
Bot can execute actions on behalf of other NPCs
backend: AI priority: A
Sometimes the action of the bot or the user results in another NPC taking an action. The slackbot should execute these actions for other NPCs. For example, the chef prepares a steak after asking him for steak.
1.0
Bot can execute actions on behalf of other NPCs - Sometimes the action of the bot or the user results in another NPC taking an action. The slackbot should execute these actions for other NPCs. For example, the chef prepares a steak after asking him for steak.
non_process
bot can execute actions on behalf of other npcs sometimes the action of the bot or the user results in another npc taking an action the slackbot should execute these actions for other npcs for example the chef prepares a steak after asking him for steak
0
300,853
25,999,830,020
IssuesEvent
2022-12-20 14:30:57
keycloak/keycloak
https://api.github.com/repos/keycloak/keycloak
closed
Flaky test: org.keycloak.testsuite.federation.kerberos.KerberosLdapTest#validatePasswordPolicyTest
kind/bug area/ci flaky-test
## org.keycloak.testsuite.federation.kerberos.KerberosLdapTest#validatePasswordPolicyTest [Run (pull_request)](https://github.com/keycloak/keycloak/actions/runs/3740969966) ### Errors
1.0
Flaky test: org.keycloak.testsuite.federation.kerberos.KerberosLdapTest#validatePasswordPolicyTest - ## org.keycloak.testsuite.federation.kerberos.KerberosLdapTest#validatePasswordPolicyTest [Run (pull_request)](https://github.com/keycloak/keycloak/actions/runs/3740969966) ### Errors
non_process
flaky test org keycloak testsuite federation kerberos kerberosldaptest validatepasswordpolicytest org keycloak testsuite federation kerberos kerberosldaptest validatepasswordpolicytest errors
0
814,813
30,522,967,310
IssuesEvent
2023-07-19 09:18:40
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID: 322061] Unchecked return value in drivers/input/input_xpt2046.c
bug priority: medium Coverity area: Input
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/18636af6d6658cf66e5f055901ffe2515116d329/drivers/input/input_xpt2046.c Category: Error handling issues Function: `xpt2046_work_handler` Component: Drivers CID: [322061](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=322061) Details: https://github.com/zephyrproject-rtos/zephyr/blob/18636af6d6658cf66e5f055901ffe2515116d329/drivers/input/input_xpt2046.c#L183 Please fix or provide comments in coverity using the link: https://scan9.scan.coverity.com/reports.htm#v29271/p12996. For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_CHECKED_RETURN). ([CWE-252](http://cwe.mitre.org/data/definitions/252.html)) Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID: 322061] Unchecked return value in drivers/input/input_xpt2046.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/18636af6d6658cf66e5f055901ffe2515116d329/drivers/input/input_xpt2046.c Category: Error handling issues Function: `xpt2046_work_handler` Component: Drivers CID: [322061](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=322061) Details: https://github.com/zephyrproject-rtos/zephyr/blob/18636af6d6658cf66e5f055901ffe2515116d329/drivers/input/input_xpt2046.c#L183 Please fix or provide comments in coverity using the link: https://scan9.scan.coverity.com/reports.htm#v29271/p12996. For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_CHECKED_RETURN). ([CWE-252](http://cwe.mitre.org/data/definitions/252.html)) Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
non_process
unchecked return value in drivers input input c static code scan issues found in file category error handling issues function work handler component drivers cid details please fix or provide comments in coverity using the link for more information about the violation check the note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
0
154,538
5,920,827,545
IssuesEvent
2017-05-22 21:15:17
infolab-csail/WikipediaBase
https://api.github.com/repos/infolab-csail/WikipediaBase
opened
Consider what to do with "duplicate" attributes in duplicate infoboxes
priority/medium question
E.g., [USS Powhatan (ID-3013)](https://en.wikipedia.org/wiki/USS_Powhatan_(ID-3013)) has two instantiations of the `ship-career` infobox template. `get-attributes` returns two instances of the attribute, one per infobox: ``` $ telnet <host> 8023 (get-attributes "wikipedia-ship-career" "USS Powhatan (ID-3013)") (... (:code "SHIP-FATE" :has-value t :rendered "Fate") ... (:code "SHIP-FATE" :has-value t :rendered "Fate") ...) ``` If we ask after either (identical) "fate" attribute, we get both values: ``` $ telnet <host> 8023 (get "wikipedia-ship-career" "USS Powhatan (ID-3013)" (:CODE "SHIP-FATE" :HAS-VALUE T :RENDERED "Fate")) ((:html "interned by the United States, 1914; seized, 1917") (:html "Transferred to War Department for use by United States Army")) ``` This is not necessarily a problem (in fact, it Works For Me), but we should make a deliberate decision about whether this is correct and expected behavior. If so, we should document it—something along the lines of "Note that if the same attribute name appears in multiple infoboxes of the same template, then `get-attributes` will return a separate copy of the attribute for each occurrence" and use USS Powhatan ID-3013 as an example. If not, we should decide what the correct behavior should be and implement it. @alvaromorales, @michaelsilver, any comments?
1.0
Consider what to do with "duplicate" attributes in duplicate infoboxes - E.g., [USS Powhatan (ID-3013)](https://en.wikipedia.org/wiki/USS_Powhatan_(ID-3013)) has two instantiations of the `ship-career` infobox template. `get-attributes` returns two instances of the attribute, one per infobox: ``` $ telnet <host> 8023 (get-attributes "wikipedia-ship-career" "USS Powhatan (ID-3013)") (... (:code "SHIP-FATE" :has-value t :rendered "Fate") ... (:code "SHIP-FATE" :has-value t :rendered "Fate") ...) ``` If we ask after either (identical) "fate" attribute, we get both values: ``` $ telnet <host> 8023 (get "wikipedia-ship-career" "USS Powhatan (ID-3013)" (:CODE "SHIP-FATE" :HAS-VALUE T :RENDERED "Fate")) ((:html "interned by the United States, 1914; seized, 1917") (:html "Transferred to War Department for use by United States Army")) ``` This is not necessarily a problem (in fact, it Works For Me), but we should make a deliberate decision about whether this is correct and expected behavior. If so, we should document it—something along the lines of "Note that if the same attribute name appears in multiple infoboxes of the same template, then `get-attributes` will return a separate copy of the attribute for each occurrence" and use USS Powhatan ID-3013 as an example. If not, we should decide what the correct behavior should be and implement it. @alvaromorales, @michaelsilver, any comments?
non_process
consider what to do with duplicate attributes in duplicate infoboxes e g has two instantiations of the ship career infobox template get attributes returns two instances of the attribute one per infobox telnet get attributes wikipedia ship career uss powhatan id code ship fate has value t rendered fate code ship fate has value t rendered fate if we ask after either identical fate attribute we get both values telnet get wikipedia ship career uss powhatan id code ship fate has value t rendered fate html interned by the united states seized html transferred to war department for use by united states army this is not necessarily a problem in fact it works for me but we should make a deliberate decision about whether this is correct and expected behavior if so we should document it—something along the lines of note that if the same attribute name appears in multiple infoboxes of the same template then get attributes will return a separate copy of the attribute for each occurrence and use uss powhatan id as an example if not we should decide what the correct behavior should be and implement it alvaromorales michaelsilver any comments
0
147,350
23,203,221,973
IssuesEvent
2022-08-02 00:47:25
antrea-io/antrea
https://api.github.com/repos/antrea-io/antrea
closed
Containerized Network Function (CNF) use case support on Antrea
kind/design lifecycle/stale
**Describe what you are trying to solve** SD-WAN/Telco CNF (Containerized Network Function) support on Antrea. Establishing network function virtualization (network application data plane) as a micro service orchestration in K8s environment, requires the below functionalities to be supported by CNI plugin (vital to support some of the most common use cases related to SD-WAN and Telco applications on the edge). Below are the list of functionalities/feature sets to be supported Antrea to enable CNF uses cases on the edge. - Multiple interface support per POD: Antrea to create and configure more than 1 network interface on a node/POD context to support internal (to cluster) and external (outside the cluster) communications between pods within and outside cluster. - Multiple subnet support per POD: Handle multiple IP address support per POD (in addition to default subnet provided by K8s). - Network application service Chaining: Dynamic route management across virtual and external networks. - SRIOV Support: Support for VF interface configuration per POD. - Load balancing: load balance data plane traffic between multiple CNF instances. **Describe the solution you have in mind** We are in the early stage of defining a solution to address CNF/VNF uses cases with Antrea. This proposal topic was created to initiate conversation with Antrea community to check and understand, if there are any existing feature support which would help us to achieve CNF use cases with Antrea (or) we would need to implement feature to add CNF configuration support as per the use case needs. **Describe how your solution impacts user flows** N/A at its current state. **Describe the main design/architecture of your solution** N/A at its current state. **Alternative solutions that you considered** TBD **Test plan** CNF use case validation (SD-WAN traffic forwarding, Network service function chaining etc.,) **Additional context** N/A
1.0
Containerized Network Function (CNF) use case support on Antrea - **Describe what you are trying to solve** SD-WAN/Telco CNF (Containerized Network Function) support on Antrea. Establishing network function virtualization (network application data plane) as a micro service orchestration in K8s environment, requires the below functionalities to be supported by CNI plugin (vital to support some of the most common use cases related to SD-WAN and Telco applications on the edge). Below are the list of functionalities/feature sets to be supported Antrea to enable CNF uses cases on the edge. - Multiple interface support per POD: Antrea to create and configure more than 1 network interface on a node/POD context to support internal (to cluster) and external (outside the cluster) communications between pods within and outside cluster. - Multiple subnet support per POD: Handle multiple IP address support per POD (in addition to default subnet provided by K8s). - Network application service Chaining: Dynamic route management across virtual and external networks. - SRIOV Support: Support for VF interface configuration per POD. - Load balancing: load balance data plane traffic between multiple CNF instances. **Describe the solution you have in mind** We are in the early stage of defining a solution to address CNF/VNF uses cases with Antrea. This proposal topic was created to initiate conversation with Antrea community to check and understand, if there are any existing feature support which would help us to achieve CNF use cases with Antrea (or) we would need to implement feature to add CNF configuration support as per the use case needs. **Describe how your solution impacts user flows** N/A at its current state. **Describe the main design/architecture of your solution** N/A at its current state. **Alternative solutions that you considered** TBD **Test plan** CNF use case validation (SD-WAN traffic forwarding, Network service function chaining etc.,) **Additional context** N/A
non_process
containerized network function cnf use case support on antrea describe what you are trying to solve sd wan telco cnf containerized network function support on antrea establishing network function virtualization network application data plane as a micro service orchestration in environment requires the below functionalities to be supported by cni plugin vital to support some of the most common use cases related to sd wan and telco applications on the edge below are the list of functionalities feature sets to be supported antrea to enable cnf uses cases on the edge multiple interface support per pod antrea to create and configure more than network interface on a node pod context to support internal to cluster and external outside the cluster communications between pods within and outside cluster multiple subnet support per pod handle multiple ip address support per pod in addition to default subnet provided by network application service chaining dynamic route management across virtual and external networks sriov support support for vf interface configuration per pod load balancing load balance data plane traffic between multiple cnf instances describe the solution you have in mind we are in the early stage of defining a solution to address cnf vnf uses cases with antrea this proposal topic was created to initiate conversation with antrea community to check and understand if there are any existing feature support which would help us to achieve cnf use cases with antrea or we would need to implement feature to add cnf configuration support as per the use case needs describe how your solution impacts user flows n a at its current state describe the main design architecture of your solution n a at its current state alternative solutions that you considered tbd test plan cnf use case validation sd wan traffic forwarding network service function chaining etc additional context n a
0
29,008
2,712,810,752
IssuesEvent
2015-04-09 15:45:08
mavoine/tarsius
https://api.github.com/repos/mavoine/tarsius
closed
time distribution graph
auto-migrated Priority-Medium Type-Enhancement
``` a time distribution graph like that of F-Spot ``` Original issue reported on code.google.com by `avoin...@gmail.com` on 11 Dec 2009 at 6:23
1.0
time distribution graph - ``` a time distribution graph like that of F-Spot ``` Original issue reported on code.google.com by `avoin...@gmail.com` on 11 Dec 2009 at 6:23
non_process
time distribution graph a time distribution graph like that of f spot original issue reported on code google com by avoin gmail com on dec at
0
48,186
10,220,389,841
IssuesEvent
2019-08-15 21:11:46
EdenServer/community
https://api.github.com/repos/EdenServer/community
closed
NPC Rondipur N. San d'Oria - displaying quest dialogue when he shouldn't
in-code-review
### Checklist <!-- Don't edit or delete this section, but tick the boxes after you have submitted your issue. If there are unticked boxes a developer may not address the issue. Make sure you comply with the checklist and then start writing in the details section below. --> - [x] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.) - [x] I have provided reproducable steps. (No "as the title says" posts please. Provide reproducable steps even if it seems like a waste of time.) - [x] I have provided my client version in the details. (type /ver into your game window) ### Details ver. 30181205_0 Rondipur currently uses dialogue from the quest 'Escort for Hire (San d'Oria)' when the player should not have access to it. Talk to him without having rank 6 reputation and he'll use the quest dialogue. It's only the text that triggers and the quest stays untriggered, as it should.
1.0
NPC Rondipur N. San d'Oria - displaying quest dialogue when he shouldn't - ### Checklist <!-- Don't edit or delete this section, but tick the boxes after you have submitted your issue. If there are unticked boxes a developer may not address the issue. Make sure you comply with the checklist and then start writing in the details section below. --> - [x] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.) - [x] I have provided reproducable steps. (No "as the title says" posts please. Provide reproducable steps even if it seems like a waste of time.) - [x] I have provided my client version in the details. (type /ver into your game window) ### Details ver. 30181205_0 Rondipur currently uses dialogue from the quest 'Escort for Hire (San d'Oria)' when the player should not have access to it. Talk to him without having rank 6 reputation and he'll use the quest dialogue. It's only the text that triggers and the quest stays untriggered, as it should.
non_process
npc rondipur n san d oria displaying quest dialogue when he shouldn t checklist don t edit or delete this section but tick the boxes after you have submitted your issue if there are unticked boxes a developer may not address the issue make sure you comply with the checklist and then start writing in the details section below i have searched for existing issues for issues like this one the issue has not been posted duplicate reports slow down development i have provided reproducable steps no as the title says posts please provide reproducable steps even if it seems like a waste of time i have provided my client version in the details type ver into your game window details ver rondipur currently uses dialogue from the quest escort for hire san d oria when the player should not have access to it talk to him without having rank reputation and he ll use the quest dialogue it s only the text that triggers and the quest stays untriggered as it should
0
19,000
24,995,253,603
IssuesEvent
2022-11-02 23:08:08
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Update OS packages in image
enhancement security process
### Problem The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume. ### Solution * Update base image for all mirror node components * Update postgresql chart and postgres-repmgr image * Add an extra command to update all OS packages so they at least are up to date as of each tag ``` RUN apt-get update \ && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends \ && rm -rf /var/lib/apt/lists/* ``` or ``` RUN apk update && apk upgrade && rm -rf /var/cache/apk/* ``` ### Alternatives _No response_
1.0
Update OS packages in image - ### Problem The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume. ### Solution * Update base image for all mirror node components * Update postgresql chart and postgres-repmgr image * Add an extra command to update all OS packages so they at least are up to date as of each tag ``` RUN apt-get update \ && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends \ && rm -rf /var/lib/apt/lists/* ``` or ``` RUN apk update && apk upgrade && rm -rf /var/cache/apk/* ``` ### Alternatives _No response_
process
update os packages in image problem the container images frequently contains out of date packages from the base image that may contain vulnerabilities we should keep them up to date even if there s not an updated base image to consume solution update base image for all mirror node components update postgresql chart and postgres repmgr image add an extra command to update all os packages so they at least are up to date as of each tag run apt get update debian frontend noninteractive apt get install y no install recommends rm rf var lib apt lists or run apk update apk upgrade rm rf var cache apk alternatives no response
1
36,534
9,819,930,580
IssuesEvent
2019-06-14 00:02:51
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Can't import tensorflow
stat:awaiting response subtype:windows type:build/install
**System information** - OS Platform and Distribution (Windows 7): - TensorFlow installed from (source or binary): - TensorFlow version: 1.13.1 - Python version: 3.6 - Installed using virtualenv? installed using pip - Bazel version (if compiling from source): - GCC/Compiler version (if compiling from source): - CUDA/cuDNN version: 9.0 - GPU model and memory: Geforce GTX 550 Ti 2811mb **Hello everyone. I just started to learn machine learning with python, and when I try to import tensorflow, I keep getting the same error:** Python 3.6.8 |Anaconda, Inc.| (default, Feb 21 2019, 18:30:04) [MSC v.1916 64 bit (AMD64)] on win32 runfile('D:/Users/Lucas/PycharmProjects/TensorENV/test.py', wdir='D:/Users/Lucas/PycharmProjects/TensorENV') Traceback (most recent call last): File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 58, in <module> from tensorflow.python.pywrap_tensorflow_internal import * File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module> _pywrap_tensorflow_internal = swig_import_helper() File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_helper _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, description) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 243, in load_module return load_dynamic(name, filename, file) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 343, in load_dynamic return _load(spec) ImportError: DLL load failed with error code -1073741795 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<input>", line 1, in <module> File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_umd.py", line 197, in runfile pydev_imports.execfile(filename, global_vars, local_vars) # execute the script File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile exec(compile(contents+"\n", file, 'exec'), glob, loc) File "D:/Users/Lucas/PycharmProjects/TensorENV/test.py", line 1, in <module> import tensorflow File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\__init__.py", line 24, in <module> from tensorflow.python import pywrap_tensorflow # pylint: disable=unused-import File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\__init__.py", line 49, in <module> from tensorflow.python import pywrap_tensorflow File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 74, in <module> raise ImportError(msg) ImportError: Traceback (most recent call last): File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 58, in <module> from tensorflow.python.pywrap_tensorflow_internal import * File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module> _pywrap_tensorflow_internal = swig_import_helper() File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_helper _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, description) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 243, in load_module return load_dynamic(name, filename, file) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 343, in load_dynamic return _load(spec) ImportError: DLL load failed with error code -1073741795 Failed to load the native TensorFlow runtime.
1.0
Can't import tensorflow - **System information** - OS Platform and Distribution (Windows 7): - TensorFlow installed from (source or binary): - TensorFlow version: 1.13.1 - Python version: 3.6 - Installed using virtualenv? installed using pip - Bazel version (if compiling from source): - GCC/Compiler version (if compiling from source): - CUDA/cuDNN version: 9.0 - GPU model and memory: Geforce GTX 550 Ti 2811mb **Hello everyone. I just started to learn machine learning with python, and when I try to import tensorflow, I keep getting the same error:** Python 3.6.8 |Anaconda, Inc.| (default, Feb 21 2019, 18:30:04) [MSC v.1916 64 bit (AMD64)] on win32 runfile('D:/Users/Lucas/PycharmProjects/TensorENV/test.py', wdir='D:/Users/Lucas/PycharmProjects/TensorENV') Traceback (most recent call last): File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 58, in <module> from tensorflow.python.pywrap_tensorflow_internal import * File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module> _pywrap_tensorflow_internal = swig_import_helper() File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_helper _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, description) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 243, in load_module return load_dynamic(name, filename, file) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 343, in load_dynamic return _load(spec) ImportError: DLL load failed with error code -1073741795 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<input>", line 1, in <module> File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_umd.py", line 197, in runfile pydev_imports.execfile(filename, global_vars, local_vars) # execute the script File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile exec(compile(contents+"\n", file, 'exec'), glob, loc) File "D:/Users/Lucas/PycharmProjects/TensorENV/test.py", line 1, in <module> import tensorflow File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\__init__.py", line 24, in <module> from tensorflow.python import pywrap_tensorflow # pylint: disable=unused-import File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\__init__.py", line 49, in <module> from tensorflow.python import pywrap_tensorflow File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 74, in <module> raise ImportError(msg) ImportError: Traceback (most recent call last): File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 58, in <module> from tensorflow.python.pywrap_tensorflow_internal import * File "D:\Program Files\JetBrains\PyCharm Community Edition 2019.1.3\helpers\pydev\_pydev_bundle\pydev_import_hook.py", line 21, in do_import module = self._system_import(name, *args, **kwargs) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module> _pywrap_tensorflow_internal = swig_import_helper() File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_helper _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, description) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 243, in load_module return load_dynamic(name, filename, file) File "D:\Users\Lucas\Anaconda3\envs\tensor\lib\imp.py", line 343, in load_dynamic return _load(spec) ImportError: DLL load failed with error code -1073741795 Failed to load the native TensorFlow runtime.
non_process
can t import tensorflow system information os platform and distribution windows tensorflow installed from source or binary tensorflow version python version installed using virtualenv installed using pip bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory geforce gtx ti hello everyone i just started to learn machine learning with python and when i try to import tensorflow i keep getting the same error python anaconda inc default feb on runfile d users lucas pycharmprojects tensorenv test py wdir d users lucas pycharmprojects tensorenv traceback most recent call last file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow py line in from tensorflow python pywrap tensorflow internal import file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev import hook py line in do import module self system import name args kwargs file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow internal py line in pywrap tensorflow internal swig import helper file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow internal py line in swig import helper mod imp load module pywrap tensorflow internal fp pathname description file d users lucas envs tensor lib imp py line in load module return load dynamic name filename file file d users lucas envs tensor lib imp py line in load dynamic return load spec importerror dll load failed with error code during handling of the above exception another exception occurred traceback most recent call last file line in file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev umd py line in runfile pydev imports execfile filename global vars local vars execute the script file d program files jetbrains pycharm community edition helpers pydev pydev imps pydev execfile py line in execfile exec compile contents n file exec glob loc file d users lucas pycharmprojects tensorenv test py line in import tensorflow file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev import hook py line in do import module self system import name args kwargs file d users lucas envs tensor lib site packages tensorflow init py line in from tensorflow python import pywrap tensorflow pylint disable unused import file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev import hook py line in do import module self system import name args kwargs file d users lucas envs tensor lib site packages tensorflow python init py line in from tensorflow python import pywrap tensorflow file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev import hook py line in do import module self system import name args kwargs file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow py line in raise importerror msg importerror traceback most recent call last file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow py line in from tensorflow python pywrap tensorflow internal import file d program files jetbrains pycharm community edition helpers pydev pydev bundle pydev import hook py line in do import module self system import name args kwargs file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow internal py line in pywrap tensorflow internal swig import helper file d users lucas envs tensor lib site packages tensorflow python pywrap tensorflow internal py line in swig import helper mod imp load module pywrap tensorflow internal fp pathname description file d users lucas envs tensor lib imp py line in load module return load dynamic name filename file file d users lucas envs tensor lib imp py line in load dynamic return load spec importerror dll load failed with error code failed to load the native tensorflow runtime
0
334,684
24,431,315,513
IssuesEvent
2022-10-06 08:20:43
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
opened
[Docs]: aws_identitystore_group example
documentation needs-triage
### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/identitystore_group ### Description The attribute `identity_store_id ` must be set with an id and not the arn of the identity store. ``` resource "aws_identitystore_group" "this" { display_name = "Example group" description = "Example description" identity_store_id = tolist(data.aws_ssoadmin_instances.example.identity_store_ids)[0] } ``` ### References https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ssoadmin_instances ### Would you like to implement a fix? Yes
1.0
[Docs]: aws_identitystore_group example - ### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/identitystore_group ### Description The attribute `identity_store_id ` must be set with an id and not the arn of the identity store. ``` resource "aws_identitystore_group" "this" { display_name = "Example group" description = "Example description" identity_store_id = tolist(data.aws_ssoadmin_instances.example.identity_store_ids)[0] } ``` ### References https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ssoadmin_instances ### Would you like to implement a fix? Yes
non_process
aws identitystore group example documentation link description the attribute identity store id must be set with an id and not the arn of the identity store resource aws identitystore group this display name example group description example description identity store id tolist data aws ssoadmin instances example identity store ids references would you like to implement a fix yes
0
19,082
25,127,778,003
IssuesEvent
2022-11-09 13:05:16
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Rethink PartialEq implementation for sql-schema-describer::ColumnTypeFamily
process/candidate kind/tech team/schema
It compares IDs, which we absolutely do not want in the migration engine differ, but is harmless everywhere else. Should we break up ColumnTypeFamily variants into scalar and enum and unsupported variants? Remove the PartialEq impl? https://github.com/prisma/prisma-engines/pull/3372
1.0
Rethink PartialEq implementation for sql-schema-describer::ColumnTypeFamily - It compares IDs, which we absolutely do not want in the migration engine differ, but is harmless everywhere else. Should we break up ColumnTypeFamily variants into scalar and enum and unsupported variants? Remove the PartialEq impl? https://github.com/prisma/prisma-engines/pull/3372
process
rethink partialeq implementation for sql schema describer columntypefamily it compares ids which we absolutely do not want in the migration engine differ but is harmless everywhere else should we break up columntypefamily variants into scalar and enum and unsupported variants remove the partialeq impl
1
16,853
22,112,506,769
IssuesEvent
2022-06-01 22:47:22
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
infra: bot template/config for running on Cloud Run
type: process
In the future, we will want extra dependencies available to our bot. We should still be able to run our probot-based bots easily with our credentials and tasks shims in `gcf-utils` within a Docker container on Cloud Run. - [x] Figure out how to run in Cloud Build - [x] Configure canary-bot to run on Cloud Build - [x] Configure scheduler-proxy to target a Cloud Build service - [x] Configure gcf-utils to queue tasks to target the Cloud Build service - [x] Configure generate-bot to templatize deployment for a new bot
1.0
infra: bot template/config for running on Cloud Run - In the future, we will want extra dependencies available to our bot. We should still be able to run our probot-based bots easily with our credentials and tasks shims in `gcf-utils` within a Docker container on Cloud Run. - [x] Figure out how to run in Cloud Build - [x] Configure canary-bot to run on Cloud Build - [x] Configure scheduler-proxy to target a Cloud Build service - [x] Configure gcf-utils to queue tasks to target the Cloud Build service - [x] Configure generate-bot to templatize deployment for a new bot
process
infra bot template config for running on cloud run in the future we will want extra dependencies available to our bot we should still be able to run our probot based bots easily with our credentials and tasks shims in gcf utils within a docker container on cloud run figure out how to run in cloud build configure canary bot to run on cloud build configure scheduler proxy to target a cloud build service configure gcf utils to queue tasks to target the cloud build service configure generate bot to templatize deployment for a new bot
1
226,718
18,043,955,427
IssuesEvent
2021-09-18 14:55:43
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit
Test_9999 logicmoo.base.examples.fol unit_test SANITY_EXISTS_03
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- module(t123). :- dynamic(t123:ttExpressionType/1). % :- process_this_script. % :- process_this_script. :- statistics. /*~ % Started at Sat Sep 18 07:55:40 2021 % 0.875 seconds cpu time for 412,769 inferences % 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes % % Limit Allocated In use % Local stack: - 52 Kb 4,216 b % Global stack: - 64 Kb 49 Kb % Trail stack: - 34 Kb 520 b % Total: 1,024 Mb 150 Kb 53 Kb % % 5 garbage collections gained 305,208 bytes in 0.000 seconds. % 1 atom garbage collections gained 1,137 atoms in 0.015 seconds. % 4 clause garbage collections gained 118 clauses in 0.000 seconds. % Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds % 3 threads, 0 finished threads used 0.000 seconds ~*/ subtest_assert(I):-kif_assert(I). subtest([subtest_assert(tAnimal(joe)), mpred_test(isa(_,tHeart))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( tHeart(Heart4))])). ============================================ ~*/ subtest([subtest_assert(tAnimal(joe)), mpred_test(hasOrgan(joe,_))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). ============================================ ~*/ subtest([subtest_assert(tHeart(_)), mpred_test(~hasOrgan(jack,_))]). %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]))))) %~ kifi = subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]). %~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) %~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tHeart(Heart7)), mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). ============================================ ~*/ dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl. %~ warn( db_expand_0( clause(unknown,cuz), %~ dbanner :- %~ nl , %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl , %~ nl, %~ dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)) %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))]]) %~ unused(save_junit_results) ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
2.0
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- module(t123). :- dynamic(t123:ttExpressionType/1). % :- process_this_script. % :- process_this_script. :- statistics. /*~ % Started at Sat Sep 18 07:55:40 2021 % 0.875 seconds cpu time for 412,769 inferences % 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes % % Limit Allocated In use % Local stack: - 52 Kb 4,216 b % Global stack: - 64 Kb 49 Kb % Trail stack: - 34 Kb 520 b % Total: 1,024 Mb 150 Kb 53 Kb % % 5 garbage collections gained 305,208 bytes in 0.000 seconds. % 1 atom garbage collections gained 1,137 atoms in 0.015 seconds. % 4 clause garbage collections gained 118 clauses in 0.000 seconds. % Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds % 3 threads, 0 finished threads used 0.000 seconds ~*/ subtest_assert(I):-kif_assert(I). subtest([subtest_assert(tAnimal(joe)), mpred_test(isa(_,tHeart))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( tHeart(Heart4))])). ============================================ ~*/ subtest([subtest_assert(tAnimal(joe)), mpred_test(hasOrgan(joe,_))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). ============================================ ~*/ subtest([subtest_assert(tHeart(_)), mpred_test(~hasOrgan(jack,_))]). %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]))))) %~ kifi = subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]). %~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) %~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tHeart(Heart7)), mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). ============================================ ~*/ dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl. %~ warn( db_expand_0( clause(unknown,cuz), %~ dbanner :- %~ nl , %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl , %~ nl, %~ dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)) %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))]]) %~ unused(save_junit_results) ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
non_process
logicmoo base examples fol sanity exists junit cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol timeout foreground preserve status s sigkill k lmoo clif sanity exists pfc pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl this test might need use module library logicmoo plarkc module dynamic ttexpressiontype process this script process this script statistics started at sat sep seconds cpu time for inferences atoms functors predicates modules vm codes limit allocated in use local stack kb b global stack kb kb trail stack kb b total mb kb kb garbage collections gained bytes in seconds atom garbage collections gained atoms in seconds clause garbage collections gained clauses in seconds stack shifts local global trail in seconds threads finished threads used seconds subtest assert i kif assert i subtest subtest assert tanimal joe mpred test isa theart var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test theart kifi subtest subtest assert tanimal joe mpred test theart kifm nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test theart var kif to boxlog subtest in english joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test theart var results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test hasorgan joe var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifi subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifm nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan kif to boxlog subtest in english joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert theart mpred test hasorgan jack debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert theart mpred test hasorgan jack hasorgan kifi subtest subtest assert theart mpred test hasorgan jack hasorgan kifm nesc subtest subtest assert theart mpred test hasorgan jack hasorgan subtest subtest assert theart var mpred test hasorgan jack var hasorgan kif to boxlog subtest in english isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert theart var mpred test not hasorgan jack var hasorgan results in the following entailment s nesc subtest it is necessarily true that isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest nesc subtest subtest assert theart mpred test hasorgan jack hasorgan dbanner nl nl dmsg nl nl warn db expand clause unknown cuz dbanner nl nl dmsg nl nl dbanner nl dmsg nl message hook type error message hook initialization exception must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax error prolog initialisation failed nl unknown message p must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax unused save junit results totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k sanity exists pfc pl returned
0
12,671
15,038,812,360
IssuesEvent
2021-02-02 17:53:15
w3c/aria-at
https://api.github.com/repos/w3c/aria-at
opened
Proposed Patterns for Test Plan Development (Q1/Q2 2021)
Agenda+ process tests
Issue #318 outlines the initial 13 patterns to be worked on by members of the PAC team. We are approaching the completion point on those patterns (status outlined below), and want to take input from the community group on which patterns to tackle next. #### Current Status * Pull requests have been submitted for 8 of the 13 patterns. * Test plans for two patterns have been prepared but are undergoing internal PAC review: Tabs With Manual Activation and Minimal Data Grid. * The test plans for two patterns are currently being written: Radio Group Using Roving tabindex and Radio Group Using aria-activedescendant. * One pattern has been intentionally skipped for now: Date Picker Combobox. It's thought that this could benefit from a more compositional approach to test development.
1.0
Proposed Patterns for Test Plan Development (Q1/Q2 2021) - Issue #318 outlines the initial 13 patterns to be worked on by members of the PAC team. We are approaching the completion point on those patterns (status outlined below), and want to take input from the community group on which patterns to tackle next. #### Current Status * Pull requests have been submitted for 8 of the 13 patterns. * Test plans for two patterns have been prepared but are undergoing internal PAC review: Tabs With Manual Activation and Minimal Data Grid. * The test plans for two patterns are currently being written: Radio Group Using Roving tabindex and Radio Group Using aria-activedescendant. * One pattern has been intentionally skipped for now: Date Picker Combobox. It's thought that this could benefit from a more compositional approach to test development.
process
proposed patterns for test plan development issue outlines the initial patterns to be worked on by members of the pac team we are approaching the completion point on those patterns status outlined below and want to take input from the community group on which patterns to tackle next current status pull requests have been submitted for of the patterns test plans for two patterns have been prepared but are undergoing internal pac review tabs with manual activation and minimal data grid the test plans for two patterns are currently being written radio group using roving tabindex and radio group using aria activedescendant one pattern has been intentionally skipped for now date picker combobox it s thought that this could benefit from a more compositional approach to test development
1
5,853
8,679,088,597
IssuesEvent
2018-11-30 22:15:03
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
[Firestore] Query start/end at/after accepts snapshots, doesn't properly apply implicit ordering.
api: firestore triaged for GA type: process
Query.start_at()/start_after()/end_at()/end_after() accepts DocumentSnapshots, but doesn’t properly apply the implicit ordering.
1.0
[Firestore] Query start/end at/after accepts snapshots, doesn't properly apply implicit ordering. - Query.start_at()/start_after()/end_at()/end_after() accepts DocumentSnapshots, but doesn’t properly apply the implicit ordering.
process
query start end at after accepts snapshots doesn t properly apply implicit ordering query start at start after end at end after accepts documentsnapshots but doesn’t properly apply the implicit ordering
1
21,038
27,979,267,458
IssuesEvent
2023-03-26 00:20:53
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Manage border for non-convex and self-intersecting path mask
feature: enhancement difficulty: hard scope: UI scope: image processing no-issue-activity
Just a remainder in case someone want to tackle this difficult task. Currently the algorithm for drawing path mask doesn't manage properly 1. self-intersencting paths: it fails to recognize the outside side of the path ![immagine](https://user-images.githubusercontent.com/43290988/139692278-11b1abc1-390a-41ff-8d29-29a89916da69.png) 2. concave paths, when the border control handle of an "inner" point is dragged away ![immagine](https://user-images.githubusercontent.com/43290988/139694096-53ba2d25-0d56-40a2-b5ab-4d7f64c33619.png)
1.0
Manage border for non-convex and self-intersecting path mask - Just a remainder in case someone want to tackle this difficult task. Currently the algorithm for drawing path mask doesn't manage properly 1. self-intersencting paths: it fails to recognize the outside side of the path ![immagine](https://user-images.githubusercontent.com/43290988/139692278-11b1abc1-390a-41ff-8d29-29a89916da69.png) 2. concave paths, when the border control handle of an "inner" point is dragged away ![immagine](https://user-images.githubusercontent.com/43290988/139694096-53ba2d25-0d56-40a2-b5ab-4d7f64c33619.png)
process
manage border for non convex and self intersecting path mask just a remainder in case someone want to tackle this difficult task currently the algorithm for drawing path mask doesn t manage properly self intersencting paths it fails to recognize the outside side of the path concave paths when the border control handle of an inner point is dragged away
1
2,560
5,315,719,707
IssuesEvent
2017-02-13 18:01:38
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
[sct_process_segmentation] Error: Extract centerline
bug priority: high sct_process_segmentation
Data: ``` /Volumes/data_shared/sct_testing/large/amu_G1_3/t2s/t2s_seg_manual.nii.gz ``` Syntax: ``` sct_process_segmentation -i /Volumes/data_shared/sct_testing/large/amu_G1_3/t2s/t2s_seg_manual.nii.gz -p centerline ``` Error: ``` Check parameters: .. segmentation file: /Users/chgroc/data/data_sdika_err/twh_e23779_t2s_seg.nii.gz Create temporary folder... mkdir tmp.170123163935_26341/ Copying data to tmp folder... sct_convert -i /Users/chgroc/data/data_sdika_err/twh_e23779_t2s_seg.nii.gz -o tmp.170123163935_26341/segmentation.nii.gz Orient centerline to RPI orientation... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz Open segmentation volume... Get data dimensions... .. matrix size: 512 x 512 x 12 .. voxel size: 0.3906mm x 0.3906mm x 8.00003mm Smooth centerline/segmentation... .. Get center of mass of the centerline/segmentation... .. Smoothing algo = nurbs Fitting centerline using B-spline approximation... Test: # of control points = 5 Error on approximation = 4.72 mm Test: # of control points = 6 WARNING: NURBS instability -> wrong reconstruction Test: # of control points = 7 Error on approximation = 36.15 mm Test: # of control points = 8 WARNING: NURBS instability -> wrong control points Test: # of control points = 9 singular matrix Test: # of control points = 10 singular matrix Test: # of control points = 11 singular matrix The fitting of the curve was done using 5 control points: the number that gave the best results. Error on approximation = 4.72 mm Traceback (most recent call last): File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 1128, in <module> main(sys.argv[1:]) File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 226, in main fname_output = extract_centerline(fname_segmentation, remove_temp_files, verbose=param.verbose, algo_fitting=param.algo_fitting) File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 467, in extract_centerline data[int(round(x_centerline_fit[iz-min_z_index])), int(round(y_centerline_fit[iz-min_z_index])), int(iz)] = 1 # if index is out of bounds here for hanning: either the segmentation has holes or labels have been added to the file IndexError: index 12 is out of bounds for axis 2 with size 12 ```
1.0
[sct_process_segmentation] Error: Extract centerline - Data: ``` /Volumes/data_shared/sct_testing/large/amu_G1_3/t2s/t2s_seg_manual.nii.gz ``` Syntax: ``` sct_process_segmentation -i /Volumes/data_shared/sct_testing/large/amu_G1_3/t2s/t2s_seg_manual.nii.gz -p centerline ``` Error: ``` Check parameters: .. segmentation file: /Users/chgroc/data/data_sdika_err/twh_e23779_t2s_seg.nii.gz Create temporary folder... mkdir tmp.170123163935_26341/ Copying data to tmp folder... sct_convert -i /Users/chgroc/data/data_sdika_err/twh_e23779_t2s_seg.nii.gz -o tmp.170123163935_26341/segmentation.nii.gz Orient centerline to RPI orientation... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz Open segmentation volume... Get data dimensions... .. matrix size: 512 x 512 x 12 .. voxel size: 0.3906mm x 0.3906mm x 8.00003mm Smooth centerline/segmentation... .. Get center of mass of the centerline/segmentation... .. Smoothing algo = nurbs Fitting centerline using B-spline approximation... Test: # of control points = 5 Error on approximation = 4.72 mm Test: # of control points = 6 WARNING: NURBS instability -> wrong reconstruction Test: # of control points = 7 Error on approximation = 36.15 mm Test: # of control points = 8 WARNING: NURBS instability -> wrong control points Test: # of control points = 9 singular matrix Test: # of control points = 10 singular matrix Test: # of control points = 11 singular matrix The fitting of the curve was done using 5 control points: the number that gave the best results. Error on approximation = 4.72 mm Traceback (most recent call last): File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 1128, in <module> main(sys.argv[1:]) File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 226, in main fname_output = extract_centerline(fname_segmentation, remove_temp_files, verbose=param.verbose, algo_fitting=param.algo_fitting) File "/Users/chgroc/spinalcordtoolbox/scripts/sct_process_segmentation.py", line 467, in extract_centerline data[int(round(x_centerline_fit[iz-min_z_index])), int(round(y_centerline_fit[iz-min_z_index])), int(iz)] = 1 # if index is out of bounds here for hanning: either the segmentation has holes or labels have been added to the file IndexError: index 12 is out of bounds for axis 2 with size 12 ```
process
error extract centerline data volumes data shared sct testing large amu seg manual nii gz syntax sct process segmentation i volumes data shared sct testing large amu seg manual nii gz p centerline error check parameters segmentation file users chgroc data data sdika err twh seg nii gz create temporary folder mkdir tmp copying data to tmp folder sct convert i users chgroc data data sdika err twh seg nii gz o tmp segmentation nii gz orient centerline to rpi orientation sct image i segmentation nii gz setorient rpi o segmentation rpi nii gz open segmentation volume get data dimensions matrix size x x voxel size x x smooth centerline segmentation get center of mass of the centerline segmentation smoothing algo nurbs fitting centerline using b spline approximation test of control points error on approximation mm test of control points warning nurbs instability wrong reconstruction test of control points error on approximation mm test of control points warning nurbs instability wrong control points test of control points singular matrix test of control points singular matrix test of control points singular matrix the fitting of the curve was done using control points the number that gave the best results error on approximation mm traceback most recent call last file users chgroc spinalcordtoolbox scripts sct process segmentation py line in main sys argv file users chgroc spinalcordtoolbox scripts sct process segmentation py line in main fname output extract centerline fname segmentation remove temp files verbose param verbose algo fitting param algo fitting file users chgroc spinalcordtoolbox scripts sct process segmentation py line in extract centerline data int round y centerline fit int iz if index is out of bounds here for hanning either the segmentation has holes or labels have been added to the file indexerror index is out of bounds for axis with size
1
15,578
19,703,728,352
IssuesEvent
2022-01-12 19:23:15
googleapis/nodejs-kms
https://api.github.com/repos/googleapis/nodejs-kms
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'kms' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'kms' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname kms invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
114,850
24,678,936,403
IssuesEvent
2022-10-18 19:28:21
bnreplah/reactvulna
https://api.github.com/repos/bnreplah/reactvulna
opened
Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS) [VID:80:src/Users.js:100]
VeracodeFlaw: Medium Veracode Pipeline Scan
**Filename:** src/Users.js **Line:** 100 **CWE:** 80 (Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS)) <span>This call to jQueryResult.replaceWith() contains a cross-site scripting (XSS) flaw. The application populates the HTTP response with untrusted input, allowing an attacker to embed malicious content, such as Javascript code, which will be executed in the context of the victim's browser. XSS vulnerabilities are commonly exploited to steal or manipulate cookies, modify presentation of content, and compromise confidential information, with new attack vectors being discovered on a regular basis. </span> <span>Use contextual escaping on all untrusted data before using it to construct any portion of an HTTP response. The escaping method should be chosen based on the specific use case of the untrusted data, otherwise it may not protect fully against the attack. For example, if the data is being written to the body of an HTML page, use HTML entity escaping; if the data is being written to an attribute, use attribute escaping; etc. Both the OWASP Java Encoder library and the Microsoft AntiXSS library provide contextual escaping methods. For more details on contextual escaping, see https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Cross_Site_Scripting_Prevention_Cheat_Sheet.md. In addition, as a best practice, always validate untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/79.html">CWE</a> <a href="https://owasp.org/www-community/attacks/xss/">OWASP</a> <a href="https://docs.veracode.com/r/review_cleansers">Supported Cleansers</a></span>
2.0
Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS) [VID:80:src/Users.js:100] - **Filename:** src/Users.js **Line:** 100 **CWE:** 80 (Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS)) <span>This call to jQueryResult.replaceWith() contains a cross-site scripting (XSS) flaw. The application populates the HTTP response with untrusted input, allowing an attacker to embed malicious content, such as Javascript code, which will be executed in the context of the victim's browser. XSS vulnerabilities are commonly exploited to steal or manipulate cookies, modify presentation of content, and compromise confidential information, with new attack vectors being discovered on a regular basis. </span> <span>Use contextual escaping on all untrusted data before using it to construct any portion of an HTTP response. The escaping method should be chosen based on the specific use case of the untrusted data, otherwise it may not protect fully against the attack. For example, if the data is being written to the body of an HTML page, use HTML entity escaping; if the data is being written to an attribute, use attribute escaping; etc. Both the OWASP Java Encoder library and the Microsoft AntiXSS library provide contextual escaping methods. For more details on contextual escaping, see https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Cross_Site_Scripting_Prevention_Cheat_Sheet.md. In addition, as a best practice, always validate untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/79.html">CWE</a> <a href="https://owasp.org/www-community/attacks/xss/">OWASP</a> <a href="https://docs.veracode.com/r/review_cleansers">Supported Cleansers</a></span>
non_process
improper neutralization of script related html tags in a web page basic xss filename src users js line cwe improper neutralization of script related html tags in a web page basic xss this call to jqueryresult replacewith contains a cross site scripting xss flaw the application populates the http response with untrusted input allowing an attacker to embed malicious content such as javascript code which will be executed in the context of the victim s browser xss vulnerabilities are commonly exploited to steal or manipulate cookies modify presentation of content and compromise confidential information with new attack vectors being discovered on a regular basis use contextual escaping on all untrusted data before using it to construct any portion of an http response the escaping method should be chosen based on the specific use case of the untrusted data otherwise it may not protect fully against the attack for example if the data is being written to the body of an html page use html entity escaping if the data is being written to an attribute use attribute escaping etc both the owasp java encoder library and the microsoft antixss library provide contextual escaping methods for more details on contextual escaping see in addition as a best practice always validate untrusted input to ensure that it conforms to the expected format using centralized data validation routines when possible references
0
249,241
18,858,174,858
IssuesEvent
2021-11-12 09:28:07
pss-coder/pe
https://api.github.com/repos/pss-coder/pe
opened
Storage component UI diagram Date and Frequency
severity.High type.DocumentationBug
Where does your Frequency and Date class comes from? From your code it seems that your Frequency class comes from the Model component, and your date comes from the internal Java class which isn't necessary to add in. ![Screenshot 2021-11-12 at 5.24.06 PM.png](https://raw.githubusercontent.com/pss-coder/pe/main/files/9b0d26e6-0bab-4a7d-8d12-6d3f6131651b.png) ![Screenshot 2021-11-12 at 5.27.36 PM.png](https://raw.githubusercontent.com/pss-coder/pe/main/files/a65a5a40-f539-4c03-a7a9-3ecf9d9ff080.png) <!--session: 1636704585877-8b82bb88-bedf-4d7a-ab66-f744764ccb3b--> <!--Version: Web v3.4.1-->
1.0
Storage component UI diagram Date and Frequency - Where does your Frequency and Date class comes from? From your code it seems that your Frequency class comes from the Model component, and your date comes from the internal Java class which isn't necessary to add in. ![Screenshot 2021-11-12 at 5.24.06 PM.png](https://raw.githubusercontent.com/pss-coder/pe/main/files/9b0d26e6-0bab-4a7d-8d12-6d3f6131651b.png) ![Screenshot 2021-11-12 at 5.27.36 PM.png](https://raw.githubusercontent.com/pss-coder/pe/main/files/a65a5a40-f539-4c03-a7a9-3ecf9d9ff080.png) <!--session: 1636704585877-8b82bb88-bedf-4d7a-ab66-f744764ccb3b--> <!--Version: Web v3.4.1-->
non_process
storage component ui diagram date and frequency where does your frequency and date class comes from from your code it seems that your frequency class comes from the model component and your date comes from the internal java class which isn t necessary to add in
0
15,028
18,740,586,385
IssuesEvent
2021-11-04 13:11:19
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Process] Cannot safely clear output
Bug Process Status: Needs Review Help wanted Stalled
**Symfony version(s) affected**: 4.3-dev **Description** When reading from a process producing rapid output, it's necessary to periodically clear the process output stream buffer to avoid memory exhaustion. Between consecutive calls to getIncrementalOutput() and clearOutput(), additional output may be asynchronously written by the process which would be unread and lost. The same issue occurs using getOutput(). **How to reproduce** The following code proves that some lines are lost between the adjacent get and clear calls: ``` $i = 0; $max = 1000000; $process = new Process("seq $i $max"); $process->start(); while ($process->isRunning()) { if (($out = $process->getOutput()) != '') { $process->clearOutput(); foreach (explode("\n", $out) as $line) { if ($line != $i) { echo "Skipped $i\n"; exit(); } ++$i; } } } ```
1.0
[Process] Cannot safely clear output - **Symfony version(s) affected**: 4.3-dev **Description** When reading from a process producing rapid output, it's necessary to periodically clear the process output stream buffer to avoid memory exhaustion. Between consecutive calls to getIncrementalOutput() and clearOutput(), additional output may be asynchronously written by the process which would be unread and lost. The same issue occurs using getOutput(). **How to reproduce** The following code proves that some lines are lost between the adjacent get and clear calls: ``` $i = 0; $max = 1000000; $process = new Process("seq $i $max"); $process->start(); while ($process->isRunning()) { if (($out = $process->getOutput()) != '') { $process->clearOutput(); foreach (explode("\n", $out) as $line) { if ($line != $i) { echo "Skipped $i\n"; exit(); } ++$i; } } } ```
process
cannot safely clear output symfony version s affected dev description when reading from a process producing rapid output it s necessary to periodically clear the process output stream buffer to avoid memory exhaustion between consecutive calls to getincrementaloutput and clearoutput additional output may be asynchronously written by the process which would be unread and lost the same issue occurs using getoutput how to reproduce the following code proves that some lines are lost between the adjacent get and clear calls i max process new process seq i max process start while process isrunning if out process getoutput process clearoutput foreach explode n out as line if line i echo skipped i n exit i
1
16,778
21,961,290,554
IssuesEvent
2022-05-24 16:02:23
google/ground-android
https://api.github.com/repos/google/ground-android
closed
[Cloud Build] Save gradle cache only if dependencies have been updated
type: process priority: p2
Currently, we zip and copy gradle cache to GCS with every successful trigger. This takes ~ 40 seconds and doesn't add any value if the dependencies haven't been updated in the current pull request. Ideas: * Generate hash (md5) for `build.gradle` files (both project and app level) * Run `Compress gradle build cache` and `Save gradle cache to GCS` steps only if the combined hash has been modified. * Add another step to save the new hashes to GCS
1.0
[Cloud Build] Save gradle cache only if dependencies have been updated - Currently, we zip and copy gradle cache to GCS with every successful trigger. This takes ~ 40 seconds and doesn't add any value if the dependencies haven't been updated in the current pull request. Ideas: * Generate hash (md5) for `build.gradle` files (both project and app level) * Run `Compress gradle build cache` and `Save gradle cache to GCS` steps only if the combined hash has been modified. * Add another step to save the new hashes to GCS
process
save gradle cache only if dependencies have been updated currently we zip and copy gradle cache to gcs with every successful trigger this takes seconds and doesn t add any value if the dependencies haven t been updated in the current pull request ideas generate hash for build gradle files both project and app level run compress gradle build cache and save gradle cache to gcs steps only if the combined hash has been modified add another step to save the new hashes to gcs
1
11,313
14,116,333,413
IssuesEvent
2020-11-08 02:23:44
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
closed
kubeflow-testing old GPU nodes
area/engprod kind/bug kind/process lifecycle/frozen lifecycle/stale priority/p2
Our kubeflow-testing cluster has a GPU node pool with 6 nodes. I don't think we have many tests actually using that. We should have autoscaling enabled. I suspect its not getting downscaled because K8s isn't smart enough to not schedule workloads on that node-pool.
1.0
kubeflow-testing old GPU nodes - Our kubeflow-testing cluster has a GPU node pool with 6 nodes. I don't think we have many tests actually using that. We should have autoscaling enabled. I suspect its not getting downscaled because K8s isn't smart enough to not schedule workloads on that node-pool.
process
kubeflow testing old gpu nodes our kubeflow testing cluster has a gpu node pool with nodes i don t think we have many tests actually using that we should have autoscaling enabled i suspect its not getting downscaled because isn t smart enough to not schedule workloads on that node pool
1
4,294
7,192,432,840
IssuesEvent
2018-02-03 03:31:59
amaster507/ifbmt
https://api.github.com/repos/amaster507/ifbmt
closed
Data Import
calendar idea importing data process
# There should be a way to import existing user data. Of course the easiest way to import data is from a CSV (excel) file and sanitizing the input and mapping the columns to the correct database fields. But, what other formats are data in that everyone is using that may need to be imported? Automatic importing may not be available for every case and we may need to implement a process to submit files to be imported and then someone knowledgeable with the database structure can format, sanitize, and import the data. Is this even something that is needed?
1.0
Data Import - # There should be a way to import existing user data. Of course the easiest way to import data is from a CSV (excel) file and sanitizing the input and mapping the columns to the correct database fields. But, what other formats are data in that everyone is using that may need to be imported? Automatic importing may not be available for every case and we may need to implement a process to submit files to be imported and then someone knowledgeable with the database structure can format, sanitize, and import the data. Is this even something that is needed?
process
data import there should be a way to import existing user data of course the easiest way to import data is from a csv excel file and sanitizing the input and mapping the columns to the correct database fields but what other formats are data in that everyone is using that may need to be imported automatic importing may not be available for every case and we may need to implement a process to submit files to be imported and then someone knowledgeable with the database structure can format sanitize and import the data is this even something that is needed
1
8,393
11,564,244,840
IssuesEvent
2020-02-20 08:13:33
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing : Perform SQL query on a PostGIS database, add the ability to select schema
Feature Request Processing
**Feature description.** Add the ability to select schema and not only the database name. **Additional context** See screenshot on a standard version : ![image](https://user-images.githubusercontent.com/5890416/74673374-59722180-51af-11ea-9cbc-5194568b7133.png) Add an other dropdown menu to select schema will be perfect.
1.0
Processing : Perform SQL query on a PostGIS database, add the ability to select schema - **Feature description.** Add the ability to select schema and not only the database name. **Additional context** See screenshot on a standard version : ![image](https://user-images.githubusercontent.com/5890416/74673374-59722180-51af-11ea-9cbc-5194568b7133.png) Add an other dropdown menu to select schema will be perfect.
process
processing perform sql query on a postgis database add the ability to select schema feature description add the ability to select schema and not only the database name additional context see screenshot on a standard version add an other dropdown menu to select schema will be perfect
1
8,484
11,645,065,464
IssuesEvent
2020-02-29 22:32:45
googleapis/nodejs-iot
https://api.github.com/repos/googleapis/nodejs-iot
closed
GA Release
api: cloudiot type: process
Package name: @google-cloud/bigquery-data-transfer Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
GA Release - Package name: @google-cloud/bigquery-data-transfer Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
ga release package name google cloud bigquery data transfer current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
18,883
24,823,858,655
IssuesEvent
2022-10-25 18:47:55
tradingstrategy-ai/frontend
https://api.github.com/repos/tradingstrategy-ai/frontend
closed
Replace Cypress testing framework with a more modern alternative
priority: P2 process size: M
# Goal Replace `cypress` integration test framework with SvelteKit standard `playwright` test framework. Tests should run in isolation from production environment or data. ## Requirements - [x] replace `cypress` with `playwright` - [x] mock all `api` requests (consider using [Mock Service Worker](https://mswjs.io) or see what built-in facilities are provided by `playwright`) - [x] migrate and/or re-write existing tests that seem to provide genuine value ### Removed from scope - add a test for #203 (see #217) - add a test for #207 (see #218)
1.0
Replace Cypress testing framework with a more modern alternative - # Goal Replace `cypress` integration test framework with SvelteKit standard `playwright` test framework. Tests should run in isolation from production environment or data. ## Requirements - [x] replace `cypress` with `playwright` - [x] mock all `api` requests (consider using [Mock Service Worker](https://mswjs.io) or see what built-in facilities are provided by `playwright`) - [x] migrate and/or re-write existing tests that seem to provide genuine value ### Removed from scope - add a test for #203 (see #217) - add a test for #207 (see #218)
process
replace cypress testing framework with a more modern alternative goal replace cypress integration test framework with sveltekit standard playwright test framework tests should run in isolation from production environment or data requirements replace cypress with playwright mock all api requests consider using or see what built in facilities are provided by playwright migrate and or re write existing tests that seem to provide genuine value removed from scope add a test for see add a test for see
1
529,556
15,391,053,222
IssuesEvent
2021-03-03 14:08:57
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Add parameter for peaks indices in things
Framework Low Priority Stale
This issue was originally [TRAC 7275](http://trac.mantidproject.org/mantid/ticket/7275) Since sometimes you want to use a subset of all peaks, add a parameter to various SCD-type algorithms to use only some of the peaks.
1.0
Add parameter for peaks indices in things - This issue was originally [TRAC 7275](http://trac.mantidproject.org/mantid/ticket/7275) Since sometimes you want to use a subset of all peaks, add a parameter to various SCD-type algorithms to use only some of the peaks.
non_process
add parameter for peaks indices in things this issue was originally since sometimes you want to use a subset of all peaks add a parameter to various scd type algorithms to use only some of the peaks
0
146
2,577,590,520
IssuesEvent
2015-02-12 17:55:26
cfpb/hmda-viz-prototype
https://api.github.com/repos/cfpb/hmda-viz-prototype
opened
MSA name and State name
Processing
Create MSA name and State name json file - non numeric, non abbreviated
1.0
MSA name and State name - Create MSA name and State name json file - non numeric, non abbreviated
process
msa name and state name create msa name and state name json file non numeric non abbreviated
1
1,626
4,238,827,101
IssuesEvent
2016-07-06 06:41:16
nodejs/node
https://api.github.com/repos/nodejs/node
closed
When I write áéíóú, process.stdout.write in Windows English version, process.stdout is closed
process windows
When I write áéíóú, process.stdout.write in Windows English version, process.stdout is closed E.g. ```javascript process.stdout.write('Aplicación') //print nothing process.stdout.write('xxxx') // Error: this.socket.is closed ``` * **Version**: node v5.9.1 * **Platform**: Windows, 7 32 bit (English Language)
1.0
When I write áéíóú, process.stdout.write in Windows English version, process.stdout is closed - When I write áéíóú, process.stdout.write in Windows English version, process.stdout is closed E.g. ```javascript process.stdout.write('Aplicación') //print nothing process.stdout.write('xxxx') // Error: this.socket.is closed ``` * **Version**: node v5.9.1 * **Platform**: Windows, 7 32 bit (English Language)
process
when i write áéíóú process stdout write in windows english version process stdout is closed when i write áéíóú process stdout write in windows english version process stdout is closed e g javascript process stdout write aplicación print nothing process stdout write xxxx error this socket is closed version node platform windows bit english language
1
95,899
12,059,242,491
IssuesEvent
2020-04-15 18:55:07
palantir/blueprint
https://api.github.com/repos/palantir/blueprint
closed
Add resizing options to Core Kit Sketch
Domain: design
#### Environment - __Package version(s)__: 3.5.1 - __Browser and OS versions__: N/A #### Feature request Having the Sketch file in the repo is super helpful for developing against Blueprint! It would be very helpful to have the `Core Kit.sketch` file have the correct resize options set, i.e., fix height/width, pin to left/right. This would make it much easier to prototype using components in the Sketch file. #### Examples The navigation bar currently stretches the icons and positions when resized horizontally. With appropriate resize options (pin to left/pin to right), prototyping would be simplified. <img width="528" alt="resize_options" src="https://user-images.githubusercontent.com/250437/45394540-03767300-b5e6-11e8-912d-cf64ff7f1286.png">
1.0
Add resizing options to Core Kit Sketch - #### Environment - __Package version(s)__: 3.5.1 - __Browser and OS versions__: N/A #### Feature request Having the Sketch file in the repo is super helpful for developing against Blueprint! It would be very helpful to have the `Core Kit.sketch` file have the correct resize options set, i.e., fix height/width, pin to left/right. This would make it much easier to prototype using components in the Sketch file. #### Examples The navigation bar currently stretches the icons and positions when resized horizontally. With appropriate resize options (pin to left/pin to right), prototyping would be simplified. <img width="528" alt="resize_options" src="https://user-images.githubusercontent.com/250437/45394540-03767300-b5e6-11e8-912d-cf64ff7f1286.png">
non_process
add resizing options to core kit sketch environment package version s browser and os versions n a feature request having the sketch file in the repo is super helpful for developing against blueprint it would be very helpful to have the core kit sketch file have the correct resize options set i e fix height width pin to left right this would make it much easier to prototype using components in the sketch file examples the navigation bar currently stretches the icons and positions when resized horizontally with appropriate resize options pin to left pin to right prototyping would be simplified img width alt resize options src
0
116,615
4,704,317,996
IssuesEvent
2016-10-13 11:01:51
PowerlineApp/powerline-mobile
https://api.github.com/repos/PowerlineApp/powerline-mobile
closed
Facebook Registration - Address Entry
bug P2 - Medium Priority
When a user registers with Facebook for the first time, they still need to provide their address. The second screen of the e-mail registration flow (showing address + DOB + phone number fields) should be displayed to user. On iOS, the user is clicking the REgister with Facebook button and is being brought to their user profile screen instead of two the "Address" screen from the registration flow. The user should be brought to the Address screen. Same behavior should apply on Android, but unable to test on Android. Since we're working on this screen anyway, we should replace the DOB field with an "I am 13 or older" checkbox (do not allow users less than 13 to register). We should also remove Phone Number field from this screen for the time being.
1.0
Facebook Registration - Address Entry - When a user registers with Facebook for the first time, they still need to provide their address. The second screen of the e-mail registration flow (showing address + DOB + phone number fields) should be displayed to user. On iOS, the user is clicking the REgister with Facebook button and is being brought to their user profile screen instead of two the "Address" screen from the registration flow. The user should be brought to the Address screen. Same behavior should apply on Android, but unable to test on Android. Since we're working on this screen anyway, we should replace the DOB field with an "I am 13 or older" checkbox (do not allow users less than 13 to register). We should also remove Phone Number field from this screen for the time being.
non_process
facebook registration address entry when a user registers with facebook for the first time they still need to provide their address the second screen of the e mail registration flow showing address dob phone number fields should be displayed to user on ios the user is clicking the register with facebook button and is being brought to their user profile screen instead of two the address screen from the registration flow the user should be brought to the address screen same behavior should apply on android but unable to test on android since we re working on this screen anyway we should replace the dob field with an i am or older checkbox do not allow users less than to register we should also remove phone number field from this screen for the time being
0
245,804
26,567,520,995
IssuesEvent
2023-01-20 21:56:44
kubecost/cost-analyzer-helm-chart
https://api.github.com/repos/kubecost/cost-analyzer-helm-chart
closed
kiwigrid/k8s-sidecar:1.15.4 impacted by 7 Critical CVEs
bug security stale
**Describe the bug** kiwigrid/k8s-sidecar:1.15.4 contains 20 CVEs -- 1 Critical CVE affecting busybox, 1 Critical affecting ssl_client, and 5 Critical affecting expat. The following image tags do not contain these or any other Medium/High/Critical vulnerabilities: - kiwigrid/k8s-sidecar:1.16.0 - kiwigrid/k8s-sidecar:1.17.0 - kiwigrid/k8s-sidecar:1.18.0 - kiwigrid/k8s-sidecar:1.18.1 - kiwigrid/k8s-sidecar:1.19.0 - kiwigrid/k8s-sidecar:1.19.1 - kiwigrid/k8s-sidecar:1.19.2 **To Reproduce** `trivy image kiwigrid/k8s-sidecar:1.15.4` **Collect logs (please complete the following information):** Table view of scan results below. ``` crisnevareswsl@AUSCNEVARW10L:~$ trivy i kiwigrid/k8s-sidecar:1.15.4 2022-07-12T11:18:25.514-0500 INFO Vulnerability scanning is enabled 2022-07-12T11:18:25.514-0500 INFO Secret scanning is enabled 2022-07-12T11:18:25.514-0500 INFO If your scanning is slow, please try '--security-checks vuln' to disable secret scanning 2022-07-12T11:18:25.514-0500 INFO Please see also https://aquasecurity.github.io/trivy/v0.29.2/docs/secret/scanning/#recommendation for faster secret detection 2022-07-12T11:18:27.829-0500 INFO Detected OS: alpine 2022-07-12T11:18:27.829-0500 INFO Detecting Alpine vulnerabilities... 2022-07-12T11:18:27.830-0500 INFO Number of language-specific files: 1 2022-07-12T11:18:27.830-0500 INFO Detecting python-pkg vulnerabilities... kiwigrid/k8s-sidecar:1.15.4 (alpine 3.15.0) Total: 20 (UNKNOWN: 0, LOW: 2, MEDIUM: 5, HIGH: 6, CRITICAL: 7) ┌──────────────┬────────────────┬──────────┬───────────────────┬───────────────┬─────────────────────────────────────────────────────────────┐ │ Library │ Vulnerability │ Severity │ Installed Version │ Fixed Version │ Title │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ busybox │ CVE-2022-28391 │ CRITICAL │ 1.34.1-r3 │ 1.34.1-r5 │ busybox: remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-28391 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-23852 │ CRITICAL │ 2.4.3-r0 │ 2.4.4-r0 │ expat: Integer overflow in function XML_GetBuffer │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-23852 │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-23990 │ │ │ │ expat: integer overflow in the doProlog function │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-23990 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-25235 │ CRITICAL │ 2.4.3-r0 │ 2.4.5-r0 │ expat: Malformed 2- and 3-byte UTF-8 sequences can lead to │ │ │ │ │ │ │ arbitrary code... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25235 │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25236 │ │ │ │ expat: Namespace-separator characters in "xmlns[:prefix]" │ │ │ │ │ │ │ attribute values can lead to arbitrary code... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25236 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-25315 │ CRITICAL │ 2.4.3-r0 │ 2.4.5-r0 │ expat: Integer overflow in storeRawNames() │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25315 │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25314 │ HIGH │ │ │ expat: integer overflow in copyString() │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25314 │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25313 │ MEDIUM │ │ │ expat: stack exhaustion in doctype parsing │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25313 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ krb5-libs │ CVE-2021-37750 │ MEDIUM │ 1.19.2-r4 │ 1.19.3-r0 │ krb5: NULL pointer dereference in process_tgs_req() in │ │ │ │ │ │ │ kdc/do_tgs_req.c via a FAST inner... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-37750 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libcrypto1.1 │ CVE-2022-0778 │ HIGH │ 1.1.1l-r8 │ 1.1.1n-r0 │ openssl: Infinite loop in BN_mod_sqrt() reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0778 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libcrypto1.1 │ CVE-2022-2097 │ LOW │ 1.1.1l-r8 │ 1.1.1q-r0 │ openssl: AES OCB fails to encrypt some bytes │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-2097 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libretls │ CVE-2022-0778 │ HIGH │ 3.3.4-r2 │ 3.3.4-r3 │ openssl: Infinite loop in BN_mod_sqrt() reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0778 │ ├──────────────┤ │ ├───────────────────┼───────────────┤ │ │ libssl1.1 │ │ │ 1.1.1l-r8 │ 1.1.1n-r0 │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libssl1.1 │ CVE-2022-2097 │ LOW │ 1.1.1l-r8 │ 1.1.1q-r0 │ openssl: AES OCB fails to encrypt some bytes │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-2097 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2021-3995 │ MEDIUM │ 2.37.2-r1 │ 2.37.3-r0 │ util-linux: Unauthorized unmount of FUSE filesystems │ │ │ │ │ │ │ belonging to users with similar uid... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-3995 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2021-3996 │ MEDIUM │ 2.37.2-r1 │ 2.37.3-r0 │ util-linux: Unauthorized unmount of filesystems in libmount │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-3996 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2022-0563 │ MEDIUM │ 2.37.2-r1 │ 2.37.4-r0 │ util-linux: partial disclosure of arbitrary files in chfn │ │ │ │ │ │ │ and chsh when compiled... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0563 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ ssl_client │ CVE-2022-28391 │ CRITICAL │ 1.34.1-r3 │ 1.34.1-r5 │ busybox: remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-28391 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ xz-libs │ CVE-2022-1271 │ HIGH │ 5.2.5-r0 │ 5.2.5-r1 │ gzip: arbitrary-file-write vulnerability │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-1271 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ zlib │ CVE-2018-25032 │ HIGH │ 1.2.11-r3 │ 1.2.12-r0 │ zlib: A flaw found in zlib when compressing (not │ │ │ │ │ │ │ decompressing) certain inputs... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2018-25032 │ └──────────────┴────────────────┴──────────┴───────────────────┴───────────────┴─────────────────────────────────────────────────────────────┘ Python (python-pkg) Total: 0 (UNKNOWN: 0, LOW: 0, MEDIUM: 0, HIGH: 0, CRITICAL: 0) ``` gz#2166<br><br><i>(related to <a href='https://kubecost.zendesk.com/agent/tickets/2166'>Zendesk ticket #2166</a>)</i>
True
kiwigrid/k8s-sidecar:1.15.4 impacted by 7 Critical CVEs - **Describe the bug** kiwigrid/k8s-sidecar:1.15.4 contains 20 CVEs -- 1 Critical CVE affecting busybox, 1 Critical affecting ssl_client, and 5 Critical affecting expat. The following image tags do not contain these or any other Medium/High/Critical vulnerabilities: - kiwigrid/k8s-sidecar:1.16.0 - kiwigrid/k8s-sidecar:1.17.0 - kiwigrid/k8s-sidecar:1.18.0 - kiwigrid/k8s-sidecar:1.18.1 - kiwigrid/k8s-sidecar:1.19.0 - kiwigrid/k8s-sidecar:1.19.1 - kiwigrid/k8s-sidecar:1.19.2 **To Reproduce** `trivy image kiwigrid/k8s-sidecar:1.15.4` **Collect logs (please complete the following information):** Table view of scan results below. ``` crisnevareswsl@AUSCNEVARW10L:~$ trivy i kiwigrid/k8s-sidecar:1.15.4 2022-07-12T11:18:25.514-0500 INFO Vulnerability scanning is enabled 2022-07-12T11:18:25.514-0500 INFO Secret scanning is enabled 2022-07-12T11:18:25.514-0500 INFO If your scanning is slow, please try '--security-checks vuln' to disable secret scanning 2022-07-12T11:18:25.514-0500 INFO Please see also https://aquasecurity.github.io/trivy/v0.29.2/docs/secret/scanning/#recommendation for faster secret detection 2022-07-12T11:18:27.829-0500 INFO Detected OS: alpine 2022-07-12T11:18:27.829-0500 INFO Detecting Alpine vulnerabilities... 2022-07-12T11:18:27.830-0500 INFO Number of language-specific files: 1 2022-07-12T11:18:27.830-0500 INFO Detecting python-pkg vulnerabilities... kiwigrid/k8s-sidecar:1.15.4 (alpine 3.15.0) Total: 20 (UNKNOWN: 0, LOW: 2, MEDIUM: 5, HIGH: 6, CRITICAL: 7) ┌──────────────┬────────────────┬──────────┬───────────────────┬───────────────┬─────────────────────────────────────────────────────────────┐ │ Library │ Vulnerability │ Severity │ Installed Version │ Fixed Version │ Title │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ busybox │ CVE-2022-28391 │ CRITICAL │ 1.34.1-r3 │ 1.34.1-r5 │ busybox: remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-28391 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-23852 │ CRITICAL │ 2.4.3-r0 │ 2.4.4-r0 │ expat: Integer overflow in function XML_GetBuffer │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-23852 │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-23990 │ │ │ │ expat: integer overflow in the doProlog function │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-23990 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-25235 │ CRITICAL │ 2.4.3-r0 │ 2.4.5-r0 │ expat: Malformed 2- and 3-byte UTF-8 sequences can lead to │ │ │ │ │ │ │ arbitrary code... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25235 │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25236 │ │ │ │ expat: Namespace-separator characters in "xmlns[:prefix]" │ │ │ │ │ │ │ attribute values can lead to arbitrary code... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25236 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ CVE-2022-25315 │ CRITICAL │ 2.4.3-r0 │ 2.4.5-r0 │ expat: Integer overflow in storeRawNames() │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25315 │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25314 │ HIGH │ │ │ expat: integer overflow in copyString() │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25314 │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ CVE-2022-25313 │ MEDIUM │ │ │ expat: stack exhaustion in doctype parsing │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-25313 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ krb5-libs │ CVE-2021-37750 │ MEDIUM │ 1.19.2-r4 │ 1.19.3-r0 │ krb5: NULL pointer dereference in process_tgs_req() in │ │ │ │ │ │ │ kdc/do_tgs_req.c via a FAST inner... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-37750 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libcrypto1.1 │ CVE-2022-0778 │ HIGH │ 1.1.1l-r8 │ 1.1.1n-r0 │ openssl: Infinite loop in BN_mod_sqrt() reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0778 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libcrypto1.1 │ CVE-2022-2097 │ LOW │ 1.1.1l-r8 │ 1.1.1q-r0 │ openssl: AES OCB fails to encrypt some bytes │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-2097 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libretls │ CVE-2022-0778 │ HIGH │ 3.3.4-r2 │ 3.3.4-r3 │ openssl: Infinite loop in BN_mod_sqrt() reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0778 │ ├──────────────┤ │ ├───────────────────┼───────────────┤ │ │ libssl1.1 │ │ │ 1.1.1l-r8 │ 1.1.1n-r0 │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libssl1.1 │ CVE-2022-2097 │ LOW │ 1.1.1l-r8 │ 1.1.1q-r0 │ openssl: AES OCB fails to encrypt some bytes │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-2097 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2021-3995 │ MEDIUM │ 2.37.2-r1 │ 2.37.3-r0 │ util-linux: Unauthorized unmount of FUSE filesystems │ │ │ │ │ │ │ belonging to users with similar uid... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-3995 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2021-3996 │ MEDIUM │ 2.37.2-r1 │ 2.37.3-r0 │ util-linux: Unauthorized unmount of filesystems in libmount │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2021-3996 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ CVE-2022-0563 │ MEDIUM │ 2.37.2-r1 │ 2.37.4-r0 │ util-linux: partial disclosure of arbitrary files in chfn │ │ │ │ │ │ │ and chsh when compiled... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-0563 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ ssl_client │ CVE-2022-28391 │ CRITICAL │ 1.34.1-r3 │ 1.34.1-r5 │ busybox: remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-28391 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ xz-libs │ CVE-2022-1271 │ HIGH │ 5.2.5-r0 │ 5.2.5-r1 │ gzip: arbitrary-file-write vulnerability │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2022-1271 │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ zlib │ CVE-2018-25032 │ HIGH │ 1.2.11-r3 │ 1.2.12-r0 │ zlib: A flaw found in zlib when compressing (not │ │ │ │ │ │ │ decompressing) certain inputs... │ │ │ │ │ │ │ https://avd.aquasec.com/nvd/cve-2018-25032 │ └──────────────┴────────────────┴──────────┴───────────────────┴───────────────┴─────────────────────────────────────────────────────────────┘ Python (python-pkg) Total: 0 (UNKNOWN: 0, LOW: 0, MEDIUM: 0, HIGH: 0, CRITICAL: 0) ``` gz#2166<br><br><i>(related to <a href='https://kubecost.zendesk.com/agent/tickets/2166'>Zendesk ticket #2166</a>)</i>
non_process
kiwigrid sidecar impacted by critical cves describe the bug kiwigrid sidecar contains cves critical cve affecting busybox critical affecting ssl client and critical affecting expat the following image tags do not contain these or any other medium high critical vulnerabilities kiwigrid sidecar kiwigrid sidecar kiwigrid sidecar kiwigrid sidecar kiwigrid sidecar kiwigrid sidecar kiwigrid sidecar to reproduce trivy image kiwigrid sidecar collect logs please complete the following information table view of scan results below crisnevareswsl trivy i kiwigrid sidecar info vulnerability scanning is enabled info secret scanning is enabled info if your scanning is slow please try security checks vuln to disable secret scanning info please see also for faster secret detection info detected os alpine info detecting alpine vulnerabilities info number of language specific files info detecting python pkg vulnerabilities kiwigrid sidecar alpine total unknown low medium high critical ┌──────────────┬────────────────┬──────────┬───────────────────┬───────────────┬─────────────────────────────────────────────────────────────┐ │ library │ vulnerability │ severity │ installed version │ fixed version │ title │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ busybox │ cve │ critical │ │ │ busybox remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ cve │ critical │ │ │ expat integer overflow in function xml getbuffer │ │ │ │ │ │ │ │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ cve │ │ │ │ expat integer overflow in the doprolog function │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ cve │ critical │ │ │ expat malformed and byte utf sequences can lead to │ │ │ │ │ │ │ arbitrary code │ │ │ │ │ │ │ │ │ ├────────────────┤ │ │ ├─────────────────────────────────────────────────────────────┤ │ │ cve │ │ │ │ expat namespace separator characters in xmlns │ │ │ │ │ │ │ attribute values can lead to arbitrary code │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ expat │ cve │ critical │ │ │ expat integer overflow in storerawnames │ │ │ │ │ │ │ │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ cve │ high │ │ │ expat integer overflow in copystring │ │ │ │ │ │ │ │ │ ├────────────────┼──────────┤ │ ├─────────────────────────────────────────────────────────────┤ │ │ cve │ medium │ │ │ expat stack exhaustion in doctype parsing │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libs │ cve │ medium │ │ │ null pointer dereference in process tgs req in │ │ │ │ │ │ │ kdc do tgs req c via a fast inner │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ │ cve │ high │ │ │ openssl infinite loop in bn mod sqrt reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ │ cve │ low │ │ │ openssl aes ocb fails to encrypt some bytes │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libretls │ cve │ high │ │ │ openssl infinite loop in bn mod sqrt reachable when │ │ │ │ │ │ │ parsing certificates │ │ │ │ │ │ │ │ ├──────────────┤ │ ├───────────────────┼───────────────┤ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ │ cve │ low │ │ │ openssl aes ocb fails to encrypt some bytes │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ cve │ medium │ │ │ util linux unauthorized unmount of fuse filesystems │ │ │ │ │ │ │ belonging to users with similar uid │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ cve │ medium │ │ │ util linux unauthorized unmount of filesystems in libmount │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ libuuid │ cve │ medium │ │ │ util linux partial disclosure of arbitrary files in chfn │ │ │ │ │ │ │ and chsh when compiled │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ ssl client │ cve │ critical │ │ │ busybox remote attackers may execute arbitrary code if │ │ │ │ │ │ │ netstat is used │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ xz libs │ cve │ high │ │ │ gzip arbitrary file write vulnerability │ │ │ │ │ │ │ │ ├──────────────┼────────────────┼──────────┼───────────────────┼───────────────┼─────────────────────────────────────────────────────────────┤ │ zlib │ cve │ high │ │ │ zlib a flaw found in zlib when compressing not │ │ │ │ │ │ │ decompressing certain inputs │ │ │ │ │ │ │ │ └──────────────┴────────────────┴──────────┴───────────────────┴───────────────┴─────────────────────────────────────────────────────────────┘ python python pkg total unknown low medium high critical gz related to
0
11,412
14,241,651,069
IssuesEvent
2020-11-18 23:53:29
googleapis/python-logging
https://api.github.com/repos/googleapis/python-logging
closed
`from google.cloud import logging` collides with `logging` from standard library
api: logging priority: p2 type: process
Also, an issue with `from google.cloud import logging` is that it collides with the `logging` standard library, and they are often [going to be used together](https://cloud.google.com/logging/docs/setup/python#connecting_the_library_to_python_logging). Right now we tell users to import as `import google.cloud.logging` and use the full path in their code. I wonder if it would make sense to rename the library to `google.cloud.cloudlogging` or something so there's no naming conflict _Originally posted by @Daniel-Sanche in https://github.com/googleapis/python-logging/pull/83#discussion_r515238862_
1.0
`from google.cloud import logging` collides with `logging` from standard library - Also, an issue with `from google.cloud import logging` is that it collides with the `logging` standard library, and they are often [going to be used together](https://cloud.google.com/logging/docs/setup/python#connecting_the_library_to_python_logging). Right now we tell users to import as `import google.cloud.logging` and use the full path in their code. I wonder if it would make sense to rename the library to `google.cloud.cloudlogging` or something so there's no naming conflict _Originally posted by @Daniel-Sanche in https://github.com/googleapis/python-logging/pull/83#discussion_r515238862_
process
from google cloud import logging collides with logging from standard library also an issue with from google cloud import logging is that it collides with the logging standard library and they are often right now we tell users to import as import google cloud logging and use the full path in their code i wonder if it would make sense to rename the library to google cloud cloudlogging or something so there s no naming conflict originally posted by daniel sanche in
1
513,135
14,916,406,163
IssuesEvent
2021-01-22 18:09:30
xwikisas/application-mocca-calendar
https://api.github.com/repos/xwikisas/application-mocca-calendar
closed
Warnings for deprecated usage of method when using Calendar Pro
Priority: Major Type: Bug
STEPS TO REPRODUCE Environment: Windows 10 Pro 64bit, Chrome 87, using an instance of XWiki 12.10.2 on PostgreSQL 13, Tomcat 9.0.41 While using Calendar Pro 2.9.2, in XWiki console there are many warnings displayed related to deprecated usage of method, like: `2021-01-19 12:14:58,119 [http-nio-1115-exec-8 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Events&page=xwiki%3AMoccaCalendar.Te-st-ed-There-are-5-French-accents-the-cedille-C-the-accent-aigu-e-the-accent-circonflexe-a-e-i-o-u-the-accent-grave-a-e-u-and-the-accent-trema-e-i-u-The-umlaut-a-e-i-o-u-is-the-only-accent-used-in-German-writing.Event1.WebHome&isResize=0&delta=-86400000&startDate=2021-01-18T12:10:00&endDate=2021-01-18T12:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:17:23,485 [http-nio-1115-exec-7 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=0&delta=172800000&startDate=2021-01-20T10:00:00&endDate=2021-01-20T10:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:18:00,067 [http-nio-1115-exec-10 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=0&delta=-174600000&startDate=2021-01-18T09:30:00&endDate=2021-01-18T10:00:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:18:01,603 [http-nio-1115-exec-8 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=1&delta=9000000&startDate=2021-01-18T09:30:00&endDate=2021-01-18T12:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25`
1.0
Warnings for deprecated usage of method when using Calendar Pro - STEPS TO REPRODUCE Environment: Windows 10 Pro 64bit, Chrome 87, using an instance of XWiki 12.10.2 on PostgreSQL 13, Tomcat 9.0.41 While using Calendar Pro 2.9.2, in XWiki console there are many warnings displayed related to deprecated usage of method, like: `2021-01-19 12:14:58,119 [http-nio-1115-exec-8 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Events&page=xwiki%3AMoccaCalendar.Te-st-ed-There-are-5-French-accents-the-cedille-C-the-accent-aigu-e-the-accent-circonflexe-a-e-i-o-u-the-accent-grave-a-e-u-and-the-accent-trema-e-i-u-The-umlaut-a-e-i-o-u-is-the-only-accent-used-in-German-writing.Event1.WebHome&isResize=0&delta=-86400000&startDate=2021-01-18T12:10:00&endDate=2021-01-18T12:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:17:23,485 [http-nio-1115-exec-7 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=0&delta=172800000&startDate=2021-01-20T10:00:00&endDate=2021-01-20T10:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:18:00,067 [http-nio-1115-exec-10 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=0&delta=-174600000&startDate=2021-01-18T09:30:00&endDate=2021-01-18T10:00:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25 2021-01-19 12:18:01,603 [http-nio-1115-exec-8 - http://localhost:1115/xwiki/bin/get/MoccaCalendar/UpdateService?xpage=plain&outputSyntax=plain&calendarDoc=MoccaCalendar.Calendar2-us.WebHome&page=xwiki%3AMoccaCalendar.Calendar2-us.Event2.WebHome&isResize=1&delta=9000000&startDate=2021-01-18T09:30:00&endDate=2021-01-18T12:30:00&allDay=] WARN o.a.v.introspection - Deprecated usage of method [org.apache.velocity.tools.generic.MathTool.toInteger] in xwiki:MoccaCalendar.UpdateService@15,25`
non_process
warnings for deprecated usage of method when using calendar pro steps to reproduce environment windows pro chrome using an instance of xwiki on postgresql tomcat while using calendar pro in xwiki console there are many warnings displayed related to deprecated usage of method like warn o a v introspection deprecated usage of method in xwiki moccacalendar updateservice warn o a v introspection deprecated usage of method in xwiki moccacalendar updateservice warn o a v introspection deprecated usage of method in xwiki moccacalendar updateservice warn o a v introspection deprecated usage of method in xwiki moccacalendar updateservice
0
69,515
13,259,899,525
IssuesEvent
2020-08-20 17:23:10
coghex/abridgefaraway
https://api.github.com/repos/coghex/abridgefaraway
opened
dynamic texture loading from lua
new code
change the number of loaded images whenever a texture is added live
1.0
dynamic texture loading from lua - change the number of loaded images whenever a texture is added live
non_process
dynamic texture loading from lua change the number of loaded images whenever a texture is added live
0
13,281
15,761,259,412
IssuesEvent
2021-03-31 09:48:23
prisma/quaint
https://api.github.com/repos/prisma/quaint
closed
Run CI on GitHub Actions
kind/tech process/candidate
Now we can't run tests from 3rd party pull requests. We could move our test runs to GitHub actions, which would make it easier for people to write pull requests. And, additionally, we could run the tests on platforms such as Windows and macOS.
1.0
Run CI on GitHub Actions - Now we can't run tests from 3rd party pull requests. We could move our test runs to GitHub actions, which would make it easier for people to write pull requests. And, additionally, we could run the tests on platforms such as Windows and macOS.
process
run ci on github actions now we can t run tests from party pull requests we could move our test runs to github actions which would make it easier for people to write pull requests and additionally we could run the tests on platforms such as windows and macos
1
139,955
11,300,003,325
IssuesEvent
2020-01-17 12:37:28
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
opened
Mockice instances is CBS integration tests has a low resources
area/core-and-supporting bug test-failing
<!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** Sometimes integration tests on CBS fail due by problem with provisioning pod of Mockice instances in tests. Probably this is a problem with low resources of instance. **Expected result** CBS should work with Mockice **Actual result** CBS tests sometimes fail due by problem with Mockice provisioning.
1.0
Mockice instances is CBS integration tests has a low resources - <!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** Sometimes integration tests on CBS fail due by problem with provisioning pod of Mockice instances in tests. Probably this is a problem with low resources of instance. **Expected result** CBS should work with Mockice **Actual result** CBS tests sometimes fail due by problem with Mockice provisioning.
non_process
mockice instances is cbs integration tests has a low resources thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description sometimes integration tests on cbs fail due by problem with provisioning pod of mockice instances in tests probably this is a problem with low resources of instance expected result cbs should work with mockice actual result cbs tests sometimes fail due by problem with mockice provisioning
0
1,404
3,968,812,007
IssuesEvent
2016-05-03 20:59:41
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
ServiceControllerTests.PauseAndContinue test failed in CI
blocking-clean-ci System.ServiceProcess
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win10_debug/142/consoleFull ``` 18:55:59 System.ServiceProcess.Tests.ServiceControllerTests.PauseAndContinue [FAIL] 18:55:59 Assert.Equal() Failure 18:55:59 Expected: Running 18:55:59 Actual: StartPending 18:55:59 Stack Trace: 18:55:59 d:\j\workspace\outerloop_win---0cba2915\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs(172,0): at System.ServiceProcess.Tests.ServiceControllerTests.PauseAndContinue() ```
1.0
ServiceControllerTests.PauseAndContinue test failed in CI - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win10_debug/142/consoleFull ``` 18:55:59 System.ServiceProcess.Tests.ServiceControllerTests.PauseAndContinue [FAIL] 18:55:59 Assert.Equal() Failure 18:55:59 Expected: Running 18:55:59 Actual: StartPending 18:55:59 Stack Trace: 18:55:59 d:\j\workspace\outerloop_win---0cba2915\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs(172,0): at System.ServiceProcess.Tests.ServiceControllerTests.PauseAndContinue() ```
process
servicecontrollertests pauseandcontinue test failed in ci system serviceprocess tests servicecontrollertests pauseandcontinue assert equal failure expected running actual startpending stack trace d j workspace outerloop win src system serviceprocess servicecontroller tests system serviceprocess servicecontroller tests servicecontrollertests cs at system serviceprocess tests servicecontrollertests pauseandcontinue
1
18,839
24,744,471,235
IssuesEvent
2022-10-21 08:34:28
fadeoutsoftware/WASDI
https://api.github.com/repos/fadeoutsoftware/WASDI
closed
Runtime error: mosaic_tile_checker failed due to code issue
bug P2 libraries app / processor
Execution of mosaic_tile_checker failed because Exception: : 'NoneType' object is not iterable
1.0
Runtime error: mosaic_tile_checker failed due to code issue - Execution of mosaic_tile_checker failed because Exception: : 'NoneType' object is not iterable
process
runtime error mosaic tile checker failed due to code issue execution of mosaic tile checker failed because exception nonetype object is not iterable
1
84,470
10,540,790,987
IssuesEvent
2019-10-02 09:14:05
pydata/sparse
https://api.github.com/repos/pydata/sparse
closed
Cython for radix argsort and fast indexing?
design decision good first issue help wanted type:performance
Since our main bottleneck is sorting, I was considering adding Cython as a dev dependency to create a - Custom radix sort solution. - Blazing fast indexing (#60) I was considering this for 0.3.
1.0
Cython for radix argsort and fast indexing? - Since our main bottleneck is sorting, I was considering adding Cython as a dev dependency to create a - Custom radix sort solution. - Blazing fast indexing (#60) I was considering this for 0.3.
non_process
cython for radix argsort and fast indexing since our main bottleneck is sorting i was considering adding cython as a dev dependency to create a custom radix sort solution blazing fast indexing i was considering this for
0
354
2,794,158,344
IssuesEvent
2015-05-11 15:16:53
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Grok Extractor, Converter Support
processing
Hi, when parsing metrics out of our logs, using a Grok extractor is really convenient. However, there's currently no way to apply converters to extracted fields? It'd be very useful to, for example, be able to attach converters to either fields in the extractor or individual Grok patterns (even better!).
1.0
Grok Extractor, Converter Support - Hi, when parsing metrics out of our logs, using a Grok extractor is really convenient. However, there's currently no way to apply converters to extracted fields? It'd be very useful to, for example, be able to attach converters to either fields in the extractor or individual Grok patterns (even better!).
process
grok extractor converter support hi when parsing metrics out of our logs using a grok extractor is really convenient however there s currently no way to apply converters to extracted fields it d be very useful to for example be able to attach converters to either fields in the extractor or individual grok patterns even better
1
224,529
17,754,769,129
IssuesEvent
2021-08-28 14:34:53
MuzaffarMohammed/kfmenterprises-ecommerce
https://api.github.com/repos/MuzaffarMohammed/kfmenterprises-ecommerce
closed
Order details page redirection in order placed mail.
enhancement good first issue Priority Tested - Dev In production
![image](https://user-images.githubusercontent.com/50576677/127009918-b8266ec0-be24-4145-b04d-9052e7ecf113.png) Acceptance Criteria: User should have an option to view order details in application by navigation. 1. Make order Id as a clickable link to order detail page in application. 2. A button to view order to same order detail page.
1.0
Order details page redirection in order placed mail. - ![image](https://user-images.githubusercontent.com/50576677/127009918-b8266ec0-be24-4145-b04d-9052e7ecf113.png) Acceptance Criteria: User should have an option to view order details in application by navigation. 1. Make order Id as a clickable link to order detail page in application. 2. A button to view order to same order detail page.
non_process
order details page redirection in order placed mail acceptance criteria user should have an option to view order details in application by navigation make order id as a clickable link to order detail page in application a button to view order to same order detail page
0
362,561
25,381,495,573
IssuesEvent
2022-11-21 17:55:48
gnosischain/documentation
https://api.github.com/repos/gnosischain/documentation
closed
Validators - Rewards & Penalties
documentation validators
Current Page: https://docs.gnosischain.com/node/incentives ## Tasks - [ ] Check validity of information ### Rewards - [ ] Calculate and show the current Yield - [ ] Embed key information: https://dune.xyz/maxaleks/Gnosis-Beacon-Chain-(Deposits) ### Penalties - [ ] Penalties - [ ] Link to Penalties docs (?) - [ ] Emphasize slashing - [ ] Emphasize inactivity leaks (including mass-inactivity leaks)
1.0
Validators - Rewards & Penalties - Current Page: https://docs.gnosischain.com/node/incentives ## Tasks - [ ] Check validity of information ### Rewards - [ ] Calculate and show the current Yield - [ ] Embed key information: https://dune.xyz/maxaleks/Gnosis-Beacon-Chain-(Deposits) ### Penalties - [ ] Penalties - [ ] Link to Penalties docs (?) - [ ] Emphasize slashing - [ ] Emphasize inactivity leaks (including mass-inactivity leaks)
non_process
validators rewards penalties current page tasks check validity of information rewards calculate and show the current yield embed key information penalties penalties link to penalties docs emphasize slashing emphasize inactivity leaks including mass inactivity leaks
0
1,270
3,799,502,479
IssuesEvent
2016-03-23 16:05:53
SIMEXP/niak
https://api.github.com/repos/SIMEXP/niak
closed
shortcut for quitting in the QC of fMRI preproc
enhancement preprocessing quality control
illdopejake wrote: It would be cool if there was a keyboard shortcut to Quit. (Maybe there is and I just don't know about it?)
1.0
shortcut for quitting in the QC of fMRI preproc - illdopejake wrote: It would be cool if there was a keyboard shortcut to Quit. (Maybe there is and I just don't know about it?)
process
shortcut for quitting in the qc of fmri preproc illdopejake wrote it would be cool if there was a keyboard shortcut to quit maybe there is and i just don t know about it
1
19,599
25,952,353,300
IssuesEvent
2022-12-17 19:26:31
RobotComponents/RobotComponents
https://api.github.com/repos/RobotComponents/RobotComponents
closed
Mod: Make controller utility base class
modification in process
Make a controller utility base class and all functionality we have to this base class. Everything is now written in the component classes. We should centralize this to avoid duplicate codes, and to make it accessible through the API of the base class.
1.0
Mod: Make controller utility base class - Make a controller utility base class and all functionality we have to this base class. Everything is now written in the component classes. We should centralize this to avoid duplicate codes, and to make it accessible through the API of the base class.
process
mod make controller utility base class make a controller utility base class and all functionality we have to this base class everything is now written in the component classes we should centralize this to avoid duplicate codes and to make it accessible through the api of the base class
1
154,205
19,710,829,880
IssuesEvent
2022-01-13 05:00:27
CanarysPlayground/Demo02063984
https://api.github.com/repos/CanarysPlayground/Demo02063984
closed
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz
security vulnerability
## CVE-2020-7693 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary> <p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p> <p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sockjs/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.0.1.tgz (Root Library) - webpack-dev-server-3.2.1.tgz - :x: **sockjs-0.3.19.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/CanarysPlayground/Demo02063984/commit/6da9ceb6f30da0e72b9f2d569d94820e4e9ec3b6">6da9ceb6f30da0e72b9f2d569d94820e4e9ec3b6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20. <p>Publish Date: 2020-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p> <p>Release Date: 2020-07-14</p> <p>Fix Resolution: sockjs - 0.3.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz - ## CVE-2020-7693 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary> <p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p> <p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sockjs/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.0.1.tgz (Root Library) - webpack-dev-server-3.2.1.tgz - :x: **sockjs-0.3.19.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/CanarysPlayground/Demo02063984/commit/6da9ceb6f30da0e72b9f2d569d94820e4e9ec3b6">6da9ceb6f30da0e72b9f2d569d94820e4e9ec3b6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20. <p>Publish Date: 2020-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p> <p>Release Date: 2020-07-14</p> <p>Fix Resolution: sockjs - 0.3.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in sockjs tgz cve medium severity vulnerability vulnerable library sockjs tgz sockjs node is a server counterpart of sockjs client a javascript library that provides a websocket like object in the browser sockjs gives you a coherent cross browser javascript api which creates a low latency full duplex cross domain communication library home page a href path to dependency file package json path to vulnerable library node modules sockjs package json dependency hierarchy react scripts tgz root library webpack dev server tgz x sockjs tgz vulnerable library found in head commit a href found in base branch master vulnerability details incorrect handling of upgrade header with the value websocket leads in crashing of containers hosting sockjs apps this affects the package sockjs before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sockjs step up your open source security game with whitesource
0
417,156
12,156,235,365
IssuesEvent
2020-04-25 16:27:56
dev-protocol/stakes.social
https://api.github.com/repos/dev-protocol/stakes.social
opened
Mobile support: ButtonCard
Priority: Medium Type: Feature
On a mobile device, it would be a good look if it looked like this: ![ButtonCard](https://user-images.githubusercontent.com/1970283/80285019-1dce7780-875d-11ea-8fb1-bd8936f8fafc.png)
1.0
Mobile support: ButtonCard - On a mobile device, it would be a good look if it looked like this: ![ButtonCard](https://user-images.githubusercontent.com/1970283/80285019-1dce7780-875d-11ea-8fb1-bd8936f8fafc.png)
non_process
mobile support buttoncard on a mobile device it would be a good look if it looked like this
0
21,462
29,498,383,903
IssuesEvent
2023-06-02 19:07:38
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Port annotate middleware to MLv2 part 1: port aggregation/expression metadata fns
Querying/Processor .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
The `annotate` middleware code to generate metadata for aggregations and expressions overlaps heavily with the stuff in MLv2 (and with the JS stuff ported in https://github.com/metabase/metabase/issues/28882), we should rework things so it uses MLv2 like everybody else does. This will be both a proof-of-concept for using MLv2 in the QP and give us a lot of extra test coverage to make sure that MLv2 metadata generation is consistent with what the QP currently generates
2.0
[MLv2] Port annotate middleware to MLv2 part 1: port aggregation/expression metadata fns - The `annotate` middleware code to generate metadata for aggregations and expressions overlaps heavily with the stuff in MLv2 (and with the JS stuff ported in https://github.com/metabase/metabase/issues/28882), we should rework things so it uses MLv2 like everybody else does. This will be both a proof-of-concept for using MLv2 in the QP and give us a lot of extra test coverage to make sure that MLv2 metadata generation is consistent with what the QP currently generates
process
port annotate middleware to part port aggregation expression metadata fns the annotate middleware code to generate metadata for aggregations and expressions overlaps heavily with the stuff in and with the js stuff ported in we should rework things so it uses like everybody else does this will be both a proof of concept for using in the qp and give us a lot of extra test coverage to make sure that metadata generation is consistent with what the qp currently generates
1
756
3,236,931,634
IssuesEvent
2015-10-14 09:00:36
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Discuss allowed use of endogenous variables outside of model-block
decision preprocessor
Consider the mod-file ``` var y, c, k, a, h, b; varexo e, u; parameters beta, rho, alpha, delta, theta, psi, tau test; alpha = 0.36; rho = 0.95; tau = 0.025; beta = 0.99; delta = 0.025; psi = 0; theta = 2.95; phi = 0.1; test=y*beta; model; c*theta*h^(1+psi)=(1-alpha)*y; k = beta*(((exp(b)*c)/(exp(b(+1))*c(+1))) *(exp(b(+1))*alpha*y(+1)+(1-delta)*k)); y = exp(a)*(k(-1)^alpha)*(h^(1-alpha)); k = exp(b)*(y-c)+(1-delta)*k(-1); a = rho*a(-1)+tau*b(-1) + e; b = tau*a(-1)+rho*b(-1) + u; end; initval; y = 1.08068253095672; c = 0.80359242014163; h = 0.29175631001732; k = 11.08360443260358; a = 0; b = 0; e = 0; u = 0; end; shocks; var e; stderr 0.009; var u; stderr 0.009; var e, u = phi*0.009*0.009; end; stoch_simul; ``` The definition ``` test=y*beta; ``` results in the preprocessor plugging in for the endogenous variable ```y``` with its yet uncomputed steady state, i.e. 0. Users thus can create a circular problem with the steady state of ``` y ``` depending on ``` test ``` with ``` test ``` depending on ``` y ``` - and would never notice the problem, because the definition does not result in an error. Would it be possible to block this behavior? If we want to allow users to access the steady state value of endogenous variables outside of the model-block, it should be through the ```steady_state```operator. Or do I miss something that makes this behavior desirable?
1.0
Discuss allowed use of endogenous variables outside of model-block - Consider the mod-file ``` var y, c, k, a, h, b; varexo e, u; parameters beta, rho, alpha, delta, theta, psi, tau test; alpha = 0.36; rho = 0.95; tau = 0.025; beta = 0.99; delta = 0.025; psi = 0; theta = 2.95; phi = 0.1; test=y*beta; model; c*theta*h^(1+psi)=(1-alpha)*y; k = beta*(((exp(b)*c)/(exp(b(+1))*c(+1))) *(exp(b(+1))*alpha*y(+1)+(1-delta)*k)); y = exp(a)*(k(-1)^alpha)*(h^(1-alpha)); k = exp(b)*(y-c)+(1-delta)*k(-1); a = rho*a(-1)+tau*b(-1) + e; b = tau*a(-1)+rho*b(-1) + u; end; initval; y = 1.08068253095672; c = 0.80359242014163; h = 0.29175631001732; k = 11.08360443260358; a = 0; b = 0; e = 0; u = 0; end; shocks; var e; stderr 0.009; var u; stderr 0.009; var e, u = phi*0.009*0.009; end; stoch_simul; ``` The definition ``` test=y*beta; ``` results in the preprocessor plugging in for the endogenous variable ```y``` with its yet uncomputed steady state, i.e. 0. Users thus can create a circular problem with the steady state of ``` y ``` depending on ``` test ``` with ``` test ``` depending on ``` y ``` - and would never notice the problem, because the definition does not result in an error. Would it be possible to block this behavior? If we want to allow users to access the steady state value of endogenous variables outside of the model-block, it should be through the ```steady_state```operator. Or do I miss something that makes this behavior desirable?
process
discuss allowed use of endogenous variables outside of model block consider the mod file var y c k a h b varexo e u parameters beta rho alpha delta theta psi tau test alpha rho tau beta delta psi theta phi test y beta model c theta h psi alpha y k beta exp b c exp b c exp b alpha y delta k y exp a k alpha h alpha k exp b y c delta k a rho a tau b e b tau a rho b u end initval y c h k a b e u end shocks var e stderr var u stderr var e u phi end stoch simul the definition test y beta results in the preprocessor plugging in for the endogenous variable y with its yet uncomputed steady state i e users thus can create a circular problem with the steady state of y depending on test with test depending on y and would never notice the problem because the definition does not result in an error would it be possible to block this behavior if we want to allow users to access the steady state value of endogenous variables outside of the model block it should be through the steady state operator or do i miss something that makes this behavior desirable
1
22,222
30,771,725,737
IssuesEvent
2023-07-31 00:37:06
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Desenvolvedor BackEnd / SRE Infra Google] - Remoto - HunterHunter
BACK-END COMERCIAL INFRAESTRUTURA MONGODB JAVASCRIPT SQL TYPESCRIPT NODE.JS DOCKER KUBERNETES REQUISITOS REMOTO GOOGLE CLOUD PROCESSOS BACKEND INGLÊS SEGURANÇA UMA C CLEAN R DOCUMENTAÇÃO MODELAGEM DE DADOS TERRAFORM HELP WANTED ESPECIALISTA TESTES DE PERFORMANCE MONITORAMENTO SRE Stale
[Desenvolvedor BackEnd / SRE Infra Google] - Remoto. R$ 8 mil CLT. Com idas a cada 3 meses uma vez para São Paulo no horário comercial. Skills técnicas relevantes: Mais de 2 anos de experiência Terraform (mandatório) Google Cloud com mais de 2 anos de experiência (mandatório) Javascript / Typescript (mandatório) Microsserviços, Docker e Kubernetes Node.js MongoDB e bancos de dados orientados a documentos SQL e bancos de dados relacionais Modelagem de dados Clean Code Análise e testes de Performance e Escalabilidade Boas práticas de Segurança Escrita de documentação de API, OpenAPI Inglês técnico (boa leitura) Quanto mais dessas skills você dominar, mais fácil vai se integrar ao time e às demandas de dia a dia. Boa parte das skills são desejáveis, mas não mandatórias. Perguntas frequentes: Preciso ser especialista em toda essa stack? De forma alguma Preciso estar disposto a ser? Com certeza! Sou mais de infra mesmo, não desenvolvo há algum tempo, será que me enquadro nessa vaga? Acredito que sim! Também valorizamos perfis híbridos, e desde que esteja disposto a focar neste escopo de Infra Google end no curto-médio prazo, você pode ter um bom fit com a gente. Sou mais de back mesmo, nunca fui especialista de infra, será que me enquadro nessa vaga? Acredito que sim! Como comentei, valorizamos perfis híbridos. O seu desafio vai ser se desenvolver e ser nosso especialista em melhores práticas e gestão da infra da nossa plataforma com seu olhar de back end - e com certeza vão surgir algumas tasks de back end puro (regras de negócio) que você também vai poder contribuir, enquanto vai estudando e entendendo melhor a empresa. Quem buscamos? Estamos procuramos desenvolvedor back-end que esteja disposto a ser nosso especialista de infraestrutura no Google Cloud. Buscamos gente para desenvolver funcionalidades, definir arquiteturas de código, nos auxiliar com melhor uso da Google Cloud no nosso dia a dia, definir boas práticas de Infra as Code e participar dos processos de monitoramento dos ambientes de produção. Buscamos pessoas com experiência o suficiente para ter autonomia para se plugar rapidamente no time e começar a entregar valor; mas mais do que isso, buscamos gente que é interessada e está sempre aprendendo e evoluindo: no seu dia a dia você vai ter espaço para estudar tecnologias e ferramentas novas, propor melhorias e sugerir como as coisas deveriam ser, além de escrever código limpo (para aplicação e também para infra como código) em uma empresa que valoriza tecnologia de ponta e atualizada. Benefícios: Auxílio computador todo mês, para manter seu setup sempre atualizado! Plano de saúde Alice Remoto para sempre (nos encontramos uma vez a cada 3 meses em São Paulo) Horários flexíveis - tentamos fazer nossos encontros síncronos sempre pela manhã Salário alinhado ao mercado. Se tem interesse e com requisitos da oportunidade entre em contato pelo Linkedin informando qual é a oportunidade. [https://www.linkedin.com/in/erickamarquesb/]
1.0
[Desenvolvedor BackEnd / SRE Infra Google] - Remoto - HunterHunter - [Desenvolvedor BackEnd / SRE Infra Google] - Remoto. R$ 8 mil CLT. Com idas a cada 3 meses uma vez para São Paulo no horário comercial. Skills técnicas relevantes: Mais de 2 anos de experiência Terraform (mandatório) Google Cloud com mais de 2 anos de experiência (mandatório) Javascript / Typescript (mandatório) Microsserviços, Docker e Kubernetes Node.js MongoDB e bancos de dados orientados a documentos SQL e bancos de dados relacionais Modelagem de dados Clean Code Análise e testes de Performance e Escalabilidade Boas práticas de Segurança Escrita de documentação de API, OpenAPI Inglês técnico (boa leitura) Quanto mais dessas skills você dominar, mais fácil vai se integrar ao time e às demandas de dia a dia. Boa parte das skills são desejáveis, mas não mandatórias. Perguntas frequentes: Preciso ser especialista em toda essa stack? De forma alguma Preciso estar disposto a ser? Com certeza! Sou mais de infra mesmo, não desenvolvo há algum tempo, será que me enquadro nessa vaga? Acredito que sim! Também valorizamos perfis híbridos, e desde que esteja disposto a focar neste escopo de Infra Google end no curto-médio prazo, você pode ter um bom fit com a gente. Sou mais de back mesmo, nunca fui especialista de infra, será que me enquadro nessa vaga? Acredito que sim! Como comentei, valorizamos perfis híbridos. O seu desafio vai ser se desenvolver e ser nosso especialista em melhores práticas e gestão da infra da nossa plataforma com seu olhar de back end - e com certeza vão surgir algumas tasks de back end puro (regras de negócio) que você também vai poder contribuir, enquanto vai estudando e entendendo melhor a empresa. Quem buscamos? Estamos procuramos desenvolvedor back-end que esteja disposto a ser nosso especialista de infraestrutura no Google Cloud. Buscamos gente para desenvolver funcionalidades, definir arquiteturas de código, nos auxiliar com melhor uso da Google Cloud no nosso dia a dia, definir boas práticas de Infra as Code e participar dos processos de monitoramento dos ambientes de produção. Buscamos pessoas com experiência o suficiente para ter autonomia para se plugar rapidamente no time e começar a entregar valor; mas mais do que isso, buscamos gente que é interessada e está sempre aprendendo e evoluindo: no seu dia a dia você vai ter espaço para estudar tecnologias e ferramentas novas, propor melhorias e sugerir como as coisas deveriam ser, além de escrever código limpo (para aplicação e também para infra como código) em uma empresa que valoriza tecnologia de ponta e atualizada. Benefícios: Auxílio computador todo mês, para manter seu setup sempre atualizado! Plano de saúde Alice Remoto para sempre (nos encontramos uma vez a cada 3 meses em São Paulo) Horários flexíveis - tentamos fazer nossos encontros síncronos sempre pela manhã Salário alinhado ao mercado. Se tem interesse e com requisitos da oportunidade entre em contato pelo Linkedin informando qual é a oportunidade. [https://www.linkedin.com/in/erickamarquesb/]
process
remoto hunterhunter remoto r mil clt com idas a cada meses uma vez para são paulo no horário comercial skills técnicas relevantes mais de anos de experiência terraform mandatório google cloud com mais de anos de experiência mandatório javascript typescript mandatório microsserviços docker e kubernetes node js mongodb e bancos de dados orientados a documentos sql e bancos de dados relacionais modelagem de dados clean code análise e testes de performance e escalabilidade boas práticas de segurança escrita de documentação de api openapi inglês técnico boa leitura quanto mais dessas skills você dominar mais fácil vai se integrar ao time e às demandas de dia a dia boa parte das skills são desejáveis mas não mandatórias perguntas frequentes preciso ser especialista em toda essa stack de forma alguma preciso estar disposto a ser com certeza sou mais de infra mesmo não desenvolvo há algum tempo será que me enquadro nessa vaga acredito que sim também valorizamos perfis híbridos e desde que esteja disposto a focar neste escopo de infra google end no curto médio prazo você pode ter um bom fit com a gente sou mais de back mesmo nunca fui especialista de infra será que me enquadro nessa vaga acredito que sim como comentei valorizamos perfis híbridos o seu desafio vai ser se desenvolver e ser nosso especialista em melhores práticas e gestão da infra da nossa plataforma com seu olhar de back end e com certeza vão surgir algumas tasks de back end puro regras de negócio que você também vai poder contribuir enquanto vai estudando e entendendo melhor a empresa quem buscamos estamos procuramos desenvolvedor back end que esteja disposto a ser nosso especialista de infraestrutura no google cloud buscamos gente para desenvolver funcionalidades definir arquiteturas de código nos auxiliar com melhor uso da google cloud no nosso dia a dia definir boas práticas de infra as code e participar dos processos de monitoramento dos ambientes de produção buscamos pessoas com experiência o suficiente para ter autonomia para se plugar rapidamente no time e começar a entregar valor mas mais do que isso buscamos gente que é interessada e está sempre aprendendo e evoluindo no seu dia a dia você vai ter espaço para estudar tecnologias e ferramentas novas propor melhorias e sugerir como as coisas deveriam ser além de escrever código limpo para aplicação e também para infra como código em uma empresa que valoriza tecnologia de ponta e atualizada benefícios auxílio computador todo mês para manter seu setup sempre atualizado plano de saúde alice remoto para sempre nos encontramos uma vez a cada meses em são paulo horários flexíveis tentamos fazer nossos encontros síncronos sempre pela manhã salário alinhado ao mercado se tem interesse e com requisitos da oportunidade entre em contato pelo linkedin informando qual é a oportunidade
1
2,282
5,108,136,168
IssuesEvent
2017-01-05 16:50:08
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[Subtitles] [FR] La revue de la semaine n°2 : Syrie, médias, Orwell, CETA, lobbies, agriculture
Language: French Process: Someone is working on this issue Process: [3] Review (1) in progress
# Video title La revue de la semaine n°2 : Syrie, médias, Orwell, CETA, lobbies, agriculture # URL https://www.youtube.com/watch?v=YxhxvRoCATY # Youtube subtitles language Français # Duration 24:14 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=fr&bl=vmp&ui=hd&v=YxhxvRoCATY&tab=captions&ref=player&action_mde_edit_form=1
2.0
[Subtitles] [FR] La revue de la semaine n°2 : Syrie, médias, Orwell, CETA, lobbies, agriculture - # Video title La revue de la semaine n°2 : Syrie, médias, Orwell, CETA, lobbies, agriculture # URL https://www.youtube.com/watch?v=YxhxvRoCATY # Youtube subtitles language Français # Duration 24:14 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=fr&bl=vmp&ui=hd&v=YxhxvRoCATY&tab=captions&ref=player&action_mde_edit_form=1
process
la revue de la semaine n° syrie médias orwell ceta lobbies agriculture video title la revue de la semaine n° syrie médias orwell ceta lobbies agriculture url youtube subtitles language français duration subtitles url
1
21,592
29,993,306,536
IssuesEvent
2023-06-26 01:49:56
global-healthy-liveable-cities/global-indicators
https://api.github.com/repos/global-healthy-liveable-cities/global-indicators
opened
Configuration needs to be simpler
enhancement user process to document
Configuration is currently the most difficult and involving part of the spatial indicator calculation process. This involves specifying a range of data sources and parameters required to run analysis and document metadata for reporting. While this has been simplified from previous iterations, users still report difficulties. There are several challenges: - flexibility of configuration is required to allow for use in diverse contexts (e.g. diversity in terms of geography, data availability, audience/user requirements and interest, etc), but this brings complexity - configuration has been seperated from data storage and data outputs for clarity, but this brings fragmentation (navigating different depths of folders in different directories) - configuration uses yaml text files to allow readability for both machines and humans -- but currently it seems over-complex for humans to complete, and in turn for machines to read - the specification of URLs has led some users to believe that data will be retrieved by the process (this could potentially be done) - there is some inconsistency in the specification of data for different aspects (eg different ways of defining the CRS); should be consistent, and only ask for necessary things (for example, licence URL is not necessary; if that were required or considered unclear it could be pasted in the licence field) - the complexity of the current configuration file makes it challenging to develop a form interface for this; if simplfiied, the capacity for completion via a web app form should be supported Changes to the configuration process addressing the above aspects would likely involve some large changes, and would likely make previous configuration files not work. So, it is worth thinking about other aspects that could be simplified at the same time to minimise disruptions.
1.0
Configuration needs to be simpler - Configuration is currently the most difficult and involving part of the spatial indicator calculation process. This involves specifying a range of data sources and parameters required to run analysis and document metadata for reporting. While this has been simplified from previous iterations, users still report difficulties. There are several challenges: - flexibility of configuration is required to allow for use in diverse contexts (e.g. diversity in terms of geography, data availability, audience/user requirements and interest, etc), but this brings complexity - configuration has been seperated from data storage and data outputs for clarity, but this brings fragmentation (navigating different depths of folders in different directories) - configuration uses yaml text files to allow readability for both machines and humans -- but currently it seems over-complex for humans to complete, and in turn for machines to read - the specification of URLs has led some users to believe that data will be retrieved by the process (this could potentially be done) - there is some inconsistency in the specification of data for different aspects (eg different ways of defining the CRS); should be consistent, and only ask for necessary things (for example, licence URL is not necessary; if that were required or considered unclear it could be pasted in the licence field) - the complexity of the current configuration file makes it challenging to develop a form interface for this; if simplfiied, the capacity for completion via a web app form should be supported Changes to the configuration process addressing the above aspects would likely involve some large changes, and would likely make previous configuration files not work. So, it is worth thinking about other aspects that could be simplified at the same time to minimise disruptions.
process
configuration needs to be simpler configuration is currently the most difficult and involving part of the spatial indicator calculation process this involves specifying a range of data sources and parameters required to run analysis and document metadata for reporting while this has been simplified from previous iterations users still report difficulties there are several challenges flexibility of configuration is required to allow for use in diverse contexts e g diversity in terms of geography data availability audience user requirements and interest etc but this brings complexity configuration has been seperated from data storage and data outputs for clarity but this brings fragmentation navigating different depths of folders in different directories configuration uses yaml text files to allow readability for both machines and humans but currently it seems over complex for humans to complete and in turn for machines to read the specification of urls has led some users to believe that data will be retrieved by the process this could potentially be done there is some inconsistency in the specification of data for different aspects eg different ways of defining the crs should be consistent and only ask for necessary things for example licence url is not necessary if that were required or considered unclear it could be pasted in the licence field the complexity of the current configuration file makes it challenging to develop a form interface for this if simplfiied the capacity for completion via a web app form should be supported changes to the configuration process addressing the above aspects would likely involve some large changes and would likely make previous configuration files not work so it is worth thinking about other aspects that could be simplified at the same time to minimise disruptions
1
17,388
23,206,870,558
IssuesEvent
2022-08-02 06:32:17
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
postprocess fail: item/fluid has no source with 5dims new mining
mod:pypostprocessing postprocess-fail compatibility
### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [x] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [X] Pypostprocessing failure - [ ] Other ### What is the problem? Auto-tech is failing to load with 5DIM's New Mining mod ![unknown](https://user-images.githubusercontent.com/65198411/178821903-b84f99b9-f570-49de-8bef-49dc2bf224b8.png) ### Steps to reproduce 1. Install Py Suite (PyAE Beta via gdrive pinned in the discord #announcements-and-news 2. Install 5DIM's New Mining 3. Load up the game and gives this error during startup ### Additional context _No response_ ### Log file _No response_
2.0
postprocess fail: item/fluid has no source with 5dims new mining - ### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [x] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [X] Pypostprocessing failure - [ ] Other ### What is the problem? Auto-tech is failing to load with 5DIM's New Mining mod ![unknown](https://user-images.githubusercontent.com/65198411/178821903-b84f99b9-f570-49de-8bef-49dc2bf224b8.png) ### Steps to reproduce 1. Install Py Suite (PyAE Beta via gdrive pinned in the discord #announcements-and-news 2. Install 5DIM's New Mining 3. Load up the game and gives this error during startup ### Additional context _No response_ ### Log file _No response_
process
postprocess fail item fluid has no source with new mining mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem auto tech is failing to load with s new mining mod steps to reproduce install py suite pyae beta via gdrive pinned in the discord announcements and news install s new mining load up the game and gives this error during startup additional context no response log file no response
1
75,893
26,127,060,269
IssuesEvent
2022-12-28 20:15:56
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: Deprecation warning says to use non-existent symbols
defect scipy.optimize needs-decision deprecated
### Describe your issue. About a year ago (commit 6b08e087171479b30fc2dc015d2074a24844789e) the ``optimize.tnc`` module gained some deprecation warnings, indicating that symbols of the form ``optimze.tnc.<SYM>`` were deprecated, use instead ``optimize.<SYM>``. Since none of these symbols exist in optimize directly, doing as suggested generates ``ImportError``. ### Reproducing Code Example ```python $ python -c 'from scipy.optimize.tnc import RCSTRINGS ; from scipy.optimize import RCSTRINGS' ``` ### Error message ```shell <string>:1: DeprecationWarning: Please use `RCSTRINGS` from the `scipy.optimize` namespace, the `scipy.optimize.tnc` namespace is deprecated. Traceback (most recent call last): File "<string>", line 1, in <module> ImportError: cannot import name 'RCSTRINGS' from 'scipy.optimize' (C:\Program Files\Python310\lib\site-packages\scipy\optimize\__init__.py) ``` ### SciPy/NumPy/Python version information 1.9.3 1.24.1 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0)
1.0
BUG: Deprecation warning says to use non-existent symbols - ### Describe your issue. About a year ago (commit 6b08e087171479b30fc2dc015d2074a24844789e) the ``optimize.tnc`` module gained some deprecation warnings, indicating that symbols of the form ``optimze.tnc.<SYM>`` were deprecated, use instead ``optimize.<SYM>``. Since none of these symbols exist in optimize directly, doing as suggested generates ``ImportError``. ### Reproducing Code Example ```python $ python -c 'from scipy.optimize.tnc import RCSTRINGS ; from scipy.optimize import RCSTRINGS' ``` ### Error message ```shell <string>:1: DeprecationWarning: Please use `RCSTRINGS` from the `scipy.optimize` namespace, the `scipy.optimize.tnc` namespace is deprecated. Traceback (most recent call last): File "<string>", line 1, in <module> ImportError: cannot import name 'RCSTRINGS' from 'scipy.optimize' (C:\Program Files\Python310\lib\site-packages\scipy\optimize\__init__.py) ``` ### SciPy/NumPy/Python version information 1.9.3 1.24.1 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0)
non_process
bug deprecation warning says to use non existent symbols describe your issue about a year ago commit the optimize tnc module gained some deprecation warnings indicating that symbols of the form optimze tnc were deprecated use instead optimize since none of these symbols exist in optimize directly doing as suggested generates importerror reproducing code example python python c from scipy optimize tnc import rcstrings from scipy optimize import rcstrings error message shell deprecationwarning please use rcstrings from the scipy optimize namespace the scipy optimize tnc namespace is deprecated traceback most recent call last file line in importerror cannot import name rcstrings from scipy optimize c program files lib site packages scipy optimize init py scipy numpy python version information sys version info major minor micro releaselevel final serial
0
12,172
14,741,848,264
IssuesEvent
2021-01-07 11:16:24
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
News - Connection Timeout
anc-process anp-urgent ant-bug has attachment
In GitLab by @kdjstudios on Feb 18, 2019, 10:16 **Submitted by:** Kyle **Helpdesk:** - Deb: http://www.servicedesk.answernet.com/profiles/ticket/7152314 - Gaylan: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-21-72387/conversation - Richard and Gary: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-20-94418/conversation - Grant: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-20-18712/conversation **Server:** Internal **Client/Site:** NA **Account:** NA **Issue:** While attempting to update the News sections in SA Billing today. I kept encountering the something went wrong error when trying to save. I would simply navigate back or refresh the page and it appeared to update accordingly. After looking into my email I see that SAB was timing out. Please look into this, thank you. [SA_Billing_Error_Report_sab_news_index__ActionViewTemplateError__Connection_timed_out_-_user_specified_timeout.msg](/uploads/59cbc735f4dd87e730c608e6f4ef2e7d/SA_Billing_Error_Report_sab_news_index__ActionViewTemplateError__Connection_timed_out_-_user_specified_timeout.msg)
1.0
News - Connection Timeout - In GitLab by @kdjstudios on Feb 18, 2019, 10:16 **Submitted by:** Kyle **Helpdesk:** - Deb: http://www.servicedesk.answernet.com/profiles/ticket/7152314 - Gaylan: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-21-72387/conversation - Richard and Gary: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-20-94418/conversation - Grant: http://www.servicedesk.answernet.com/profiles/ticket/2019-02-20-18712/conversation **Server:** Internal **Client/Site:** NA **Account:** NA **Issue:** While attempting to update the News sections in SA Billing today. I kept encountering the something went wrong error when trying to save. I would simply navigate back or refresh the page and it appeared to update accordingly. After looking into my email I see that SAB was timing out. Please look into this, thank you. [SA_Billing_Error_Report_sab_news_index__ActionViewTemplateError__Connection_timed_out_-_user_specified_timeout.msg](/uploads/59cbc735f4dd87e730c608e6f4ef2e7d/SA_Billing_Error_Report_sab_news_index__ActionViewTemplateError__Connection_timed_out_-_user_specified_timeout.msg)
process
news connection timeout in gitlab by kdjstudios on feb submitted by kyle helpdesk deb gaylan richard and gary grant server internal client site na account na issue while attempting to update the news sections in sa billing today i kept encountering the something went wrong error when trying to save i would simply navigate back or refresh the page and it appeared to update accordingly after looking into my email i see that sab was timing out please look into this thank you uploads sa billing error report sab news index actionviewtemplateerror connection timed out user specified timeout msg
1
20,988
10,565,011,058
IssuesEvent
2019-10-05 07:35:04
plan-player-analytics/Plan
https://api.github.com/repos/plan-player-analytics/Plan
closed
XSS Attack possible with Bungee-Bukkit connection system [4.1.0 -> 4.8.7]
Security Vulnerability p: 25 server: Networks
### How to reproduce the vulnerability Version 4.5.2 - Find a BungeeCord server that has Plan on it - ~~Find a Bukkit/Sponge server UUID from the /debug page~~ Snoop on the HTTP traffic between Bungee and Bukkit servers for a server UUID - Send a valid CacheAnalysisPageRequest to /info/cacheanalysispagerequest that contains malicious html - Wait for the user to open the infected server page ### Mitigation steps (4.5.2) - Enable HTTPS on all servers that can be accessed from outside the machines ~~to limit access to /debug page~~ - (Extreme measure) Disable BungeeCord Plan webserver ### How to check if you are affected - Check BungeeCord /debug page connection log for any IPs that look out of place ### Impact Cross-site-scripting vulnerability to viewers of the server and player pages Affected requests: CacheAnalysisPageRequest, CachePlayerPluginsTabRequest Risk to users: LOW (Cross-server traffic hard to snoop) ### Underlying cause - Requests are authenticated using the server uuids ### Planned fixes - [x] Remove Server UUIDs from the debug page - [x] ~~Config setting to disable debug page (to be default disabled)~~ - [ ] Request numbering to all cross-server requests that is placed into the database before each request. This will allow more secure request validation.
True
XSS Attack possible with Bungee-Bukkit connection system [4.1.0 -> 4.8.7] - ### How to reproduce the vulnerability Version 4.5.2 - Find a BungeeCord server that has Plan on it - ~~Find a Bukkit/Sponge server UUID from the /debug page~~ Snoop on the HTTP traffic between Bungee and Bukkit servers for a server UUID - Send a valid CacheAnalysisPageRequest to /info/cacheanalysispagerequest that contains malicious html - Wait for the user to open the infected server page ### Mitigation steps (4.5.2) - Enable HTTPS on all servers that can be accessed from outside the machines ~~to limit access to /debug page~~ - (Extreme measure) Disable BungeeCord Plan webserver ### How to check if you are affected - Check BungeeCord /debug page connection log for any IPs that look out of place ### Impact Cross-site-scripting vulnerability to viewers of the server and player pages Affected requests: CacheAnalysisPageRequest, CachePlayerPluginsTabRequest Risk to users: LOW (Cross-server traffic hard to snoop) ### Underlying cause - Requests are authenticated using the server uuids ### Planned fixes - [x] Remove Server UUIDs from the debug page - [x] ~~Config setting to disable debug page (to be default disabled)~~ - [ ] Request numbering to all cross-server requests that is placed into the database before each request. This will allow more secure request validation.
non_process
xss attack possible with bungee bukkit connection system how to reproduce the vulnerability version find a bungeecord server that has plan on it find a bukkit sponge server uuid from the debug page snoop on the http traffic between bungee and bukkit servers for a server uuid send a valid cacheanalysispagerequest to info cacheanalysispagerequest that contains malicious html wait for the user to open the infected server page mitigation steps enable https on all servers that can be accessed from outside the machines to limit access to debug page extreme measure disable bungeecord plan webserver how to check if you are affected check bungeecord debug page connection log for any ips that look out of place impact cross site scripting vulnerability to viewers of the server and player pages affected requests cacheanalysispagerequest cacheplayerpluginstabrequest risk to users low cross server traffic hard to snoop underlying cause requests are authenticated using the server uuids planned fixes remove server uuids from the debug page config setting to disable debug page to be default disabled request numbering to all cross server requests that is placed into the database before each request this will allow more secure request validation
0
113,132
14,370,337,201
IssuesEvent
2020-12-01 10:57:57
openaustralia/righttoknow
https://api.github.com/repos/openaustralia/righttoknow
closed
Version 0.31.0.4
design enhancement
# 0.31.0.4 ## Highlighted Features * Updated translations from Transifex (Liz Conlan) # 0.31.0.3 ## Highlighted Features * Fix broken translation string (Gareth Rees) # 0.31.0.2 ## Highlighted Features * Remove obsolete pro msgids (Gareth Rees) # 0.31.0.1 ## Highlighted Features * Updated translations from Transifex (Gareth Rees) # 0.31.0.0 ## Highlighted Features * Rescue from POP poller timeouts (Graeme Porteous) * Fixed issue where an attempted password reset with cookies disabled caused a redirect loop (Graeme Porteous) * Improved user drop down positioning (Martin Wright) * Fixed an issue where selectize was blocking other scripts from running (Graeme Porteous) * Dropped support for Ruby 1.9 (Liz Conlan) * Default to Ruby 2.x for all install scripts (Liz Conlan) * Removed support for Debian Wheezy (Liz Conlan) * Add Debian Stretch support (Louise Crow, Gareth Rees) * Replace out of support zip gem with rubyzip to address an issue where some zip files created by the site may not be openable on Windows PCs (Liz Conlan, Graeme Porteous) * Fix bug in `stats:show` task (Liz Conlan, Gareth Rees) * Use `.eml` file extension when downloading raw emails through the admin interface (Gareth Rees) * Reduce usage of auto-login links in emails (Gareth Rees) * Remove rendering of exceptions in admin interface (Gareth Rees) * Pass through sign-in form if a user is already signed in (Gareth Rees) * Make the event history table responsive (Miroslav Schlossberg) * Fix bug that prevented private requests from being published across the whole site once the embargo period had expired (Liz Conlan) * Update format of `robots.txt` for Baidu compatibility (Gareth Rees) * Removed support for Ubuntu Precise (Louise Crow) * Remove the use of purge requests to Varnish (Louise Crow) * Add a temp task to recache any attachments whose content has changed (Louise Crow) ## Upgrade Notes * This release drops support for Ruby 1.9.x. If you are using Ubuntu Trusty you will need to install a newer ruby version either using a ruby environment manager like [rbenv](https://github.com/rbenv/rbenv#basic-github-checkout) or by installing the ruby2.1 (and ruby2.1-dev) or ruby2.3 (and ruby2.3-dev) [Ubuntu packages from Brightbox](https://www.brightbox.com/docs/ruby/ubuntu/). (If you are setting up a fresh Trusty box using our script, the 2.1 Brightbox package is supplied.) * Please note that this release removes support for Ubuntu Precise as it has reached End of Life and will no longer receive security patches. If you are running Alaveteli on Ubuntu Precise, you should upgrade your OS to Ubuntu Trusty before upgrading to this release. This [Ubuntu upgrade guide](https://wiki.ubuntu.com/TrustyTahr/ReleaseNotes#Upgrading_from_Ubuntu_12.04_LTS_or_Ubuntu_13.10) can guide you through the process. If you have questions about upgrading OS, please don't hesitate to ask on the [alaveteli-dev](https://groups.google.com/forum/#!forum/alaveteli-dev) group. * Please note that this release also removes support for Debian Wheezy as it only packages Ruby 1.9.3. If you are running Alaveteli on Debian Wheezy, you should upgrade your OS to Debian Jessie before upgrading to this release. This [Debian upgrade guide](https://www.debian.org/releases/jessie/amd64/release-notes/ch-upgrading.en.html) can guide you through the process. If you have questions about upgrading OS, please don't hesitate to ask on the [alaveteli-dev](https://groups.google.com/forum/#!forum/alaveteli-dev) group. * This release removes the use of purge requests to Varnish. Please make sure your site works with `VARNISH_HOST` empty before upgrading. * There's a new temp task that can be used to recache any attachments whose content has slightly changed (e.g. due to an upgrade in the `mail` gem that alters e.g the trailing space on attachment bodies). You can run it with `bundle exec rake temp:populate_missing_attachment_files` if you're seeing `No such file or directory @ rb_sysopen` errors from `foi_attachment.rb`. * There are some database structure updates so remember to `rake db:migrate` ### Changed Templates app/views/admin_general/_to_do_list.html.erb app/views/admin_general/index.html.erb app/views/admin_raw_email/show.html.erb app/views/admin_request/show.html.erb app/views/admin_user/show.html.erb app/views/alaveteli_pro/account_request/new.html.erb app/views/alaveteli_pro/batch_request_authority_searches/index.html.erb app/views/alaveteli_pro/dashboard/_projects.html.erb app/views/alaveteli_pro/dashboard/index.html.erb app/views/alaveteli_pro/general/_log_in_bar_links.html.erb app/views/alaveteli_pro/info_request_batches/_embargo_form.html.erb app/views/alaveteli_pro/info_request_batches/_embargo_info.html.erb app/views/alaveteli_pro/info_request_batches/_form.html.erb app/views/alaveteli_pro/info_request_batches/_info_request.html.erb app/views/alaveteli_pro/info_requests/_embargo_form.html.erb app/views/alaveteli_pro/info_requests/_embargo_info.html.erb app/views/alaveteli_pro/info_requests/_form.html.erb app/views/alaveteli_pro/info_requests/_info_request.html.erb app/views/alaveteli_pro/info_requests/_new_request_advice.html.erb app/views/alaveteli_pro/info_requests/_select_authority_form.html.erb app/views/alaveteli_pro/info_requests/_sidebar.html.erb app/views/alaveteli_pro/info_requests/index.html.erb app/views/alaveteli_pro/info_requests/new.html.erb app/views/alaveteli_pro/info_requests/preview.html.erb app/views/general/_frontpage_bodies_list.html.erb app/views/general/_frontpage_requests_list.html.erb app/views/general/_log_in_bar.html.erb app/views/general/_responsive_footer.html.erb app/views/general/exception_caught.html.erb app/views/notification_mailer/info_request_batches/messages/_overdue.text.erb app/views/notification_mailer/info_request_batches/messages/_very_overdue.text.erb app/views/notification_mailer/info_requests/messages/_overdue.text.erb app/views/notification_mailer/info_requests/messages/_very_overdue.text.erb app/views/outgoing_mailer/_followup_footer.text.erb app/views/password_changes/new.html.erb app/views/request/_act.html.erb app/views/request/_request_listing_single.html.erb app/views/request/_request_listing_via_event.html.erb app/views/request/_request_sent.html.erb app/views/request/_sidebar.html.erb app/views/request/details.html.erb app/views/request/show.html.erb app/views/track/_rss_feed.html.erb app/views/user/_signin.html.erb app/views/user/_signup.html.erb app/views/user/show.html.erb app/views/user/sign.html.erb app/views/users/sessions/show.html.erb
1.0
Version 0.31.0.4 - # 0.31.0.4 ## Highlighted Features * Updated translations from Transifex (Liz Conlan) # 0.31.0.3 ## Highlighted Features * Fix broken translation string (Gareth Rees) # 0.31.0.2 ## Highlighted Features * Remove obsolete pro msgids (Gareth Rees) # 0.31.0.1 ## Highlighted Features * Updated translations from Transifex (Gareth Rees) # 0.31.0.0 ## Highlighted Features * Rescue from POP poller timeouts (Graeme Porteous) * Fixed issue where an attempted password reset with cookies disabled caused a redirect loop (Graeme Porteous) * Improved user drop down positioning (Martin Wright) * Fixed an issue where selectize was blocking other scripts from running (Graeme Porteous) * Dropped support for Ruby 1.9 (Liz Conlan) * Default to Ruby 2.x for all install scripts (Liz Conlan) * Removed support for Debian Wheezy (Liz Conlan) * Add Debian Stretch support (Louise Crow, Gareth Rees) * Replace out of support zip gem with rubyzip to address an issue where some zip files created by the site may not be openable on Windows PCs (Liz Conlan, Graeme Porteous) * Fix bug in `stats:show` task (Liz Conlan, Gareth Rees) * Use `.eml` file extension when downloading raw emails through the admin interface (Gareth Rees) * Reduce usage of auto-login links in emails (Gareth Rees) * Remove rendering of exceptions in admin interface (Gareth Rees) * Pass through sign-in form if a user is already signed in (Gareth Rees) * Make the event history table responsive (Miroslav Schlossberg) * Fix bug that prevented private requests from being published across the whole site once the embargo period had expired (Liz Conlan) * Update format of `robots.txt` for Baidu compatibility (Gareth Rees) * Removed support for Ubuntu Precise (Louise Crow) * Remove the use of purge requests to Varnish (Louise Crow) * Add a temp task to recache any attachments whose content has changed (Louise Crow) ## Upgrade Notes * This release drops support for Ruby 1.9.x. If you are using Ubuntu Trusty you will need to install a newer ruby version either using a ruby environment manager like [rbenv](https://github.com/rbenv/rbenv#basic-github-checkout) or by installing the ruby2.1 (and ruby2.1-dev) or ruby2.3 (and ruby2.3-dev) [Ubuntu packages from Brightbox](https://www.brightbox.com/docs/ruby/ubuntu/). (If you are setting up a fresh Trusty box using our script, the 2.1 Brightbox package is supplied.) * Please note that this release removes support for Ubuntu Precise as it has reached End of Life and will no longer receive security patches. If you are running Alaveteli on Ubuntu Precise, you should upgrade your OS to Ubuntu Trusty before upgrading to this release. This [Ubuntu upgrade guide](https://wiki.ubuntu.com/TrustyTahr/ReleaseNotes#Upgrading_from_Ubuntu_12.04_LTS_or_Ubuntu_13.10) can guide you through the process. If you have questions about upgrading OS, please don't hesitate to ask on the [alaveteli-dev](https://groups.google.com/forum/#!forum/alaveteli-dev) group. * Please note that this release also removes support for Debian Wheezy as it only packages Ruby 1.9.3. If you are running Alaveteli on Debian Wheezy, you should upgrade your OS to Debian Jessie before upgrading to this release. This [Debian upgrade guide](https://www.debian.org/releases/jessie/amd64/release-notes/ch-upgrading.en.html) can guide you through the process. If you have questions about upgrading OS, please don't hesitate to ask on the [alaveteli-dev](https://groups.google.com/forum/#!forum/alaveteli-dev) group. * This release removes the use of purge requests to Varnish. Please make sure your site works with `VARNISH_HOST` empty before upgrading. * There's a new temp task that can be used to recache any attachments whose content has slightly changed (e.g. due to an upgrade in the `mail` gem that alters e.g the trailing space on attachment bodies). You can run it with `bundle exec rake temp:populate_missing_attachment_files` if you're seeing `No such file or directory @ rb_sysopen` errors from `foi_attachment.rb`. * There are some database structure updates so remember to `rake db:migrate` ### Changed Templates app/views/admin_general/_to_do_list.html.erb app/views/admin_general/index.html.erb app/views/admin_raw_email/show.html.erb app/views/admin_request/show.html.erb app/views/admin_user/show.html.erb app/views/alaveteli_pro/account_request/new.html.erb app/views/alaveteli_pro/batch_request_authority_searches/index.html.erb app/views/alaveteli_pro/dashboard/_projects.html.erb app/views/alaveteli_pro/dashboard/index.html.erb app/views/alaveteli_pro/general/_log_in_bar_links.html.erb app/views/alaveteli_pro/info_request_batches/_embargo_form.html.erb app/views/alaveteli_pro/info_request_batches/_embargo_info.html.erb app/views/alaveteli_pro/info_request_batches/_form.html.erb app/views/alaveteli_pro/info_request_batches/_info_request.html.erb app/views/alaveteli_pro/info_requests/_embargo_form.html.erb app/views/alaveteli_pro/info_requests/_embargo_info.html.erb app/views/alaveteli_pro/info_requests/_form.html.erb app/views/alaveteli_pro/info_requests/_info_request.html.erb app/views/alaveteli_pro/info_requests/_new_request_advice.html.erb app/views/alaveteli_pro/info_requests/_select_authority_form.html.erb app/views/alaveteli_pro/info_requests/_sidebar.html.erb app/views/alaveteli_pro/info_requests/index.html.erb app/views/alaveteli_pro/info_requests/new.html.erb app/views/alaveteli_pro/info_requests/preview.html.erb app/views/general/_frontpage_bodies_list.html.erb app/views/general/_frontpage_requests_list.html.erb app/views/general/_log_in_bar.html.erb app/views/general/_responsive_footer.html.erb app/views/general/exception_caught.html.erb app/views/notification_mailer/info_request_batches/messages/_overdue.text.erb app/views/notification_mailer/info_request_batches/messages/_very_overdue.text.erb app/views/notification_mailer/info_requests/messages/_overdue.text.erb app/views/notification_mailer/info_requests/messages/_very_overdue.text.erb app/views/outgoing_mailer/_followup_footer.text.erb app/views/password_changes/new.html.erb app/views/request/_act.html.erb app/views/request/_request_listing_single.html.erb app/views/request/_request_listing_via_event.html.erb app/views/request/_request_sent.html.erb app/views/request/_sidebar.html.erb app/views/request/details.html.erb app/views/request/show.html.erb app/views/track/_rss_feed.html.erb app/views/user/_signin.html.erb app/views/user/_signup.html.erb app/views/user/show.html.erb app/views/user/sign.html.erb app/views/users/sessions/show.html.erb
non_process
version highlighted features updated translations from transifex liz conlan highlighted features fix broken translation string gareth rees highlighted features remove obsolete pro msgids gareth rees highlighted features updated translations from transifex gareth rees highlighted features rescue from pop poller timeouts graeme porteous fixed issue where an attempted password reset with cookies disabled caused a redirect loop graeme porteous improved user drop down positioning martin wright fixed an issue where selectize was blocking other scripts from running graeme porteous dropped support for ruby liz conlan default to ruby x for all install scripts liz conlan removed support for debian wheezy liz conlan add debian stretch support louise crow gareth rees replace out of support zip gem with rubyzip to address an issue where some zip files created by the site may not be openable on windows pcs liz conlan graeme porteous fix bug in stats show task liz conlan gareth rees use eml file extension when downloading raw emails through the admin interface gareth rees reduce usage of auto login links in emails gareth rees remove rendering of exceptions in admin interface gareth rees pass through sign in form if a user is already signed in gareth rees make the event history table responsive miroslav schlossberg fix bug that prevented private requests from being published across the whole site once the embargo period had expired liz conlan update format of robots txt for baidu compatibility gareth rees removed support for ubuntu precise louise crow remove the use of purge requests to varnish louise crow add a temp task to recache any attachments whose content has changed louise crow upgrade notes this release drops support for ruby x if you are using ubuntu trusty you will need to install a newer ruby version either using a ruby environment manager like or by installing the and dev or and dev if you are setting up a fresh trusty box using our script the brightbox package is supplied please note that this release removes support for ubuntu precise as it has reached end of life and will no longer receive security patches if you are running alaveteli on ubuntu precise you should upgrade your os to ubuntu trusty before upgrading to this release this can guide you through the process if you have questions about upgrading os please don t hesitate to ask on the group please note that this release also removes support for debian wheezy as it only packages ruby if you are running alaveteli on debian wheezy you should upgrade your os to debian jessie before upgrading to this release this can guide you through the process if you have questions about upgrading os please don t hesitate to ask on the group this release removes the use of purge requests to varnish please make sure your site works with varnish host empty before upgrading there s a new temp task that can be used to recache any attachments whose content has slightly changed e g due to an upgrade in the mail gem that alters e g the trailing space on attachment bodies you can run it with bundle exec rake temp populate missing attachment files if you re seeing no such file or directory rb sysopen errors from foi attachment rb there are some database structure updates so remember to rake db migrate changed templates app views admin general to do list html erb app views admin general index html erb app views admin raw email show html erb app views admin request show html erb app views admin user show html erb app views alaveteli pro account request new html erb app views alaveteli pro batch request authority searches index html erb app views alaveteli pro dashboard projects html erb app views alaveteli pro dashboard index html erb app views alaveteli pro general log in bar links html erb app views alaveteli pro info request batches embargo form html erb app views alaveteli pro info request batches embargo info html erb app views alaveteli pro info request batches form html erb app views alaveteli pro info request batches info request html erb app views alaveteli pro info requests embargo form html erb app views alaveteli pro info requests embargo info html erb app views alaveteli pro info requests form html erb app views alaveteli pro info requests info request html erb app views alaveteli pro info requests new request advice html erb app views alaveteli pro info requests select authority form html erb app views alaveteli pro info requests sidebar html erb app views alaveteli pro info requests index html erb app views alaveteli pro info requests new html erb app views alaveteli pro info requests preview html erb app views general frontpage bodies list html erb app views general frontpage requests list html erb app views general log in bar html erb app views general responsive footer html erb app views general exception caught html erb app views notification mailer info request batches messages overdue text erb app views notification mailer info request batches messages very overdue text erb app views notification mailer info requests messages overdue text erb app views notification mailer info requests messages very overdue text erb app views outgoing mailer followup footer text erb app views password changes new html erb app views request act html erb app views request request listing single html erb app views request request listing via event html erb app views request request sent html erb app views request sidebar html erb app views request details html erb app views request show html erb app views track rss feed html erb app views user signin html erb app views user signup html erb app views user show html erb app views user sign html erb app views users sessions show html erb
0
19,064
25,083,667,851
IssuesEvent
2022-11-07 21:37:39
googleapis/gaxios
https://api.github.com/repos/googleapis/gaxios
closed
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * must have required property 'library_type' in .repo-metadata.json * must have required property 'release_level' in .repo-metadata.json * must have required property 'client_documentation' in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * must have required property 'library_type' in .repo-metadata.json * must have required property 'release_level' in .repo-metadata.json * must have required property 'client_documentation' in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 must have required property library type in repo metadata json must have required property release level in repo metadata json must have required property client documentation in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
7,855
11,029,633,772
IssuesEvent
2019-12-06 14:17:14
prisma/photonjs
https://api.github.com/repos/prisma/photonjs
closed
UniqueConstraintViolation error data as JSON object
kind/improvement process/next-milestone
Hello! Do you have any plans to support returning the UniqueConstraintViolation error data as JSON object which contains the name of the field with duplicate value? At this moment this info is returned as a string which I need to parse.
1.0
UniqueConstraintViolation error data as JSON object - Hello! Do you have any plans to support returning the UniqueConstraintViolation error data as JSON object which contains the name of the field with duplicate value? At this moment this info is returned as a string which I need to parse.
process
uniqueconstraintviolation error data as json object hello do you have any plans to support returning the uniqueconstraintviolation error data as json object which contains the name of the field with duplicate value at this moment this info is returned as a string which i need to parse
1
765,531
26,850,545,770
IssuesEvent
2023-02-03 10:40:41
kartoza/ckanext-dalrrd-emc-dcpr
https://api.github.com/repos/kartoza/ckanext-dalrrd-emc-dcpr
opened
Make json the default on the server
priority:low
The current EMC STAC API root endpoint is https://csw-testing.emc.kartoza.com/, but accessing this returns an html page which isn't readable in the STAC Browser QGIS plugin. Appending ?f=json to the endpoint to make it https://csw-testing.emc.kartoza.com/?f=json will return a JSON response that the plugin can work with.
1.0
Make json the default on the server - The current EMC STAC API root endpoint is https://csw-testing.emc.kartoza.com/, but accessing this returns an html page which isn't readable in the STAC Browser QGIS plugin. Appending ?f=json to the endpoint to make it https://csw-testing.emc.kartoza.com/?f=json will return a JSON response that the plugin can work with.
non_process
make json the default on the server the current emc stac api root endpoint is but accessing this returns an html page which isn t readable in the stac browser qgis plugin appending f json to the endpoint to make it will return a json response that the plugin can work with
0
89,450
15,829,054,201
IssuesEvent
2021-04-06 10:40:10
bci-oss/keycloak
https://api.github.com/repos/bci-oss/keycloak
opened
WS-2016-7057 (Medium) detected in plexus-utils-3.0.20.jar
security vulnerability
## WS-2016-7057 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>plexus-utils-3.0.20.jar</b></p></summary> <p>A collection of various utility classes to ease working with strings, files, command lines, XML and more.</p> <p>Path to dependency file: keycloak/testsuite/model/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/plexus/plexus-utils/3.0.20/plexus-utils-3.0.20.jar</p> <p> Dependency Hierarchy: - integration-arquillian-tests-base-13.0.0-SNAPSHOT.jar (Root Library) - arquillian-phantom-driver-1.2.1.Final.jar - shrinkwrap-resolver-depchain-2.2.2.pom - shrinkwrap-resolver-impl-maven-2.2.2.jar - :x: **plexus-utils-3.0.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bci-oss/keycloak/commit/4ece64b821b25688d9fbfacd091a88584ffc90e0">4ece64b821b25688d9fbfacd091a88584ffc90e0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Plexus-utils before 3.0.24 are vulnerable to Directory Traversal <p>Publish Date: 2016-05-07 <p>URL: <a href=https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef>WS-2016-7057</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef">https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef</a></p> <p>Release Date: 2019-05-30</p> <p>Fix Resolution: 3.0.24</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.codehaus.plexus","packageName":"plexus-utils","packageVersion":"3.0.20","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak.testsuite:integration-arquillian-tests-base:13.0.0-SNAPSHOT;org.jboss.arquillian.extension:arquillian-phantom-driver:1.2.1.Final;org.jboss.shrinkwrap.resolver:shrinkwrap-resolver-depchain:2.2.2;org.jboss.shrinkwrap.resolver:shrinkwrap-resolver-impl-maven:2.2.2;org.codehaus.plexus:plexus-utils:3.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.24"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7057","vulnerabilityDetails":"Plexus-utils before 3.0.24 are vulnerable to Directory Traversal","vulnerabilityUrl":"https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2016-7057 (Medium) detected in plexus-utils-3.0.20.jar - ## WS-2016-7057 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>plexus-utils-3.0.20.jar</b></p></summary> <p>A collection of various utility classes to ease working with strings, files, command lines, XML and more.</p> <p>Path to dependency file: keycloak/testsuite/model/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/codehaus/plexus/plexus-utils/3.0.20/plexus-utils-3.0.20.jar</p> <p> Dependency Hierarchy: - integration-arquillian-tests-base-13.0.0-SNAPSHOT.jar (Root Library) - arquillian-phantom-driver-1.2.1.Final.jar - shrinkwrap-resolver-depchain-2.2.2.pom - shrinkwrap-resolver-impl-maven-2.2.2.jar - :x: **plexus-utils-3.0.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bci-oss/keycloak/commit/4ece64b821b25688d9fbfacd091a88584ffc90e0">4ece64b821b25688d9fbfacd091a88584ffc90e0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Plexus-utils before 3.0.24 are vulnerable to Directory Traversal <p>Publish Date: 2016-05-07 <p>URL: <a href=https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef>WS-2016-7057</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef">https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef</a></p> <p>Release Date: 2019-05-30</p> <p>Fix Resolution: 3.0.24</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.codehaus.plexus","packageName":"plexus-utils","packageVersion":"3.0.20","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak.testsuite:integration-arquillian-tests-base:13.0.0-SNAPSHOT;org.jboss.arquillian.extension:arquillian-phantom-driver:1.2.1.Final;org.jboss.shrinkwrap.resolver:shrinkwrap-resolver-depchain:2.2.2;org.jboss.shrinkwrap.resolver:shrinkwrap-resolver-impl-maven:2.2.2;org.codehaus.plexus:plexus-utils:3.0.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.24"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7057","vulnerabilityDetails":"Plexus-utils before 3.0.24 are vulnerable to Directory Traversal","vulnerabilityUrl":"https://github.com/codehaus-plexus/plexus-utils/commit/33a2853df8185b4519b1b8bfae284f03392618ef","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in plexus utils jar ws medium severity vulnerability vulnerable library plexus utils jar a collection of various utility classes to ease working with strings files command lines xml and more path to dependency file keycloak testsuite model pom xml path to vulnerable library home wss scanner repository org codehaus plexus plexus utils plexus utils jar dependency hierarchy integration arquillian tests base snapshot jar root library arquillian phantom driver final jar shrinkwrap resolver depchain pom shrinkwrap resolver impl maven jar x plexus utils jar vulnerable library found in head commit a href found in base branch master vulnerability details plexus utils before are vulnerable to directory traversal publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org keycloak testsuite integration arquillian tests base snapshot org jboss arquillian extension arquillian phantom driver final org jboss shrinkwrap resolver shrinkwrap resolver depchain org jboss shrinkwrap resolver shrinkwrap resolver impl maven org codehaus plexus plexus utils isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails plexus utils before are vulnerable to directory traversal vulnerabilityurl
0
7,838
11,012,109,158
IssuesEvent
2019-12-04 17:33:39
90301/TextReplace
https://api.github.com/repos/90301/TextReplace
opened
XPath based operations
Log Processor
need support for treating a document as XML, and being able to do XPath based search and possibly XPath based replacement.
1.0
XPath based operations - need support for treating a document as XML, and being able to do XPath based search and possibly XPath based replacement.
process
xpath based operations need support for treating a document as xml and being able to do xpath based search and possibly xpath based replacement
1
20,680
27,352,721,486
IssuesEvent
2023-02-27 10:41:37
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Ability to suppress warnings by type (or just experimental warnings)
feature request process
**Is your feature request related to a problem? Please describe.** I'd like to suppress experimental warnings while still seeing any other errors. In particular when I am using native ES modules I do not want the experimental warning printed for every process, but I do want unrelated warnings to still be printed. **Describe the solution you'd like** Allow `--no-warnings` to optionally accept an option string such as `--no-warnings=type1,type2`. Using `--no-warnings` without any option would continue to disable all warnings. This would allow `--no-warnings=ExperimentalWarning` to suppress ExperimentalWarning only. **Describe alternatives you've considered** `--no-experimental-warnings` or a similarly named new flag could be created. This has the drawback that `node --no-experimental-warnings` on node.js 13.3.0 exit with an error where `--no-warnings=ExperimentalWarnings` will not currently error (it causes all warnings to be ignored). In my own repo which uses ES modules I've created `suppress-experimental.cjs` which gets loaded with `NODE_OPTIONS='--require=./suppress-experimental.cjs'`: ```js 'use strict'; const {emitWarning} = process; process.emitWarning = (warning, ...args) => { if (args[0] === 'ExperimentalWarning') { return; } if (args[0] && typeof args[0] === 'object' && args[0].type === 'ExperimentalWarning') { return; } return emitWarning(warning, ...args); }; ``` Obviously patching node.js internals like this is undesirable but it accomplishes my goal.
1.0
Ability to suppress warnings by type (or just experimental warnings) - **Is your feature request related to a problem? Please describe.** I'd like to suppress experimental warnings while still seeing any other errors. In particular when I am using native ES modules I do not want the experimental warning printed for every process, but I do want unrelated warnings to still be printed. **Describe the solution you'd like** Allow `--no-warnings` to optionally accept an option string such as `--no-warnings=type1,type2`. Using `--no-warnings` without any option would continue to disable all warnings. This would allow `--no-warnings=ExperimentalWarning` to suppress ExperimentalWarning only. **Describe alternatives you've considered** `--no-experimental-warnings` or a similarly named new flag could be created. This has the drawback that `node --no-experimental-warnings` on node.js 13.3.0 exit with an error where `--no-warnings=ExperimentalWarnings` will not currently error (it causes all warnings to be ignored). In my own repo which uses ES modules I've created `suppress-experimental.cjs` which gets loaded with `NODE_OPTIONS='--require=./suppress-experimental.cjs'`: ```js 'use strict'; const {emitWarning} = process; process.emitWarning = (warning, ...args) => { if (args[0] === 'ExperimentalWarning') { return; } if (args[0] && typeof args[0] === 'object' && args[0].type === 'ExperimentalWarning') { return; } return emitWarning(warning, ...args); }; ``` Obviously patching node.js internals like this is undesirable but it accomplishes my goal.
process
ability to suppress warnings by type or just experimental warnings is your feature request related to a problem please describe i d like to suppress experimental warnings while still seeing any other errors in particular when i am using native es modules i do not want the experimental warning printed for every process but i do want unrelated warnings to still be printed describe the solution you d like allow no warnings to optionally accept an option string such as no warnings using no warnings without any option would continue to disable all warnings this would allow no warnings experimentalwarning to suppress experimentalwarning only describe alternatives you ve considered no experimental warnings or a similarly named new flag could be created this has the drawback that node no experimental warnings on node js exit with an error where no warnings experimentalwarnings will not currently error it causes all warnings to be ignored in my own repo which uses es modules i ve created suppress experimental cjs which gets loaded with node options require suppress experimental cjs js use strict const emitwarning process process emitwarning warning args if args experimentalwarning return if args typeof args object args type experimentalwarning return return emitwarning warning args obviously patching node js internals like this is undesirable but it accomplishes my goal
1
292,904
25,249,659,085
IssuesEvent
2022-11-15 13:49:18
cloudquery/cloudquery
https://api.github.com/repos/cloudquery/cloudquery
closed
digitalocean plugin is broken
bug digitalocean integration-test-gap
### Describe the Bug Panics, and empty database after `cloudquery sync`. ``` export DIGITALOCEAN_ACCESS_TOKEN="<...>" ~/projects/cqdo % cloudquery sync . Loading spec(s) from . Starting migration for: digitalocean -> [postgresql] Migration completed successfully. Starting sync for: digitalocean -> [postgresql] Sync completed successfully. Summary: resources: 493, errors: 2, panic: 2, failed_writes: 0, time: 8s ``` ### CloudQuery Version `cloudquery: v1.4.0` `digitalocean: v2.1.0` `postgresql: v1.3.11` ### Steps to Reproduce `digitalocean.yml`: ```yml kind: source spec: name: digitalocean path: cloudquery/digitalocean version: "v2.1.0" # latest version of digitalocean plugin tables: ["*"] destinations: ["postgresql"] spec: ```
1.0
digitalocean plugin is broken - ### Describe the Bug Panics, and empty database after `cloudquery sync`. ``` export DIGITALOCEAN_ACCESS_TOKEN="<...>" ~/projects/cqdo % cloudquery sync . Loading spec(s) from . Starting migration for: digitalocean -> [postgresql] Migration completed successfully. Starting sync for: digitalocean -> [postgresql] Sync completed successfully. Summary: resources: 493, errors: 2, panic: 2, failed_writes: 0, time: 8s ``` ### CloudQuery Version `cloudquery: v1.4.0` `digitalocean: v2.1.0` `postgresql: v1.3.11` ### Steps to Reproduce `digitalocean.yml`: ```yml kind: source spec: name: digitalocean path: cloudquery/digitalocean version: "v2.1.0" # latest version of digitalocean plugin tables: ["*"] destinations: ["postgresql"] spec: ```
non_process
digitalocean plugin is broken describe the bug panics and empty database after cloudquery sync export digitalocean access token projects cqdo cloudquery sync loading spec s from starting migration for digitalocean migration completed successfully starting sync for digitalocean sync completed successfully summary resources errors panic failed writes time cloudquery version cloudquery digitalocean postgresql steps to reproduce digitalocean yml yml kind source spec name digitalocean path cloudquery digitalocean version latest version of digitalocean plugin tables destinations spec
0
48,302
13,325,308,022
IssuesEvent
2020-08-27 09:45:59
solidify/fitbit-api-demo
https://api.github.com/repos/solidify/fitbit-api-demo
opened
CVE-2019-14892 (High) detected in jackson-databind-2.8.1.jar
security vulnerability
## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/fitbit-api-demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/solidify/fitbit-api-demo/commit/0b2ec4ebe69f408230a396984bb6ebaef3b6b3ab">0b2ec4ebe69f408230a396984bb6ebaef3b6b3ab</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14892 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/fitbit-api-demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/solidify/fitbit-api-demo/commit/0b2ec4ebe69f408230a396984bb6ebaef3b6b3ab">0b2ec4ebe69f408230a396984bb6ebaef3b6b3ab</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm fitbit api demo pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
17,221
22,831,775,221
IssuesEvent
2022-07-12 13:31:15
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/transform] Improve processor's ability to handle enums
priority:p2 processor/transform
The transform processor currently has messy support for fields that are enums in the protobufs. Some fields, like `metric.aggregation_temporality`, use the numeric representation in some functions but a string representation in others. Other fields are accessed using pdata representation that don't allow comparison with the query language's parse fields. For example, the where clause `... where kind == 2` does not work because `kind` is returned for comparison as an int32 but `2` is interpreted by the parser as an `int64` The processor as a whole needs a solution for handling enums. Some enums, like `severity_number`, should be able to be interacted with via the numeric representation so that comparisons like `<` and `>` can be used, but should also be able to be represented as a string for readability. Others enums, such as `aggregation_temporality` might only ever needs represented as a string. All fields should be usable in a condition. At the moment comparisons's of literal values in a condition are always string, int64, or float64, so we need a way to access enums so that they are comparable with at least one of these types. I've reviewed the access functions for each signal and identified this fields as needing dealt with **Traces** - `kind`: Currently accessed as an int32 so does not work in conditions. Should it be accessible as a string? - `status.code`: Currently accessed as an int32 so does not work in conditions. Should it be accessible as a string? **Metrics** - `metric.type`: Currently accessed as a string. This field isn't a enum in the protos but is represented as a numeric in pdata so adding it here to be reviewed. - `metric.aggregation_temporality`: Accessed as a int64, but that makes it hard to interpret queries. Should be allowed to use a string representation ("unspecified, "delta", and "cumulative"). - `flags`: Accessed as a uint32 so doesn't work in comparison. Should it be accessible as a string? **Logs** - `severity_number`: Accessed as an int32 so doesn't work in conditions. Should it be accessible as a string? **Additional context** There are other fields that have access bugs when being used in a condition, but I will open a separate issue for those.
1.0
[processor/transform] Improve processor's ability to handle enums - The transform processor currently has messy support for fields that are enums in the protobufs. Some fields, like `metric.aggregation_temporality`, use the numeric representation in some functions but a string representation in others. Other fields are accessed using pdata representation that don't allow comparison with the query language's parse fields. For example, the where clause `... where kind == 2` does not work because `kind` is returned for comparison as an int32 but `2` is interpreted by the parser as an `int64` The processor as a whole needs a solution for handling enums. Some enums, like `severity_number`, should be able to be interacted with via the numeric representation so that comparisons like `<` and `>` can be used, but should also be able to be represented as a string for readability. Others enums, such as `aggregation_temporality` might only ever needs represented as a string. All fields should be usable in a condition. At the moment comparisons's of literal values in a condition are always string, int64, or float64, so we need a way to access enums so that they are comparable with at least one of these types. I've reviewed the access functions for each signal and identified this fields as needing dealt with **Traces** - `kind`: Currently accessed as an int32 so does not work in conditions. Should it be accessible as a string? - `status.code`: Currently accessed as an int32 so does not work in conditions. Should it be accessible as a string? **Metrics** - `metric.type`: Currently accessed as a string. This field isn't a enum in the protos but is represented as a numeric in pdata so adding it here to be reviewed. - `metric.aggregation_temporality`: Accessed as a int64, but that makes it hard to interpret queries. Should be allowed to use a string representation ("unspecified, "delta", and "cumulative"). - `flags`: Accessed as a uint32 so doesn't work in comparison. Should it be accessible as a string? **Logs** - `severity_number`: Accessed as an int32 so doesn't work in conditions. Should it be accessible as a string? **Additional context** There are other fields that have access bugs when being used in a condition, but I will open a separate issue for those.
process
improve processor s ability to handle enums the transform processor currently has messy support for fields that are enums in the protobufs some fields like metric aggregation temporality use the numeric representation in some functions but a string representation in others other fields are accessed using pdata representation that don t allow comparison with the query language s parse fields for example the where clause where kind does not work because kind is returned for comparison as an but is interpreted by the parser as an the processor as a whole needs a solution for handling enums some enums like severity number should be able to be interacted with via the numeric representation so that comparisons like can be used but should also be able to be represented as a string for readability others enums such as aggregation temporality might only ever needs represented as a string all fields should be usable in a condition at the moment comparisons s of literal values in a condition are always string or so we need a way to access enums so that they are comparable with at least one of these types i ve reviewed the access functions for each signal and identified this fields as needing dealt with traces kind currently accessed as an so does not work in conditions should it be accessible as a string status code currently accessed as an so does not work in conditions should it be accessible as a string metrics metric type currently accessed as a string this field isn t a enum in the protos but is represented as a numeric in pdata so adding it here to be reviewed metric aggregation temporality accessed as a but that makes it hard to interpret queries should be allowed to use a string representation unspecified delta and cumulative flags accessed as a so doesn t work in comparison should it be accessible as a string logs severity number accessed as an so doesn t work in conditions should it be accessible as a string additional context there are other fields that have access bugs when being used in a condition but i will open a separate issue for those
1
8,037
11,214,947,397
IssuesEvent
2020-01-07 00:14:56
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Test failure : System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated
arch-x64 area-System.Diagnostics.Process os-linux test-corefx
**Job:** [coreclr-corefx-jitstress:20191105.1](https://dev.azure.com/dnceng/9ee6d478-d288-47f7-aacc-f6e6d082ae6d/_build/results?buildId=417295) **Details:** https://helix.dot.net/api/2019-06-17/jobs/627db8f7-fc83-49e4-a05e-643d903e29f1/workitems/System.Diagnostics.Process.Tests/console **OS & Arch:** Linux x64 **Mode:** export COMPlus_TieredCompilation=0 export COMPlus_DbgEnableMiniDump=1 export COMPlus_DbgMiniDumpName=$HELIX_DUMP_FOLDER/coredump.%d.dmp export COMPlus_ReadyToRun=0 export COMPlus_ZapDisable=1 **Log:** ``` ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Diagnostics.Process.Tests (found 214 of 272 test cases) Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2) System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP] Condition(s) not met: "IsWindowsNanoServer" Linux Linux Linux Linux None of the following programs were installed on this machine: xdg-open,gnome-open,kfmclient. System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated [FAIL] Assert.Equal() Failure Expected: Boolean[] [True, True, True] Actual: List<Boolean> [True, True, False] Stack Trace: /_/src/System.Diagnostics.Process/tests/Helpers.cs(29,0): at System.Diagnostics.Tests.Helpers.RetryWithBackoff(Action action, Int32 delayInMilliseconds, Int32 times) /_/src/System.Diagnostics.Process/tests/ProcessTests.cs(2130,0): at System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated() --- End of stack trace from previous location --- System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP] Condition(s) not met: "IsStressModeEnabled" Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 268, Errors: 0, Failed: 1, Skipped: 2, Time: 129.059s ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec ----- end Wed Nov 6 08:09:22 UTC 2019 ----- exit code 1 ---------------------------------------------------------- ```
1.0
Test failure : System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated - **Job:** [coreclr-corefx-jitstress:20191105.1](https://dev.azure.com/dnceng/9ee6d478-d288-47f7-aacc-f6e6d082ae6d/_build/results?buildId=417295) **Details:** https://helix.dot.net/api/2019-06-17/jobs/627db8f7-fc83-49e4-a05e-643d903e29f1/workitems/System.Diagnostics.Process.Tests/console **OS & Arch:** Linux x64 **Mode:** export COMPlus_TieredCompilation=0 export COMPlus_DbgEnableMiniDump=1 export COMPlus_DbgMiniDumpName=$HELIX_DUMP_FOLDER/coredump.%d.dmp export COMPlus_ReadyToRun=0 export COMPlus_ZapDisable=1 **Log:** ``` ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Diagnostics.Process.Tests (found 214 of 272 test cases) Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2) System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP] Condition(s) not met: "IsWindowsNanoServer" Linux Linux Linux Linux None of the following programs were installed on this machine: xdg-open,gnome-open,kfmclient. System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated [FAIL] Assert.Equal() Failure Expected: Boolean[] [True, True, True] Actual: List<Boolean> [True, True, False] Stack Trace: /_/src/System.Diagnostics.Process/tests/Helpers.cs(29,0): at System.Diagnostics.Tests.Helpers.RetryWithBackoff(Action action, Int32 delayInMilliseconds, Int32 times) /_/src/System.Diagnostics.Process/tests/ProcessTests.cs(2130,0): at System.Diagnostics.Tests.ProcessTests.Kill_EntireProcessTree_True_EntireTreeTerminated() --- End of stack trace from previous location --- System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP] Condition(s) not met: "IsStressModeEnabled" Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 268, Errors: 0, Failed: 1, Skipped: 2, Time: 129.059s ~/work/627db8f7-fc83-49e4-a05e-643d903e29f1/Work/760c06b2-9f16-40c8-b3a7-da9077b7c051/Exec ----- end Wed Nov 6 08:09:22 UTC 2019 ----- exit code 1 ---------------------------------------------------------- ```
process
test failure system diagnostics tests processtests kill entireprocesstree true entiretreeterminated job details os arch linux mode export complus tieredcompilation export complus dbgenableminidump export complus dbgminidumpname helix dump folder coredump d dmp export complus readytorun export complus zapdisable log work work exec work work exec discovering system diagnostics process tests method display classandmethod method display options none discovered system diagnostics process tests found of test cases starting system diagnostics process tests parallel test collections on max threads system diagnostics tests processstartinfotests shellexecute nano fails start condition s not met iswindowsnanoserver linux linux linux linux none of the following programs were installed on this machine xdg open gnome open kfmclient system diagnostics tests processtests kill entireprocesstree true entiretreeterminated assert equal failure expected boolean actual list stack trace src system diagnostics process tests helpers cs at system diagnostics tests helpers retrywithbackoff action action delayinmilliseconds times src system diagnostics process tests processtests cs at system diagnostics tests processtests kill entireprocesstree true entiretreeterminated end of stack trace from previous location system diagnostics tests processtests testprocessrecycledpid condition s not met isstressmodeenabled finished system diagnostics process tests test execution summary system diagnostics process tests total errors failed skipped time work work exec end wed nov utc exit code
1
111,807
11,742,337,729
IssuesEvent
2020-03-12 00:25:15
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
Improve events module documentation
contributions-welcome documentation events priority: maybe someday technical change
In GitLab by @thomwiggers on Feb 13, 2018, 11:54 The ~events module could use improved docs.
1.0
Improve events module documentation - In GitLab by @thomwiggers on Feb 13, 2018, 11:54 The ~events module could use improved docs.
non_process
improve events module documentation in gitlab by thomwiggers on feb the events module could use improved docs
0
13,749
16,502,591,531
IssuesEvent
2021-05-25 15:42:35
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
In the new version of Python azure.storage.blob package no BlockBlobService exists
Pri2 assigned-to-author machine-learning/svc product-question team-data-science-process/subsvc triaged
Hi Microsoft, in the new version of Python azure.storage.blob package no BlockBlobService exists!!! from azure.storage.blob import BlockBlobService returns en error: ImportError: cannot import name 'BlockBlobService' from 'azure.storage.blob' (C:\ProgramDat..... --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2f45a6b5-0fea-7fbb-5d4d-37e0e2583fd7 * Version Independent ID: 7be6f792-09f8-c22f-86b6-d0f690e9b3a4 * Content: [Explore data in Azure Blob Storage with pandas - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/explore-data-blob) * Content Source: [articles/machine-learning/team-data-science-process/explore-data-blob.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/machine-learning/team-data-science-process/explore-data-blob.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
1.0
In the new version of Python azure.storage.blob package no BlockBlobService exists - Hi Microsoft, in the new version of Python azure.storage.blob package no BlockBlobService exists!!! from azure.storage.blob import BlockBlobService returns en error: ImportError: cannot import name 'BlockBlobService' from 'azure.storage.blob' (C:\ProgramDat..... --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2f45a6b5-0fea-7fbb-5d4d-37e0e2583fd7 * Version Independent ID: 7be6f792-09f8-c22f-86b6-d0f690e9b3a4 * Content: [Explore data in Azure Blob Storage with pandas - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/explore-data-blob) * Content Source: [articles/machine-learning/team-data-science-process/explore-data-blob.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/machine-learning/team-data-science-process/explore-data-blob.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
process
in the new version of python azure storage blob package no blockblobservice exists hi microsoft in the new version of python azure storage blob package no blockblobservice exists from azure storage blob import blockblobservice returns en error importerror cannot import name blockblobservice from azure storage blob c programdat document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service machine learning sub service team data science process github login marktab microsoft alias tdsp
1
7,400
10,523,138,640
IssuesEvent
2019-09-30 10:16:39
didi/mpx
https://api.github.com/repos/didi/mpx
closed
mpx 微信小程序 转义字符 报错
processing
页面中需要显示 "<" 这种特殊字符, 按照微信的写法是 ` <text decode="true"> A &lt; B</text>`, 但 mpx编译后的代码为: `<view><list mpxPageStatus="{{mpxPageStatus}}"></list> <text decode="true"> A < B</text></view> ` 先将特殊字符转了, 导致微信开发工具中报错了
1.0
mpx 微信小程序 转义字符 报错 - 页面中需要显示 "<" 这种特殊字符, 按照微信的写法是 ` <text decode="true"> A &lt; B</text>`, 但 mpx编译后的代码为: `<view><list mpxPageStatus="{{mpxPageStatus}}"></list> <text decode="true"> A < B</text></view> ` 先将特殊字符转了, 导致微信开发工具中报错了
process
mpx 微信小程序 转义字符 报错 页面中需要显示 a lt b 但 mpx编译后的代码为 a 先将特殊字符转了 导致微信开发工具中报错了
1
13,427
15,881,156,445
IssuesEvent
2021-04-09 14:30:19
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Hyperlinks are displayed as normal text in resources in android
Android Bug P1 Process: Fixed Process: Tested QA Process: Tested dev
Steps:- 1. Add a resource text with hyperlinks. eg:- http://www.google.com in WCP 2. Open the application and enroll into study 3. Navigate to resource section and find the resource created 4. Verify the hyperlinks in the resource A/R: Hyperlinks are displayed as normal text and not clickable E/R:- Hyperlinks should be clickable links and navigate to respective site ![image](https://user-images.githubusercontent.com/60500517/110322057-8f8f7700-8038-11eb-8814-2a3886d6307f.png)
3.0
[Android] Hyperlinks are displayed as normal text in resources in android - Steps:- 1. Add a resource text with hyperlinks. eg:- http://www.google.com in WCP 2. Open the application and enroll into study 3. Navigate to resource section and find the resource created 4. Verify the hyperlinks in the resource A/R: Hyperlinks are displayed as normal text and not clickable E/R:- Hyperlinks should be clickable links and navigate to respective site ![image](https://user-images.githubusercontent.com/60500517/110322057-8f8f7700-8038-11eb-8814-2a3886d6307f.png)
process
hyperlinks are displayed as normal text in resources in android steps add a resource text with hyperlinks eg in wcp open the application and enroll into study navigate to resource section and find the resource created verify the hyperlinks in the resource a r hyperlinks are displayed as normal text and not clickable e r hyperlinks should be clickable links and navigate to respective site
1
20,351
27,010,135,908
IssuesEvent
2023-02-10 14:48:32
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Larger CI Runners to Prevent MIRI OOMing and Improve CI Times
question enhancement development-process
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.** Since updating MIRI in #1828 it is periodically OOMing - https://github.com/apache/arrow-rs/actions/workflows/miri.yaml ![image](https://user-images.githubusercontent.com/1781103/173018511-f61ee955-f90a-467e-8666-f9e1f017d088.png) https://github.com/apache/arrow-rs/actions/runs/2473012537 **Describe the solution you'd like** I'm not entirely sure what the best course of action here is, rolling back to a 6 month old MIRI is not ideal and would require backing out changes in #1822, but then neither is having it randomly fail. It has been a long-time annoyance of mine that the the CI currently takes ~40 minutes to chug through, despite significant caching. This is largely because the runners are rather piddly - https://docs.github.com/en/actions/using-github-hosted-runners/about-github-hosted-runners#supported-runners-and-hardware-resources. This also precludes automatically running any meaningful benchmarks (#1274). Perhaps we should invest some time into a more powerful CI system such as [buildkite](https://buildkite.com/apache-arrow) which is used by other arrow projects... **Describe alternatives you've considered** None
1.0
Larger CI Runners to Prevent MIRI OOMing and Improve CI Times - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.** Since updating MIRI in #1828 it is periodically OOMing - https://github.com/apache/arrow-rs/actions/workflows/miri.yaml ![image](https://user-images.githubusercontent.com/1781103/173018511-f61ee955-f90a-467e-8666-f9e1f017d088.png) https://github.com/apache/arrow-rs/actions/runs/2473012537 **Describe the solution you'd like** I'm not entirely sure what the best course of action here is, rolling back to a 6 month old MIRI is not ideal and would require backing out changes in #1822, but then neither is having it randomly fail. It has been a long-time annoyance of mine that the the CI currently takes ~40 minutes to chug through, despite significant caching. This is largely because the runners are rather piddly - https://docs.github.com/en/actions/using-github-hosted-runners/about-github-hosted-runners#supported-runners-and-hardware-resources. This also precludes automatically running any meaningful benchmarks (#1274). Perhaps we should invest some time into a more powerful CI system such as [buildkite](https://buildkite.com/apache-arrow) which is used by other arrow projects... **Describe alternatives you've considered** None
process
larger ci runners to prevent miri ooming and improve ci times is your feature request related to a problem or challenge please describe what you are trying to do since updating miri in it is periodically ooming describe the solution you d like i m not entirely sure what the best course of action here is rolling back to a month old miri is not ideal and would require backing out changes in but then neither is having it randomly fail it has been a long time annoyance of mine that the the ci currently takes minutes to chug through despite significant caching this is largely because the runners are rather piddly this also precludes automatically running any meaningful benchmarks perhaps we should invest some time into a more powerful ci system such as which is used by other arrow projects describe alternatives you ve considered none
1
107,981
11,577,768,295
IssuesEvent
2020-02-21 14:44:20
saros-project/saros
https://api.github.com/repos/saros-project/saros
closed
Reference last commit of whiteboard
Area: Documentation
In order to allow interested developers to find the last Whiteboard commit we should reference the last commit of #831
1.0
Reference last commit of whiteboard - In order to allow interested developers to find the last Whiteboard commit we should reference the last commit of #831
non_process
reference last commit of whiteboard in order to allow interested developers to find the last whiteboard commit we should reference the last commit of
0
15,571
19,703,505,734
IssuesEvent
2022-01-12 19:08:09
googleapis/nodejs-analytics-data
https://api.github.com/repos/googleapis/nodejs-analytics-data
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'analytics-data' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'analytics-data' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname analytics data invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
264,500
23,120,973,243
IssuesEvent
2022-07-27 21:27:04
w3c/aria-at
https://api.github.com/repos/w3c/aria-at
opened
Asserting screen reader mode changes in ARIA-AT tests
Agenda+Community Group tests AT-Automation Agenda+Test Writing
#### Background Screen readers offer different modes for reading and interacting with content and controls on a web page. These modes are required to suit the various situations a user may find themselves in, and the actions they must carry out to deal with them. For example: 1. When on a text-heavy page, such as a Wikipedia article, users need to read text in sequence using the arrow keys, jump to different headed sections using H/Shift+H, etc. As such, the screen reader's own specific cursor will be active, and many keystrokes will be entirely handled by the screen reader. 2. When filling out a form, users need to type text in the various inputs, scroll through the choices in radio groups and select dropdowns, and toggle checkboxes. Accordingly, the screen reader's own cursor will not be active, and many keystrokes that would otherwise be used for controlling that cursor will be passed through to the browser/web page instead. This includes the arrow keys and similar, e.g. to control the position of the system caret within a text field. 3. When operating more complex controls, including so-called "composite widgets", keyboard commands are often required to be handled by the web page directly. For example, the arrow keys are used to move within grids, tab lists and toolbars. Similar to the case of filling out a form, this requires a screen reader's own cursor to be inactive, with the appropriate reduction in keystroke handling. As these affordances are shared across screen readers, but with slightly different names and connotations, the ARIA-AT project has chosen to adopt an abstracted set of terms to describe them. Please see the [Screen Reader Terminology Translation page on the wiki](https://github.com/w3c/aria-at/wiki/Screen-Reader-Terminology-Translation) for more details. #### Automatic Mode Switching As web pages continue to increase in complexity, frequently changing the active mode is a requirement. To save users the overhead of having to carry this out manually, screen readers implement patterns of automatic mode switching to respond to common scenarios (e.g. focus moving into a text field). It is of critical importance that these mode switches be applicable and consistent, to create a predictable experience and avoid users having to second-guess their screen reader's behaviour. However, the majority of ARIA-AT tests do not currently include assertions targeting this aspect of functionality; the sole exception is the [Menubar Editor plan](https://aria-at.netlify.app/review/menubar-editor.html), which includes assertions following the wording: > Change of mode from reading to interaction is conveyed This is a situation we would like to change, but there are a number of questions/concerns to be addressed before doing so. #### Outstanding Questions/Concerns Note: these are what we've come up with at PAC, there may be more. ##### Sound-Based Feedback With JAWS and NVDA configured to use default settings, mode changes are only conveyed via sound. That is, the speech output does not reflect whether or not the mode has switched. This may make it more difficult for testers to unambiguously determine that a mode switch has occurred, and/or which one, particularly if a screen reader switches modes multiple times in rapid succession. In such a case, a sound may interrupt or overlap with the playback of another, and testers would need to take additional steps to verify the new state. It also prevents deaf-blind braille users from running tests, which they are currently able to do. Meanwhile, the lack of speech output is a problem while reviewing conflicts within results. When in doubt about what a particular tester experienced, the test admin should be able to use the recorded output as a canonical source of truth, and map it to all of the assertions accordingly. If the output does not include a significant area of feedback used by the tester, this will not be possible to the same degree. ##### Indefinite Assertion Wording Related to the lack of speech output, the assertion text may not be immediately understandable. We may just want to stick with what exists for the menubar tests, but possibly link to the wiki page somewhere for testers to reference. ##### Mode Applicability Inconsistencies On the previously linked [Screen Reader Terminology Translation wiki page](https://github.com/w3c/aria-at/wiki/Screen-Reader-Terminology-Translation), the "Desktop Screen Reader Terms" section states that: > ... MacOS VoiceOver technically does not have a mode equivalent to JAWS virtual cursor mode. However, behavior of VoiceOver is sufficiently similar when quick nav keys are toggled on for the ARIA-AT project to treat the quick nav on state of VoiceOver as equivalent to JAWS virtual cursor mode. The table in the same section creates equivalencies between Reading Mode and Quick Nav being enabled, and Interaction Mode with Quick Nav being disabled. But none of the tests reflect this, including the aforementioned Menubar plan. Testers are simply instructed to turn Quick Nav off, always, and VoiceOver is essentially treated as "modeless". Note: I believe VO also uses sounds by default, to indicate when Quick Nav is on or off. Needs to be verified. ##### Automation Approach Once test running is automated, it isn't currently planned to capture non-speech-based feedback from a screen reader, so a solution will eventually need to be found to the lack of it. It may benefit everyone to find that solution now and apply it to human testing. #### Potential Solutions (not necessarily mutually exclusive) 1. Instruct testers to change the relevant screen reader settings to use speech feedback instead of sounds. Not sure if this is doable in JAWS with the default mode active. Also don't know if this can be changed independently in VO, or whether the tester would have to turn off all sound-based feedback. 2. Treat VoiceOver as a modal screen reader, i.e. include it within Reading and Interaction Mode tests as is the case for NVDA and JAWS, and ensure that all tests are carried out with Quick Nav on and off. 3. Provide a means for a tester to indicate that they heard the Reading and/or Interaction Mode sounds, e.g. with checkboxes and buttons to hear what they sound like in the app. Note that this is exclusionary for certain audiences, as already described, and would not resolve it for automation purposes. 4. Have the automation spec include capture of played sound files, and/or special handling of screen reader internal state.
2.0
Asserting screen reader mode changes in ARIA-AT tests - #### Background Screen readers offer different modes for reading and interacting with content and controls on a web page. These modes are required to suit the various situations a user may find themselves in, and the actions they must carry out to deal with them. For example: 1. When on a text-heavy page, such as a Wikipedia article, users need to read text in sequence using the arrow keys, jump to different headed sections using H/Shift+H, etc. As such, the screen reader's own specific cursor will be active, and many keystrokes will be entirely handled by the screen reader. 2. When filling out a form, users need to type text in the various inputs, scroll through the choices in radio groups and select dropdowns, and toggle checkboxes. Accordingly, the screen reader's own cursor will not be active, and many keystrokes that would otherwise be used for controlling that cursor will be passed through to the browser/web page instead. This includes the arrow keys and similar, e.g. to control the position of the system caret within a text field. 3. When operating more complex controls, including so-called "composite widgets", keyboard commands are often required to be handled by the web page directly. For example, the arrow keys are used to move within grids, tab lists and toolbars. Similar to the case of filling out a form, this requires a screen reader's own cursor to be inactive, with the appropriate reduction in keystroke handling. As these affordances are shared across screen readers, but with slightly different names and connotations, the ARIA-AT project has chosen to adopt an abstracted set of terms to describe them. Please see the [Screen Reader Terminology Translation page on the wiki](https://github.com/w3c/aria-at/wiki/Screen-Reader-Terminology-Translation) for more details. #### Automatic Mode Switching As web pages continue to increase in complexity, frequently changing the active mode is a requirement. To save users the overhead of having to carry this out manually, screen readers implement patterns of automatic mode switching to respond to common scenarios (e.g. focus moving into a text field). It is of critical importance that these mode switches be applicable and consistent, to create a predictable experience and avoid users having to second-guess their screen reader's behaviour. However, the majority of ARIA-AT tests do not currently include assertions targeting this aspect of functionality; the sole exception is the [Menubar Editor plan](https://aria-at.netlify.app/review/menubar-editor.html), which includes assertions following the wording: > Change of mode from reading to interaction is conveyed This is a situation we would like to change, but there are a number of questions/concerns to be addressed before doing so. #### Outstanding Questions/Concerns Note: these are what we've come up with at PAC, there may be more. ##### Sound-Based Feedback With JAWS and NVDA configured to use default settings, mode changes are only conveyed via sound. That is, the speech output does not reflect whether or not the mode has switched. This may make it more difficult for testers to unambiguously determine that a mode switch has occurred, and/or which one, particularly if a screen reader switches modes multiple times in rapid succession. In such a case, a sound may interrupt or overlap with the playback of another, and testers would need to take additional steps to verify the new state. It also prevents deaf-blind braille users from running tests, which they are currently able to do. Meanwhile, the lack of speech output is a problem while reviewing conflicts within results. When in doubt about what a particular tester experienced, the test admin should be able to use the recorded output as a canonical source of truth, and map it to all of the assertions accordingly. If the output does not include a significant area of feedback used by the tester, this will not be possible to the same degree. ##### Indefinite Assertion Wording Related to the lack of speech output, the assertion text may not be immediately understandable. We may just want to stick with what exists for the menubar tests, but possibly link to the wiki page somewhere for testers to reference. ##### Mode Applicability Inconsistencies On the previously linked [Screen Reader Terminology Translation wiki page](https://github.com/w3c/aria-at/wiki/Screen-Reader-Terminology-Translation), the "Desktop Screen Reader Terms" section states that: > ... MacOS VoiceOver technically does not have a mode equivalent to JAWS virtual cursor mode. However, behavior of VoiceOver is sufficiently similar when quick nav keys are toggled on for the ARIA-AT project to treat the quick nav on state of VoiceOver as equivalent to JAWS virtual cursor mode. The table in the same section creates equivalencies between Reading Mode and Quick Nav being enabled, and Interaction Mode with Quick Nav being disabled. But none of the tests reflect this, including the aforementioned Menubar plan. Testers are simply instructed to turn Quick Nav off, always, and VoiceOver is essentially treated as "modeless". Note: I believe VO also uses sounds by default, to indicate when Quick Nav is on or off. Needs to be verified. ##### Automation Approach Once test running is automated, it isn't currently planned to capture non-speech-based feedback from a screen reader, so a solution will eventually need to be found to the lack of it. It may benefit everyone to find that solution now and apply it to human testing. #### Potential Solutions (not necessarily mutually exclusive) 1. Instruct testers to change the relevant screen reader settings to use speech feedback instead of sounds. Not sure if this is doable in JAWS with the default mode active. Also don't know if this can be changed independently in VO, or whether the tester would have to turn off all sound-based feedback. 2. Treat VoiceOver as a modal screen reader, i.e. include it within Reading and Interaction Mode tests as is the case for NVDA and JAWS, and ensure that all tests are carried out with Quick Nav on and off. 3. Provide a means for a tester to indicate that they heard the Reading and/or Interaction Mode sounds, e.g. with checkboxes and buttons to hear what they sound like in the app. Note that this is exclusionary for certain audiences, as already described, and would not resolve it for automation purposes. 4. Have the automation spec include capture of played sound files, and/or special handling of screen reader internal state.
non_process
asserting screen reader mode changes in aria at tests background screen readers offer different modes for reading and interacting with content and controls on a web page these modes are required to suit the various situations a user may find themselves in and the actions they must carry out to deal with them for example when on a text heavy page such as a wikipedia article users need to read text in sequence using the arrow keys jump to different headed sections using h shift h etc as such the screen reader s own specific cursor will be active and many keystrokes will be entirely handled by the screen reader when filling out a form users need to type text in the various inputs scroll through the choices in radio groups and select dropdowns and toggle checkboxes accordingly the screen reader s own cursor will not be active and many keystrokes that would otherwise be used for controlling that cursor will be passed through to the browser web page instead this includes the arrow keys and similar e g to control the position of the system caret within a text field when operating more complex controls including so called composite widgets keyboard commands are often required to be handled by the web page directly for example the arrow keys are used to move within grids tab lists and toolbars similar to the case of filling out a form this requires a screen reader s own cursor to be inactive with the appropriate reduction in keystroke handling as these affordances are shared across screen readers but with slightly different names and connotations the aria at project has chosen to adopt an abstracted set of terms to describe them please see the for more details automatic mode switching as web pages continue to increase in complexity frequently changing the active mode is a requirement to save users the overhead of having to carry this out manually screen readers implement patterns of automatic mode switching to respond to common scenarios e g focus moving into a text field it is of critical importance that these mode switches be applicable and consistent to create a predictable experience and avoid users having to second guess their screen reader s behaviour however the majority of aria at tests do not currently include assertions targeting this aspect of functionality the sole exception is the which includes assertions following the wording change of mode from reading to interaction is conveyed this is a situation we would like to change but there are a number of questions concerns to be addressed before doing so outstanding questions concerns note these are what we ve come up with at pac there may be more sound based feedback with jaws and nvda configured to use default settings mode changes are only conveyed via sound that is the speech output does not reflect whether or not the mode has switched this may make it more difficult for testers to unambiguously determine that a mode switch has occurred and or which one particularly if a screen reader switches modes multiple times in rapid succession in such a case a sound may interrupt or overlap with the playback of another and testers would need to take additional steps to verify the new state it also prevents deaf blind braille users from running tests which they are currently able to do meanwhile the lack of speech output is a problem while reviewing conflicts within results when in doubt about what a particular tester experienced the test admin should be able to use the recorded output as a canonical source of truth and map it to all of the assertions accordingly if the output does not include a significant area of feedback used by the tester this will not be possible to the same degree indefinite assertion wording related to the lack of speech output the assertion text may not be immediately understandable we may just want to stick with what exists for the menubar tests but possibly link to the wiki page somewhere for testers to reference mode applicability inconsistencies on the previously linked the desktop screen reader terms section states that macos voiceover technically does not have a mode equivalent to jaws virtual cursor mode however behavior of voiceover is sufficiently similar when quick nav keys are toggled on for the aria at project to treat the quick nav on state of voiceover as equivalent to jaws virtual cursor mode the table in the same section creates equivalencies between reading mode and quick nav being enabled and interaction mode with quick nav being disabled but none of the tests reflect this including the aforementioned menubar plan testers are simply instructed to turn quick nav off always and voiceover is essentially treated as modeless note i believe vo also uses sounds by default to indicate when quick nav is on or off needs to be verified automation approach once test running is automated it isn t currently planned to capture non speech based feedback from a screen reader so a solution will eventually need to be found to the lack of it it may benefit everyone to find that solution now and apply it to human testing potential solutions not necessarily mutually exclusive instruct testers to change the relevant screen reader settings to use speech feedback instead of sounds not sure if this is doable in jaws with the default mode active also don t know if this can be changed independently in vo or whether the tester would have to turn off all sound based feedback treat voiceover as a modal screen reader i e include it within reading and interaction mode tests as is the case for nvda and jaws and ensure that all tests are carried out with quick nav on and off provide a means for a tester to indicate that they heard the reading and or interaction mode sounds e g with checkboxes and buttons to hear what they sound like in the app note that this is exclusionary for certain audiences as already described and would not resolve it for automation purposes have the automation spec include capture of played sound files and or special handling of screen reader internal state
0
418,951
12,215,813,503
IssuesEvent
2020-05-01 13:49:31
googleapis/nodejs-firestore
https://api.github.com/repos/googleapis/nodejs-firestore
opened
Synthesis failed for nodejs-firestore
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate nodejs-firestore. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth' 2020-05-01 05:10:29 [INFO] Running synthtool 2020-05-01 05:10:29 [INFO] ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-01 05:10:29,800 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. On branch autosynth nothing to commit, working tree clean 2020-05-01 05:10:29,945 synthtool > Ensuring dependencies. 2020-05-01 05:10:29,949 synthtool > Cloning googleapis. 2020-05-01 05:10:31,596 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:32,481 synthtool > Generating code for: google/firestore/admin/v1. 2020-05-01 05:10:33,848 synthtool > Generated code into /tmpfs/tmp/tmplmq06gly. 2020-05-01 05:10:33,848 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:34,724 synthtool > Generating code for: google/firestore/v1beta1. 2020-05-01 05:10:35,968 synthtool > Generated code into /tmpfs/tmp/tmpzoiqc1hf. 2020-05-01 05:10:35,969 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:36,842 synthtool > Generating code for: google/firestore/v1. 2020-05-01 05:10:38,070 synthtool > Generated code into /tmpfs/tmp/tmpyv68ng_z. 2020-05-01 05:10:38,144 synthtool > No replacements made in dev/src/v1beta1/firestore_client.ts for pattern return this\._innerApiCalls\.listen\(options\);, maybe replacement is no longer needed? 2020-05-01 05:10:38,144 synthtool > No replacements made in dev/src/v1/firestore_client.ts for pattern return this\._innerApiCalls\.listen\(options\);, maybe replacement is no longer needed? .eslintignore .eslintrc.json .github/ISSUE_TEMPLATE/bug_report.md .github/ISSUE_TEMPLATE/feature_request.md .github/ISSUE_TEMPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/publish.yml .github/release-please.yml .github/workflows/ci.yaml .kokoro/common.cfg .kokoro/continuous/node10/common.cfg .kokoro/continuous/node10/docs.cfg .kokoro/continuous/node10/lint.cfg .kokoro/continuous/node10/samples-test.cfg .kokoro/continuous/node10/system-test.cfg .kokoro/continuous/node10/test.cfg .kokoro/continuous/node12/common.cfg .kokoro/continuous/node12/test.cfg .kokoro/continuous/node8/common.cfg .kokoro/continuous/node8/test.cfg .kokoro/docs.sh .kokoro/lint.sh .kokoro/presubmit/node10/common.cfg .kokoro/presubmit/node10/docs.cfg .kokoro/presubmit/node10/lint.cfg .kokoro/presubmit/node10/samples-test.cfg .kokoro/presubmit/node10/system-test.cfg .kokoro/presubmit/node10/test.cfg .kokoro/presubmit/node12/common.cfg .kokoro/presubmit/node12/test.cfg .kokoro/presubmit/node8/common.cfg .kokoro/presubmit/node8/test.cfg .kokoro/presubmit/windows/common.cfg .kokoro/presubmit/windows/test.cfg .kokoro/publish.sh .kokoro/release/docs.cfg .kokoro/release/docs.sh .kokoro/release/publish.cfg .kokoro/samples-test.sh .kokoro/system-test.sh .kokoro/test.bat .kokoro/test.sh .kokoro/trampoline.sh .mocharc.js .nycrc .prettierignore .prettierrc.js CODE_OF_CONDUCT.md CONTRIBUTING.md LICENSE README.md renovate.json samples/README.md 2020-05-01 05:10:38,825 synthtool > Replaced "/protos/protos'" in dev/src/v1/firestore_client.ts. 2020-05-01 05:10:38,825 synthtool > No replacements made in dev/test/gapic-firestore-v1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,826 synthtool > Replaced "/protos/protos'" in dev/src/v1/firestore_admin_client.ts. 2020-05-01 05:10:38,826 synthtool > No replacements made in dev/test/gapic-firestore_admin-v1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,826 synthtool > Replaced "/protos/protos'" in dev/src/v1beta1/firestore_client.ts. 2020-05-01 05:10:38,827 synthtool > No replacements made in dev/test/gapic-firestore-v1beta1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,829 synthtool > Installing dependencies... npm WARN npm npm does not support Node.js v12.16.3 npm WARN npm You should probably upgrade to a newer version of node as we npm WARN npm can't make any promises that npm will work with this version. npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11. npm WARN npm You can find the latest version at https://nodejs.org/ npm WARN deprecated @types/moment@2.13.0: This is a stub types definition for Moment (https://github.com/moment/moment). Moment provides its own type definitions, so you don't need @types/moment installed! > protobufjs@6.9.0 postinstall /tmpfs/src/git/autosynth/working_repo/node_modules/protobufjs > node scripts/postinstall > @google-cloud/firestore@3.7.5 prepare /tmpfs/src/git/autosynth/working_repo > npm run compile npm WARN npm npm does not support Node.js v12.16.3 npm WARN npm You should probably upgrade to a newer version of node as we npm WARN npm can't make any promises that npm will work with this version. npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11. npm WARN npm You can find the latest version at https://nodejs.org/ > @google-cloud/firestore@3.7.5 precompile /tmpfs/src/git/autosynth/working_repo > gts clean Removing build ... > @google-cloud/firestore@3.7.5 compile /tmpfs/src/git/autosynth/working_repo > tsc -p . error TS2688: Cannot find type definition file for 'moment'. dev/src/serializer.ts:17:22 - error TS2307: Cannot find module 'moment'. 17 import {Moment} from 'moment';    ~~~~~~~~ dev/src/serializer.ts:149:9 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 149 if (val instanceof Buffer || val instanceof Uint8Array) {    ~~~ dev/src/serializer.ts:149:34 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 149 if (val instanceof Buffer || val instanceof Uint8Array) {    ~~~ dev/src/serializer.ts:158:24 - error TS2339: Property 'bind' does not exist on type 'never'. 158 return toProto.bind(val)();    ~~~~ dev/src/serializer.ts:162:9 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 162 if (val instanceof Array) {    ~~~ dev/src/serializer.ts:167:15 - error TS2339: Property 'length' does not exist on type 'never'. 167 if (val.length > 0) {    ~~~~~~ dev/src/serializer.ts:169:33 - error TS2339: Property 'length' does not exist on type 'never'. 169 for (let i = 0; i < val.length; ++i) {    ~~~~~~ dev/src/serializer.ts:386:14 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 386 } else if (value instanceof Buffer || value instanceof Uint8Array) {    ~~~~~ dev/src/serializer.ts:386:41 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 386 } else if (value instanceof Buffer || value instanceof Uint8Array) {    ~~~~~ dev/src/v1/firestore_admin_client.ts:1046:46 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1046 return this.descriptors.page.listIndexes.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_admin_client.ts:1252:45 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1252 return this.descriptors.page.listFields.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_client.ts:1199:48 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1199 return this.descriptors.page.listDocuments.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_client.ts:1385:52 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1385 return this.descriptors.page.listCollectionIds.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1beta1/firestore_client.ts:1207:48 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1207 return this.descriptors.page.listDocuments.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1beta1/firestore_client.ts:1393:52 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1393 return this.descriptors.page.listCollectionIds.asyncIterate(    ~~~~~~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:107:50 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 107 const servicePath = firestoreadminModule.v1.FirestoreAdminClient.servicePath;    ~~ dev/test/gapic_firestore_admin_v1.ts:112:50 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const apiEndpoint = firestoreadminModule.v1.FirestoreAdminClient.apiEndpoint;    ~~ dev/test/gapic_firestore_admin_v1.ts:117:43 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const port = firestoreadminModule.v1.FirestoreAdminClient.port;    ~~ dev/test/gapic_firestore_admin_v1.ts:123:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 123 const client = new firestoreadminModule.v1.FirestoreAdminClient();    ~~ dev/test/gapic_firestore_admin_v1.ts:128:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:135:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 135 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:145:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 145 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:154:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 154 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:166:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 166 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:186:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 186 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:210:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 210 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:245:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 245 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:270:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 270 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:294:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 294 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:329:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 329 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:354:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 354 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:378:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 378 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:413:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 413 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:438:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 438 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:463:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 463 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:501:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 501 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:524:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 524 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:550:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 550 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:576:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 576 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:615:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 615 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:639:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 639 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:666:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 666 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:691:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 691 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:729:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 729 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:752:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 752 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:778:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 778 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:803:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 803 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:841:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 841 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:864:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 864 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:890:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 890 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:918:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 918 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:957:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 957 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:980:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 980 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1019:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1019 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1053:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1053 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1083:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1083 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1112:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1112 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1140:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1140 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1179:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1179 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1202:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1202 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1241:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1241 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1275:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1275 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1305:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1305 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1341:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1341 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1386:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1386 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1426:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1426 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1480:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1480 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_v1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_v1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1.ts:112:45 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const servicePath = firestoreModule.v1.FirestoreClient.servicePath;    ~~ dev/test/gapic_firestore_v1.ts:117:45 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const apiEndpoint = firestoreModule.v1.FirestoreClient.apiEndpoint;    ~~ dev/test/gapic_firestore_v1.ts:122:38 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 122 const port = firestoreModule.v1.FirestoreClient.port;    ~~ dev/test/gapic_firestore_v1.ts:128:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreModule.v1.FirestoreClient();    ~~ dev/test/gapic_firestore_v1.ts:133:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 133 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:140:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 140 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:150:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 150 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:159:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 159 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:171:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 171 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:191:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 191 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:215:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 215 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:250:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 250 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:275:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 275 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:300:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 300 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:336:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 336 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:362:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 362 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:386:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 386 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:421:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 421 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:446:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 446 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:470:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 470 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:505:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 505 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:530:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 530 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:554:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 554 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:589:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 589 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:614:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 614 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:638:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 638 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:673:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 673 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:698:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 698 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:722:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 722 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:757:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 757 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:782:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 782 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:815:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 815 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:849:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 849 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:882:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 882 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:916:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 916 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:944:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 944 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:974:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 974 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1002:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1002 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1032:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1032 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1060:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1060 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1099:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1099 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1122:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1122 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1161:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1161 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1195:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1195 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1225:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1225 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1254:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1254 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1278:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1278 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1313:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1313 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1336:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1336 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1371:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1371 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1405:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1405 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1431:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1431 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1beta1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:112:45 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const servicePath = firestoreModule.v1beta1.FirestoreClient.servicePath;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:117:45 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const apiEndpoint = firestoreModule.v1beta1.FirestoreClient.apiEndpoint;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:122:38 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 122 const port = firestoreModule.v1beta1.FirestoreClient.port;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:128:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreModule.v1beta1.FirestoreClient();    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:133:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 133 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:140:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 140 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:150:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 150 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:159:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 159 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:171:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 171 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:191:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 191 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:215:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 215 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:250:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 250 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:275:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 275 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:299:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 299 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:334:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 334 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:359:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 359 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:384:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 384 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:420:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 420 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:446:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 446 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:470:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 470 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:505:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 505 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:530:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 530 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:554:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 554 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:589:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 589 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:614:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 614 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:638:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 638 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:673:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 673 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:698:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 698 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:722:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 722 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:757:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 757 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:782:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 782 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:815:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 815 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:849:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 849 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:882:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 882 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:916:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 916 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:944:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 944 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:974:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 974 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1002:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1002 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1032:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1032 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1060:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1060 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1099:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1099 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1122:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1122 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1161:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1161 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1195:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1195 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1225:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1225 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1254:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1254 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1278:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1278 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1313:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1313 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1336:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1336 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1371:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1371 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1405:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1405 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1431:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1431 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ Found 181 errors. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! @google-cloud/firestore@3.7.5 compile: `tsc -p .` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the @google-cloud/firestore@3.7.5 compile script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-05-01T12_11_01_818Z-debug.log npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! @google-cloud/firestore@3.7.5 prepare: `npm run compile` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the @google-cloud/firestore@3.7.5 prepare script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-05-01T12_11_01_868Z-debug.log 2020-05-01 05:11:01,936 synthtool > Failed executing npm install: None 2020-05-01 05:11:01,938 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 91, in <module> node.install() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/languages/node.py", line 102, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2020-05-01 05:11:01 [ERROR] Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 578, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 459, in main return _inner_main(temp_dir) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 532, in _inner_main deprecated_execution=args.deprecated_execution, File "/tmpfs/src/git/autosynth/autosynth/synthesizer.py", line 115, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/2ccb8acc-3741-4738-b22c-5d6eb208779b).
1.0
Synthesis failed for nodejs-firestore - Hello! Autosynth couldn't regenerate nodejs-firestore. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth' 2020-05-01 05:10:29 [INFO] Running synthtool 2020-05-01 05:10:29 [INFO] ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-01 05:10:29,800 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. On branch autosynth nothing to commit, working tree clean 2020-05-01 05:10:29,945 synthtool > Ensuring dependencies. 2020-05-01 05:10:29,949 synthtool > Cloning googleapis. 2020-05-01 05:10:31,596 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:32,481 synthtool > Generating code for: google/firestore/admin/v1. 2020-05-01 05:10:33,848 synthtool > Generated code into /tmpfs/tmp/tmplmq06gly. 2020-05-01 05:10:33,848 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:34,724 synthtool > Generating code for: google/firestore/v1beta1. 2020-05-01 05:10:35,968 synthtool > Generated code into /tmpfs/tmp/tmpzoiqc1hf. 2020-05-01 05:10:35,969 synthtool > Pulling Docker image: gapic-generator-typescript:latest latest: Pulling from gapic-images/gapic-generator-typescript Digest: sha256:d2a53b51de1a001915bdde9380785e412ccd75a38d5427fc259407c66f962a57 Status: Image is up to date for gcr.io/gapic-images/gapic-generator-typescript:latest 2020-05-01 05:10:36,842 synthtool > Generating code for: google/firestore/v1. 2020-05-01 05:10:38,070 synthtool > Generated code into /tmpfs/tmp/tmpyv68ng_z. 2020-05-01 05:10:38,144 synthtool > No replacements made in dev/src/v1beta1/firestore_client.ts for pattern return this\._innerApiCalls\.listen\(options\);, maybe replacement is no longer needed? 2020-05-01 05:10:38,144 synthtool > No replacements made in dev/src/v1/firestore_client.ts for pattern return this\._innerApiCalls\.listen\(options\);, maybe replacement is no longer needed? .eslintignore .eslintrc.json .github/ISSUE_TEMPLATE/bug_report.md .github/ISSUE_TEMPLATE/feature_request.md .github/ISSUE_TEMPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/publish.yml .github/release-please.yml .github/workflows/ci.yaml .kokoro/common.cfg .kokoro/continuous/node10/common.cfg .kokoro/continuous/node10/docs.cfg .kokoro/continuous/node10/lint.cfg .kokoro/continuous/node10/samples-test.cfg .kokoro/continuous/node10/system-test.cfg .kokoro/continuous/node10/test.cfg .kokoro/continuous/node12/common.cfg .kokoro/continuous/node12/test.cfg .kokoro/continuous/node8/common.cfg .kokoro/continuous/node8/test.cfg .kokoro/docs.sh .kokoro/lint.sh .kokoro/presubmit/node10/common.cfg .kokoro/presubmit/node10/docs.cfg .kokoro/presubmit/node10/lint.cfg .kokoro/presubmit/node10/samples-test.cfg .kokoro/presubmit/node10/system-test.cfg .kokoro/presubmit/node10/test.cfg .kokoro/presubmit/node12/common.cfg .kokoro/presubmit/node12/test.cfg .kokoro/presubmit/node8/common.cfg .kokoro/presubmit/node8/test.cfg .kokoro/presubmit/windows/common.cfg .kokoro/presubmit/windows/test.cfg .kokoro/publish.sh .kokoro/release/docs.cfg .kokoro/release/docs.sh .kokoro/release/publish.cfg .kokoro/samples-test.sh .kokoro/system-test.sh .kokoro/test.bat .kokoro/test.sh .kokoro/trampoline.sh .mocharc.js .nycrc .prettierignore .prettierrc.js CODE_OF_CONDUCT.md CONTRIBUTING.md LICENSE README.md renovate.json samples/README.md 2020-05-01 05:10:38,825 synthtool > Replaced "/protos/protos'" in dev/src/v1/firestore_client.ts. 2020-05-01 05:10:38,825 synthtool > No replacements made in dev/test/gapic-firestore-v1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,826 synthtool > Replaced "/protos/protos'" in dev/src/v1/firestore_admin_client.ts. 2020-05-01 05:10:38,826 synthtool > No replacements made in dev/test/gapic-firestore_admin-v1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,826 synthtool > Replaced "/protos/protos'" in dev/src/v1beta1/firestore_client.ts. 2020-05-01 05:10:38,827 synthtool > No replacements made in dev/test/gapic-firestore-v1beta1.ts for pattern /protos/protos', maybe replacement is no longer needed? 2020-05-01 05:10:38,829 synthtool > Installing dependencies... npm WARN npm npm does not support Node.js v12.16.3 npm WARN npm You should probably upgrade to a newer version of node as we npm WARN npm can't make any promises that npm will work with this version. npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11. npm WARN npm You can find the latest version at https://nodejs.org/ npm WARN deprecated @types/moment@2.13.0: This is a stub types definition for Moment (https://github.com/moment/moment). Moment provides its own type definitions, so you don't need @types/moment installed! > protobufjs@6.9.0 postinstall /tmpfs/src/git/autosynth/working_repo/node_modules/protobufjs > node scripts/postinstall > @google-cloud/firestore@3.7.5 prepare /tmpfs/src/git/autosynth/working_repo > npm run compile npm WARN npm npm does not support Node.js v12.16.3 npm WARN npm You should probably upgrade to a newer version of node as we npm WARN npm can't make any promises that npm will work with this version. npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11. npm WARN npm You can find the latest version at https://nodejs.org/ > @google-cloud/firestore@3.7.5 precompile /tmpfs/src/git/autosynth/working_repo > gts clean Removing build ... > @google-cloud/firestore@3.7.5 compile /tmpfs/src/git/autosynth/working_repo > tsc -p . error TS2688: Cannot find type definition file for 'moment'. dev/src/serializer.ts:17:22 - error TS2307: Cannot find module 'moment'. 17 import {Moment} from 'moment';    ~~~~~~~~ dev/src/serializer.ts:149:9 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 149 if (val instanceof Buffer || val instanceof Uint8Array) {    ~~~ dev/src/serializer.ts:149:34 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 149 if (val instanceof Buffer || val instanceof Uint8Array) {    ~~~ dev/src/serializer.ts:158:24 - error TS2339: Property 'bind' does not exist on type 'never'. 158 return toProto.bind(val)();    ~~~~ dev/src/serializer.ts:162:9 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 162 if (val instanceof Array) {    ~~~ dev/src/serializer.ts:167:15 - error TS2339: Property 'length' does not exist on type 'never'. 167 if (val.length > 0) {    ~~~~~~ dev/src/serializer.ts:169:33 - error TS2339: Property 'length' does not exist on type 'never'. 169 for (let i = 0; i < val.length; ++i) {    ~~~~~~ dev/src/serializer.ts:386:14 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 386 } else if (value instanceof Buffer || value instanceof Uint8Array) {    ~~~~~ dev/src/serializer.ts:386:41 - error TS2358: The left-hand side of an 'instanceof' expression must be of type 'any', an object type or a type parameter. 386 } else if (value instanceof Buffer || value instanceof Uint8Array) {    ~~~~~ dev/src/v1/firestore_admin_client.ts:1046:46 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1046 return this.descriptors.page.listIndexes.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_admin_client.ts:1252:45 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1252 return this.descriptors.page.listFields.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_client.ts:1199:48 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1199 return this.descriptors.page.listDocuments.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1/firestore_client.ts:1385:52 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1385 return this.descriptors.page.listCollectionIds.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1beta1/firestore_client.ts:1207:48 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1207 return this.descriptors.page.listDocuments.asyncIterate(    ~~~~~~~~~~~~ dev/src/v1beta1/firestore_client.ts:1393:52 - error TS2339: Property 'asyncIterate' does not exist on type 'PageDescriptor'. 1393 return this.descriptors.page.listCollectionIds.asyncIterate(    ~~~~~~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_admin_v1.ts:107:50 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 107 const servicePath = firestoreadminModule.v1.FirestoreAdminClient.servicePath;    ~~ dev/test/gapic_firestore_admin_v1.ts:112:50 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const apiEndpoint = firestoreadminModule.v1.FirestoreAdminClient.apiEndpoint;    ~~ dev/test/gapic_firestore_admin_v1.ts:117:43 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const port = firestoreadminModule.v1.FirestoreAdminClient.port;    ~~ dev/test/gapic_firestore_admin_v1.ts:123:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 123 const client = new firestoreadminModule.v1.FirestoreAdminClient();    ~~ dev/test/gapic_firestore_admin_v1.ts:128:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:135:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 135 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:145:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 145 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:154:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 154 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:166:49 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 166 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:186:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 186 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:210:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 210 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:245:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 245 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:270:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 270 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:294:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 294 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:329:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 329 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:354:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 354 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:378:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 378 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:413:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 413 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:438:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 438 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:463:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 463 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:501:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 501 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:524:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 524 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:550:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 550 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:576:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 576 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:615:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 615 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:639:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 639 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:666:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 666 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:691:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 691 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:729:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 729 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:752:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 752 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:778:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 778 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:803:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 803 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:841:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 841 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:864:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 864 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:890:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 890 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:918:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 918 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:957:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 957 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:980:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 980 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1019:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1019 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1053:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1053 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1083:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1083 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1112:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1112 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1140:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1140 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1179:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1179 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1202:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1202 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1241:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1241 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1275:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1275 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1305:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1305 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1341:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1341 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1386:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1386 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1426:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1426 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_admin_v1.ts:1480:53 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1480 const client = new firestoreadminModule.v1.FirestoreAdminClient({    ~~ dev/test/gapic_firestore_v1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_v1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1.ts:112:45 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const servicePath = firestoreModule.v1.FirestoreClient.servicePath;    ~~ dev/test/gapic_firestore_v1.ts:117:45 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const apiEndpoint = firestoreModule.v1.FirestoreClient.apiEndpoint;    ~~ dev/test/gapic_firestore_v1.ts:122:38 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 122 const port = firestoreModule.v1.FirestoreClient.port;    ~~ dev/test/gapic_firestore_v1.ts:128:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreModule.v1.FirestoreClient();    ~~ dev/test/gapic_firestore_v1.ts:133:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 133 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:140:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 140 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:150:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 150 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:159:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 159 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:171:44 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 171 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:191:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 191 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:215:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 215 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:250:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 250 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:275:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 275 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:300:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 300 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:336:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 336 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:362:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 362 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:386:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 386 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:421:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 421 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:446:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 446 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:470:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 470 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:505:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 505 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:530:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 530 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:554:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 554 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:589:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 589 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:614:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 614 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:638:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 638 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:673:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 673 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:698:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 698 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:722:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 722 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:757:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 757 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:782:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 782 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:815:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 815 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:849:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 849 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:882:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 882 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:916:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 916 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:944:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 944 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:974:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 974 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1002:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1002 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1032:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1032 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1060:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1060 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1099:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1099 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1122:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1122 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1161:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1161 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1195:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1195 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1225:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1225 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1254:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1254 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1278:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1278 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1313:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1313 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1336:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1336 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1371:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1371 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1405:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1405 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1.ts:1431:48 - error TS2339: Property 'v1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1431 const client = new firestoreModule.v1.FirestoreClient({    ~~ dev/test/gapic_firestore_v1beta1.ts:19:25 - error TS2307: Cannot find module '../protos/protos'. 19 import * as protos from '../protos/protos';    ~~~~~~~~~~~~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:21:24 - error TS2307: Cannot find module 'sinon'. 21 import * as sinon from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:22:25 - error TS2307: Cannot find module 'sinon'. 22 import {SinonStub} from 'sinon';    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:112:45 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 112 const servicePath = firestoreModule.v1beta1.FirestoreClient.servicePath;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:117:45 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 117 const apiEndpoint = firestoreModule.v1beta1.FirestoreClient.apiEndpoint;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:122:38 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 122 const port = firestoreModule.v1beta1.FirestoreClient.port;    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:128:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 128 const client = new firestoreModule.v1beta1.FirestoreClient();    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:133:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 133 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:140:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 140 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:150:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 150 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:159:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 159 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:171:44 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 171 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:191:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 191 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:215:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 215 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:250:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 250 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:275:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 275 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:299:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 299 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:334:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 334 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:359:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 359 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:384:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 384 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:420:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 420 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:446:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 446 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:470:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 470 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:505:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 505 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:530:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 530 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:554:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 554 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:589:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 589 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:614:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 614 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:638:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 638 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:673:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 673 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:698:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 698 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:722:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 722 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:757:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 757 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:782:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 782 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:815:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 815 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:849:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 849 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:882:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 882 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:916:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 916 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:944:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 944 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:974:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 974 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1002:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1002 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1032:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1032 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1060:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1060 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1099:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1099 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1122:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1122 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1161:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1161 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1195:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1195 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1225:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1225 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1254:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1254 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1278:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1278 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1313:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1313 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1336:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1336 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1371:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1371 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1405:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1405 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ dev/test/gapic_firestore_v1beta1.ts:1431:48 - error TS2339: Property 'v1beta1' does not exist on type 'typeof import("/tmpfs/src/git/autosynth/working_repo/dev/src/index")'. 1431 const client = new firestoreModule.v1beta1.FirestoreClient({    ~~~~~~~ Found 181 errors. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! @google-cloud/firestore@3.7.5 compile: `tsc -p .` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the @google-cloud/firestore@3.7.5 compile script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-05-01T12_11_01_818Z-debug.log npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! @google-cloud/firestore@3.7.5 prepare: `npm run compile` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the @google-cloud/firestore@3.7.5 prepare script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-05-01T12_11_01_868Z-debug.log 2020-05-01 05:11:01,936 synthtool > Failed executing npm install: None 2020-05-01 05:11:01,938 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 91, in <module> node.install() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/languages/node.py", line 102, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2020-05-01 05:11:01 [ERROR] Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 578, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 459, in main return _inner_main(temp_dir) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 532, in _inner_main deprecated_execution=args.deprecated_execution, File "/tmpfs/src/git/autosynth/autosynth/synthesizer.py", line 115, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/2ccb8acc-3741-4738-b22c-5d6eb208779b).
non_process
synthesis failed for nodejs firestore hello autosynth couldn t regenerate nodejs firestore broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth running synthtool synthtool executing tmpfs src git autosynth working repo synth py on branch autosynth nothing to commit working tree clean synthtool ensuring dependencies synthtool cloning googleapis synthtool pulling docker image gapic generator typescript latest latest pulling from gapic images gapic generator typescript digest status image is up to date for gcr io gapic images gapic generator typescript latest synthtool generating code for google firestore admin synthtool generated code into tmpfs tmp synthtool pulling docker image gapic generator typescript latest latest pulling from gapic images gapic generator typescript digest status image is up to date for gcr io gapic images gapic generator typescript latest synthtool generating code for google firestore synthtool generated code into tmpfs tmp synthtool pulling docker image gapic generator typescript latest latest pulling from gapic images gapic generator typescript digest status image is up to date for gcr io gapic images gapic generator typescript latest synthtool generating code for google firestore synthtool generated code into tmpfs tmp z synthtool no replacements made in dev src firestore client ts for pattern return this innerapicalls listen options maybe replacement is no longer needed synthtool no replacements made in dev src firestore client ts for pattern return this innerapicalls listen options maybe replacement is no longer needed eslintignore eslintrc json github issue template bug report md github issue template feature request md github issue template support request md github pull request template md github publish yml github release please yml github workflows ci yaml kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro presubmit common cfg kokoro presubmit docs cfg kokoro presubmit lint cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro presubmit common cfg kokoro presubmit test cfg kokoro presubmit common cfg kokoro presubmit test cfg kokoro presubmit windows common cfg kokoro presubmit windows test cfg kokoro publish sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js code of conduct md contributing md license readme md renovate json samples readme md synthtool replaced protos protos in dev src firestore client ts synthtool no replacements made in dev test gapic firestore ts for pattern protos protos maybe replacement is no longer needed synthtool replaced protos protos in dev src firestore admin client ts synthtool no replacements made in dev test gapic firestore admin ts for pattern protos protos maybe replacement is no longer needed synthtool replaced protos protos in dev src firestore client ts synthtool no replacements made in dev test gapic firestore ts for pattern protos protos maybe replacement is no longer needed synthtool installing dependencies npm warn npm npm does not support node js npm warn npm you should probably upgrade to a newer version of node as we npm warn npm can t make any promises that npm will work with this version npm warn npm supported releases of node js are the latest release of npm warn npm you can find the latest version at npm warn deprecated types moment this is a stub types definition for moment moment provides its own type definitions so you don t need types moment installed protobufjs postinstall tmpfs src git autosynth working repo node modules protobufjs node scripts postinstall google cloud firestore prepare tmpfs src git autosynth working repo npm run compile npm warn npm npm does not support node js npm warn npm you should probably upgrade to a newer version of node as we npm warn npm can t make any promises that npm will work with this version npm warn npm supported releases of node js are the latest release of npm warn npm you can find the latest version at google cloud firestore precompile tmpfs src git autosynth working repo gts clean removing build google cloud firestore compile tmpfs src git autosynth working repo tsc p     find type definition file for moment  src serializer ts         find module moment   import moment from moment      src serializer ts         left hand side of an instanceof expression must be of type any an object type or a type parameter   if val instanceof buffer val instanceof      src serializer ts         left hand side of an instanceof expression must be of type any an object type or a type parameter   if val instanceof buffer val instanceof      src serializer ts         bind does not exist on type never   return toproto bind val      src serializer ts         left hand side of an instanceof expression must be of type any an object type or a type parameter   if val instanceof array      src serializer ts         length does not exist on type never   if val length      src serializer ts         length does not exist on type never   for let i i val length i      src serializer ts         left hand side of an instanceof expression must be of type any an object type or a type parameter   else if value instanceof buffer value instanceof      src serializer ts         left hand side of an instanceof expression must be of type any an object type or a type parameter   else if value instanceof buffer value instanceof      src firestore admin client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listindexes asynciterate      src firestore admin client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listfields asynciterate      src firestore client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listdocuments asynciterate      src firestore client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listcollectionids asynciterate      src firestore client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listdocuments asynciterate      src firestore client ts         asynciterate does not exist on type pagedescriptor   return this descriptors page listcollectionids asynciterate      test gapic firestore admin ts         find module protos protos   import as protos from protos protos      test gapic firestore admin ts         find module sinon   import as sinon from sinon      test gapic firestore admin ts         find module sinon   import sinonstub from sinon      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const servicepath firestoreadminmodule firestoreadminclient servicepath      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const apiendpoint firestoreadminmodule firestoreadminclient apiendpoint      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const port firestoreadminmodule firestoreadminclient port      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore admin ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoreadminmodule firestoreadminclient      test gapic firestore ts         find module protos protos   import as protos from protos protos      test gapic firestore ts         find module sinon   import as sinon from sinon      test gapic firestore ts         find module sinon   import sinonstub from sinon      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const servicepath firestoremodule firestoreclient servicepath      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const apiendpoint firestoremodule firestoreclient apiendpoint      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const port firestoremodule firestoreclient port      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         find module protos protos   import as protos from protos protos      test gapic firestore ts         find module sinon   import as sinon from sinon      test gapic firestore ts         find module sinon   import sinonstub from sinon      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const servicepath firestoremodule firestoreclient servicepath      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const apiendpoint firestoremodule firestoreclient apiendpoint      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const port firestoremodule firestoreclient port      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient      test gapic firestore ts         does not exist on type typeof import tmpfs src git autosynth working repo dev src index   const client new firestoremodule firestoreclient     found errors npm err code elifecycle npm err errno npm err google cloud firestore compile tsc p npm err exit status npm err npm err failed at the google cloud firestore compile script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log npm err code elifecycle npm err errno npm err google cloud firestore prepare npm run compile npm err exit status npm err npm err failed at the google cloud firestore prepare script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm install none synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo synth py line in node install file tmpfs src git autosynth env lib site packages synthtool languages node py line in install shell run hide output hide output file tmpfs src git autosynth env lib site packages synthtool shell py line in run raise exc file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main return inner main temp dir file tmpfs src git autosynth autosynth synth py line in inner main deprecated execution args deprecated execution file tmpfs src git autosynth autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
165,975
12,887,237,029
IssuesEvent
2020-07-13 10:51:46
ansible/awx
https://api.github.com/repos/ansible/awx
closed
ANY/ALL convergence setting doesn't save sometimes for approval nodes
component:ui priority:high state:needs_test type:bug
##### ISSUE TYPE - Bug Report ##### STEPS TO REPRODUCE - Create and save workflow that looks like this: ![Screenshot from 2020-05-12 10-12-06](https://user-images.githubusercontent.com/9753817/81702291-12579d80-9439-11ea-9e3a-1039c131bc5c.png) - Open the workflow visualizer for the workflow, change the convergence setting of the approval node to `ALL` and save. ##### EXPECTED RESULTS - The new convergence setting should save and the approval node should now have an `ALL` tag. - There should be api requests in the console log showing that the updated convergence values were sent to the API by the browser. ##### ACTUAL RESULTS - The convergence setting doesn't save - No requests in the browser console logs showing the updated convergence setting. ##### ADDITIONAL INFORMATION Interestingly, the topology of the workflow seems to be a factor. if you create a workflow that looks like this: ![Screenshot from 2020-05-12 10-04-55](https://user-images.githubusercontent.com/9753817/81702626-898d3180-9439-11ea-8339-4883673a0cb8.png) The approval node's convergence setting can be updated as expected.
1.0
ANY/ALL convergence setting doesn't save sometimes for approval nodes - ##### ISSUE TYPE - Bug Report ##### STEPS TO REPRODUCE - Create and save workflow that looks like this: ![Screenshot from 2020-05-12 10-12-06](https://user-images.githubusercontent.com/9753817/81702291-12579d80-9439-11ea-9e3a-1039c131bc5c.png) - Open the workflow visualizer for the workflow, change the convergence setting of the approval node to `ALL` and save. ##### EXPECTED RESULTS - The new convergence setting should save and the approval node should now have an `ALL` tag. - There should be api requests in the console log showing that the updated convergence values were sent to the API by the browser. ##### ACTUAL RESULTS - The convergence setting doesn't save - No requests in the browser console logs showing the updated convergence setting. ##### ADDITIONAL INFORMATION Interestingly, the topology of the workflow seems to be a factor. if you create a workflow that looks like this: ![Screenshot from 2020-05-12 10-04-55](https://user-images.githubusercontent.com/9753817/81702626-898d3180-9439-11ea-8339-4883673a0cb8.png) The approval node's convergence setting can be updated as expected.
non_process
any all convergence setting doesn t save sometimes for approval nodes issue type bug report steps to reproduce create and save workflow that looks like this open the workflow visualizer for the workflow change the convergence setting of the approval node to all and save expected results the new convergence setting should save and the approval node should now have an all tag there should be api requests in the console log showing that the updated convergence values were sent to the api by the browser actual results the convergence setting doesn t save no requests in the browser console logs showing the updated convergence setting additional information interestingly the topology of the workflow seems to be a factor if you create a workflow that looks like this the approval node s convergence setting can be updated as expected
0
16,961
22,321,388,389
IssuesEvent
2022-06-14 06:48:12
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[SQL Connector] Support Sink Configuration.
compute/data-processing type/feature
Support the following Sink Configuration - SINK_TOPIC_ROUTER - SINK_MESSAGE_DELAY_DURATION AC: - testing shows these configs works - testing shows default implementation works.
1.0
[SQL Connector] Support Sink Configuration. - Support the following Sink Configuration - SINK_TOPIC_ROUTER - SINK_MESSAGE_DELAY_DURATION AC: - testing shows these configs works - testing shows default implementation works.
process
support sink configuration support the following sink configuration sink topic router sink message delay duration ac testing shows these configs works testing shows default implementation works
1
2,739
5,630,065,662
IssuesEvent
2017-04-05 11:10:36
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
closed
User image not shown correctly for invoked user in task details
comp: activiti-processList comp: activiti-taskList New Feature
**Type of issue:** (check with "[x]") ``` - [ ] New feature request - [ x ] Bug - [ ] Support request - [ ] Documentation ``` **Current behavior:** Please find attached image where an involved user image is shown differently to how it is stored in the activiti. In this example, the logged in user is the same as involved user. The logged in user image on top left corner is shown correctly. However the involved user is shown incorrectly. <img width="1585" alt="screen shot 2017-02-09 at 12 01 00 am" src="https://cloud.githubusercontent.com/assets/2080360/22771332/b9ff62e8-ee5b-11e6-9c5e-c4d7ab4825d2.png"> **Expected behavior:** Expecting the involved user image to be fetched from activiti **Steps to reproduce the issue:** Create a task and involve a user to the task. Add an image to the involved user in activiti. Check the involved user image in ADF. **Component name and version:** ng2-alfresco-task. v1.1.0 **Browser and version:** Tested in chrome.
1.0
User image not shown correctly for invoked user in task details - **Type of issue:** (check with "[x]") ``` - [ ] New feature request - [ x ] Bug - [ ] Support request - [ ] Documentation ``` **Current behavior:** Please find attached image where an involved user image is shown differently to how it is stored in the activiti. In this example, the logged in user is the same as involved user. The logged in user image on top left corner is shown correctly. However the involved user is shown incorrectly. <img width="1585" alt="screen shot 2017-02-09 at 12 01 00 am" src="https://cloud.githubusercontent.com/assets/2080360/22771332/b9ff62e8-ee5b-11e6-9c5e-c4d7ab4825d2.png"> **Expected behavior:** Expecting the involved user image to be fetched from activiti **Steps to reproduce the issue:** Create a task and involve a user to the task. Add an image to the involved user in activiti. Check the involved user image in ADF. **Component name and version:** ng2-alfresco-task. v1.1.0 **Browser and version:** Tested in chrome.
process
user image not shown correctly for invoked user in task details type of issue check with new feature request bug support request documentation current behavior please find attached image where an involved user image is shown differently to how it is stored in the activiti in this example the logged in user is the same as involved user the logged in user image on top left corner is shown correctly however the involved user is shown incorrectly img width alt screen shot at am src expected behavior expecting the involved user image to be fetched from activiti steps to reproduce the issue create a task and involve a user to the task add an image to the involved user in activiti check the involved user image in adf component name and version alfresco task browser and version tested in chrome
1
81,726
15,630,123,488
IssuesEvent
2021-03-22 01:25:45
ghuangsnl/spring-boot
https://api.github.com/repos/ghuangsnl/spring-boot
opened
CVE-2021-22134 (Medium) detected in elasticsearch-7.7.1.jar
security vulnerability
## CVE-2021-22134 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elasticsearch-7.7.1.jar</b></p></summary> <p>Elasticsearch subproject :server</p> <p>Library home page: <a href="https://github.com/elastic/elasticsearch">https://github.com/elastic/elasticsearch</a></p> <p>Path to dependency file: spring-boot/spring-boot-tests/spring-boot-smoke-tests/spring-boot-smoke-test-data-elasticsearch/build.gradle</p> <p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar,canner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar</p> <p> Dependency Hierarchy: - elasticsearch-rest-high-level-client-7.7.1.jar (Root Library) - :x: **elasticsearch-7.7.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A document disclosure flaw was found in Elasticsearch versions after 7.6.0 and before 7.11.0 when Document or Field Level Security is used. Get requests do not properly apply security permissions when executing a query against a recently updated document. This affects documents that have been updated and not yet refreshed in the index. This could result in the search disclosing the existence of documents and fields the attacker should not be able to view. <p>Publish Date: 2021-03-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22134>CVE-2021-22134</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-hwvv-438r-mhvj">https://github.com/advisories/GHSA-hwvv-438r-mhvj</a></p> <p>Release Date: 2021-03-08</p> <p>Fix Resolution: org.elasticsearch:elasticsearch:7.11.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22134 (Medium) detected in elasticsearch-7.7.1.jar - ## CVE-2021-22134 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elasticsearch-7.7.1.jar</b></p></summary> <p>Elasticsearch subproject :server</p> <p>Library home page: <a href="https://github.com/elastic/elasticsearch">https://github.com/elastic/elasticsearch</a></p> <p>Path to dependency file: spring-boot/spring-boot-tests/spring-boot-smoke-tests/spring-boot-smoke-test-data-elasticsearch/build.gradle</p> <p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar,canner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.elasticsearch/elasticsearch/7.7.1/9fd7d094cb41fe97bafbe8a7183c4036dc5ad8db/elasticsearch-7.7.1.jar</p> <p> Dependency Hierarchy: - elasticsearch-rest-high-level-client-7.7.1.jar (Root Library) - :x: **elasticsearch-7.7.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A document disclosure flaw was found in Elasticsearch versions after 7.6.0 and before 7.11.0 when Document or Field Level Security is used. Get requests do not properly apply security permissions when executing a query against a recently updated document. This affects documents that have been updated and not yet refreshed in the index. This could result in the search disclosing the existence of documents and fields the attacker should not be able to view. <p>Publish Date: 2021-03-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22134>CVE-2021-22134</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-hwvv-438r-mhvj">https://github.com/advisories/GHSA-hwvv-438r-mhvj</a></p> <p>Release Date: 2021-03-08</p> <p>Fix Resolution: org.elasticsearch:elasticsearch:7.11.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in elasticsearch jar cve medium severity vulnerability vulnerable library elasticsearch jar elasticsearch subproject server library home page a href path to dependency file spring boot spring boot tests spring boot smoke tests spring boot smoke test data elasticsearch build gradle path to vulnerable library canner gradle caches modules files org elasticsearch elasticsearch elasticsearch jar canner gradle caches modules files org elasticsearch elasticsearch elasticsearch jar home wss scanner gradle caches modules files org elasticsearch elasticsearch elasticsearch jar dependency hierarchy elasticsearch rest high level client jar root library x elasticsearch jar vulnerable library vulnerability details a document disclosure flaw was found in elasticsearch versions after and before when document or field level security is used get requests do not properly apply security permissions when executing a query against a recently updated document this affects documents that have been updated and not yet refreshed in the index this could result in the search disclosing the existence of documents and fields the attacker should not be able to view publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org elasticsearch elasticsearch step up your open source security game with whitesource
0
189,913
22,047,151,758
IssuesEvent
2022-05-30 04:00:14
nanopathi/linux-4.19.72_CVE-2021-32399
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-32399
closed
CVE-2019-19058 (Medium) detected in linuxlinux-4.19.236 - autoclosed
security vulnerability
## CVE-2019-19058 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the alloc_sgtable() function in drivers/net/wireless/intel/iwlwifi/fw/dbg.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering alloc_page() failures, aka CID-b4b814fec1a5. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19058>CVE-2019-19058</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19058">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19058</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19058 (Medium) detected in linuxlinux-4.19.236 - autoclosed - ## CVE-2019-19058 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the alloc_sgtable() function in drivers/net/wireless/intel/iwlwifi/fw/dbg.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering alloc_page() failures, aka CID-b4b814fec1a5. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19058>CVE-2019-19058</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19058">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19058</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a memory leak in the alloc sgtable function in drivers net wireless intel iwlwifi fw dbg c in the linux kernel through allows attackers to cause a denial of service memory consumption by triggering alloc page failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
5,023
7,845,787,576
IssuesEvent
2018-06-19 13:53:48
openvstorage/alba-asdmanager
https://api.github.com/repos/openvstorage/alba-asdmanager
closed
asd-manager uninstall
process_wontfix type_enhancement
## Issue Right now, the ASD-manager uninstall proceeds as follows: - In the GUI, the ASD node is removed - ` asd-manager remove` command is ran in the command line. ## Fix I propose to make the asd-manager removal independent of the GUI: - No need to first clear the ASD-node through the GUI - Automatically inform alba of recent changes
1.0
asd-manager uninstall - ## Issue Right now, the ASD-manager uninstall proceeds as follows: - In the GUI, the ASD node is removed - ` asd-manager remove` command is ran in the command line. ## Fix I propose to make the asd-manager removal independent of the GUI: - No need to first clear the ASD-node through the GUI - Automatically inform alba of recent changes
process
asd manager uninstall issue right now the asd manager uninstall proceeds as follows in the gui the asd node is removed asd manager remove command is ran in the command line fix i propose to make the asd manager removal independent of the gui no need to first clear the asd node through the gui automatically inform alba of recent changes
1
350,763
24,999,047,453
IssuesEvent
2022-11-03 05:27:08
AY2223S1-CS2103T-W12-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-W12-3/tp
closed
[PE-D][Tester A] command unrecoverable after pressing up
type.Bug type.Documentation type.Chore severity.Low
when pressing the up key, it will go back to the previous input command. But the current command is missing however. <!--session: 1666944150833-55c1343a-f384-4328-84b4-415f06211dcf--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FeatureFlaw` `severity.Medium` original: Echomo-Xinyu/ped#8
1.0
[PE-D][Tester A] command unrecoverable after pressing up - when pressing the up key, it will go back to the previous input command. But the current command is missing however. <!--session: 1666944150833-55c1343a-f384-4328-84b4-415f06211dcf--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FeatureFlaw` `severity.Medium` original: Echomo-Xinyu/ped#8
non_process
command unrecoverable after pressing up when pressing the up key it will go back to the previous input command but the current command is missing however labels type featureflaw severity medium original echomo xinyu ped
0